diff --git a/.Config.yml.swp b/.Config.yml.swp
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/.bandit b/.bandit
new file mode 100644
index 0000000000000000000000000000000000000000..2d81286aee751afae8630a21986eb04f36e7e6d7
--- /dev/null
+++ b/.bandit
@@ -0,0 +1,3 @@
+[bandit]
+exclude = tests
+skips = B101
diff --git a/.config.yml.swp b/.config.yml.swp
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/.editorconfig b/.editorconfig
new file mode 100644
index 0000000000000000000000000000000000000000..df80c53a12dbe60f3e3f60c486cedad044e81ecd
--- /dev/null
+++ b/.editorconfig
@@ -0,0 +1,14 @@
+root = true
+
+[*]
+end_of_line = lf
+insert_final_newline = true
+trim_trailing_whitespace = true
+
+[*.py]
+indent_style = space
+indent_size = 4
+
+[**.yml]
+indent_style = space
+indent_size = 2
diff --git a/.flake8 b/.flake8
new file mode 100644
index 0000000000000000000000000000000000000000..fd69af775696d49c5397839d5bd6320fe6a73f4b
--- /dev/null
+++ b/.flake8
@@ -0,0 +1,5 @@
+[flake8]
+max-line-length = 88
+
+select = C,E,F,W,B,B950
+extend-ignore = E203, E501, W503
diff --git a/.gitattributes b/.gitattributes
index 28df5f900b358436f0267334b3e3e9af33f917ba..66e2f361826a38d642508b386a31909cbf17042c 100644
--- a/.gitattributes
+++ b/.gitattributes
@@ -53,3 +53,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
*.jpg filter=lfs diff=lfs merge=lfs -text
*.jpeg filter=lfs diff=lfs merge=lfs -text
*.webp filter=lfs diff=lfs merge=lfs -text
+wandb/run-20240512_103207-vfqtzg1a/logs/debug-internal.log filter=lfs diff=lfs merge=lfs -text
+wandb/run-20240512_103207-vfqtzg1a/run-vfqtzg1a.wandb filter=lfs diff=lfs merge=lfs -text
+wandb/run-20240515_132302-ky6h7dv9/run-ky6h7dv9.wandb filter=lfs diff=lfs merge=lfs -text
diff --git a/.isort.cfg b/.isort.cfg
new file mode 100644
index 0000000000000000000000000000000000000000..79067a7c91c9364b6a26abeaa8420725f25050c4
--- /dev/null
+++ b/.isort.cfg
@@ -0,0 +1,3 @@
+[settings]
+profile=black
+known_third_party=wandb
diff --git a/.mypy.ini b/.mypy.ini
new file mode 100644
index 0000000000000000000000000000000000000000..ede9fef8879415168f4207eea47aed4d668d1f44
--- /dev/null
+++ b/.mypy.ini
@@ -0,0 +1,51 @@
+[mypy]
+plugins = pydantic.mypy
+exclude = venv
+
+[mypy-alpaca_lora_4bit.*]
+ignore_missing_imports = True
+
+[mypy-axolotl.monkeypatch.*]
+ignore_errors = True
+
+[mypy-axolotl.models.mixtral.*]
+ignore_errors = True
+
+[mypy-axolotl.models.phi.*]
+ignore_errors = True
+
+[mypy-flash_attn.*]
+ignore_missing_imports = True
+
+[mypy-huggingface_hub]
+ignore_missing_imports = True
+
+[mypy-transformers.*]
+ignore_missing_imports = True
+
+[mypy-peft]
+ignore_missing_imports = True
+
+[mypy-wandb]
+ignore_missing_imports = True
+
+[mypy-bitsandbytes]
+ignore_missing_imports = True
+
+[mypy-requests]
+ignore_missing_imports = True
+
+[mypy-datasets]
+ignore_missing_imports = True
+
+[mypy-fire]
+ignore_missing_imports = True
+
+[mypy-setuptools]
+ignore_missing_imports = True
+
+[mypy-addict]
+ignore_missing_imports = True
+
+[mypy-xformers.*]
+ignore_missing_imports = True
diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..6c5f205897ed234667a47afe87f74b364cf119a3
--- /dev/null
+++ b/.pre-commit-config.yaml
@@ -0,0 +1,43 @@
+default_language_version:
+ python: python3
+
+repos:
+- repo: https://github.com/pre-commit/pre-commit-hooks
+ rev: v4.4.0
+ hooks:
+ - id: check-yaml
+ - id: end-of-file-fixer
+ - id: trailing-whitespace
+- repo: https://github.com/psf/black
+ rev: 23.3.0
+ hooks:
+ - id: black
+- repo: https://github.com/pycqa/isort
+ rev: 5.12.0
+ hooks:
+ - id: isort
+- repo: https://github.com/PyCQA/flake8
+ rev: 6.0.0
+ hooks:
+ - id: flake8
+- repo: https://github.com/PyCQA/pylint
+ rev: v2.17.4
+ hooks:
+ - id: pylint
+- repo: https://github.com/pre-commit/mirrors-mypy
+ rev: v1.3.0
+ hooks:
+ - id: mypy
+ additional_dependencies:
+ [
+ 'types-PyYAML',
+ 'pydantic>=2.5.3',
+ ]
+- repo: https://github.com/PyCQA/bandit
+ rev: 1.7.5
+ hooks:
+ - id: bandit
+ args: [
+ '--ini',
+ '.bandit',
+ ]
diff --git a/.pylintrc b/.pylintrc
new file mode 100644
index 0000000000000000000000000000000000000000..ed973d2859b1faa54ec53706a2e4ac15b70556a4
--- /dev/null
+++ b/.pylintrc
@@ -0,0 +1,14 @@
+[MASTER]
+init-hook="from pylint.config import find_pylintrc; import os, sys; sys.path.append(os.path.dirname(find_pylintrc()))"
+
+[TYPECHECK]
+
+# List of members which are set dynamically and missed by Pylint inference
+# system, and so shouldn't trigger E1101 when accessed.
+generated-members=numpy.*, torch.*
+
+
+[pylint.messages_control]
+disable=missing-function-docstring, line-too-long, import-error,
+ too-many-arguments, too-many-locals, too-many-statements, too-many-branches, too-few-public-methods,
+ too-many-instance-attributes, fixme, import-outside-toplevel, logging-fstring-interpolation,
diff --git a/.vscode/README.md b/.vscode/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..b98a3574f4ae089aeef65d324d43ca8a1b8a8c65
--- /dev/null
+++ b/.vscode/README.md
@@ -0,0 +1 @@
+See [docs/debugging.md](../docs/debugging.md) for guidance on how to modify these files to debug axolotl with VSCode.
diff --git a/.vscode/launch.json b/.vscode/launch.json
new file mode 100644
index 0000000000000000000000000000000000000000..ec1914dc8c4f0199b7a6f94dc83a556ef973ab11
--- /dev/null
+++ b/.vscode/launch.json
@@ -0,0 +1,34 @@
+{
+ // Use IntelliSense to learn about possible attributes.
+ // Hover to view descriptions of existing attributes.
+ // For more information, visit: https://go.microsoft.com/fwlink/?linkid=830387
+ "version": "0.2.0",
+ "configurations": [
+ {
+ "name": "Debug axolotl prompt - sharegpt",
+ "type": "python",
+ "module": "accelerate.commands.launch",
+ "request": "launch",
+ "args": [
+ "-m", "axolotl.cli.train", "dev_sharegpt.yml",
+ // The flags below simplify debugging by overriding the axolotl config
+ // with the debugging tips above. Modify as needed.
+ "--dataset_processes=1", // limits data preprocessing to one process
+ "--max_steps=1", // limits training to just one step
+ "--batch_size=1", // minimizes batch size
+ "--micro_batch_size=1", // minimizes batch size
+ "--val_set_size=0", // disables validation
+ "--sample_packing=False", // disables sample packing which is necessary for small datasets
+ "--eval_sample_packing=False",// disables sample packing on eval set
+ "--dataset_prepared_path=temp_debug/axolotl_outputs/data", // send data outputs to a temp folder
+ "--output_dir=temp_debug/axolotl_outputs/model" // send model outputs to a temp folder
+ ],
+ "console": "integratedTerminal", // show output in the integrated terminal
+ "cwd": "${workspaceFolder}/devtools", // set working directory to devtools from the root of the project
+ "justMyCode": true, // step through only axolotl code
+ "env": {"CUDA_VISIBLE_DEVICES": "0", // Since we aren't doing distributed training, we need to limit to one GPU
+ "HF_HOME": "${workspaceFolder}/devtools/temp_debug/.hf-cache"}, // send HF cache to a temp folder
+ "preLaunchTask": "cleanup-for-dataprep", // delete temp folders (see below)
+ }
+ ]
+}
diff --git a/.vscode/tasks.json b/.vscode/tasks.json
new file mode 100644
index 0000000000000000000000000000000000000000..be9356f9a52a1e1f3b515ce48b3bb0a140e1a245
--- /dev/null
+++ b/.vscode/tasks.json
@@ -0,0 +1,27 @@
+//this file is used by launch.json
+{
+ "version": "2.0.0",
+ "tasks": [
+ // this task changes into the devtools directory and deletes the temp_debug/axolotl_outputs folder
+ {
+ "label": "delete-outputs",
+ "type": "shell",
+ "command": "rm -rf temp_debug/axolotl_outputs",
+ "options":{ "cwd": "${workspaceFolder}/devtools"},
+ "problemMatcher": []
+ },
+ // this task changes into the devtools directory and deletes the `temp_debug/.hf-cache/datasets` folder
+ {
+ "label": "delete-temp-hf-dataset-cache",
+ "type": "shell",
+ "command": "rm -rf temp_debug/.hf-cache/datasets",
+ "options":{ "cwd": "${workspaceFolder}/devtools"},
+ "problemMatcher": []
+ },
+ // this task combines the two tasks above
+ {
+ "label": "cleanup-for-dataprep",
+ "dependsOn": ["delete-outputs", "delete-temp-hf-dataset-cache"],
+ }
+ ]
+}
diff --git a/Config.yml b/Config.yml
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/FAQS.md b/FAQS.md
new file mode 100644
index 0000000000000000000000000000000000000000..f3c9dd52536aa7e9579f8058961b437e80cc8162
--- /dev/null
+++ b/FAQS.md
@@ -0,0 +1,7 @@
+# FAQs
+
+- Can you train StableLM with this? Yes, but only with a single GPU atm. Multi GPU support is coming soon! Just waiting on this [PR](https://github.com/huggingface/transformers/pull/22874)
+- Will this work with Deepspeed? That's still a WIP, but setting `export ACCELERATE_USE_DEEPSPEED=true` should work in some cases
+- `Error invalid argument at line 359 in file /workspace/bitsandbytes/csrc/pythonInterface.c`
+`/arrow/cpp/src/arrow/filesystem/s3fs.cc:2598: arrow::fs::FinalizeS3 was not called even though S3 was initialized.`
+This could lead to a segmentation fault at exit. Try reinstalling bitsandbytes and transformers from source.
diff --git a/LICENSE b/LICENSE
new file mode 100644
index 0000000000000000000000000000000000000000..d645695673349e3947e8e5ae42332d0ac3164cd7
--- /dev/null
+++ b/LICENSE
@@ -0,0 +1,202 @@
+
+ Apache License
+ Version 2.0, January 2004
+ http://www.apache.org/licenses/
+
+ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
+
+ 1. Definitions.
+
+ "License" shall mean the terms and conditions for use, reproduction,
+ and distribution as defined by Sections 1 through 9 of this document.
+
+ "Licensor" shall mean the copyright owner or entity authorized by
+ the copyright owner that is granting the License.
+
+ "Legal Entity" shall mean the union of the acting entity and all
+ other entities that control, are controlled by, or are under common
+ control with that entity. For the purposes of this definition,
+ "control" means (i) the power, direct or indirect, to cause the
+ direction or management of such entity, whether by contract or
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
+ outstanding shares, or (iii) beneficial ownership of such entity.
+
+ "You" (or "Your") shall mean an individual or Legal Entity
+ exercising permissions granted by this License.
+
+ "Source" form shall mean the preferred form for making modifications,
+ including but not limited to software source code, documentation
+ source, and configuration files.
+
+ "Object" form shall mean any form resulting from mechanical
+ transformation or translation of a Source form, including but
+ not limited to compiled object code, generated documentation,
+ and conversions to other media types.
+
+ "Work" shall mean the work of authorship, whether in Source or
+ Object form, made available under the License, as indicated by a
+ copyright notice that is included in or attached to the work
+ (an example is provided in the Appendix below).
+
+ "Derivative Works" shall mean any work, whether in Source or Object
+ form, that is based on (or derived from) the Work and for which the
+ editorial revisions, annotations, elaborations, or other modifications
+ represent, as a whole, an original work of authorship. For the purposes
+ of this License, Derivative Works shall not include works that remain
+ separable from, or merely link (or bind by name) to the interfaces of,
+ the Work and Derivative Works thereof.
+
+ "Contribution" shall mean any work of authorship, including
+ the original version of the Work and any modifications or additions
+ to that Work or Derivative Works thereof, that is intentionally
+ submitted to Licensor for inclusion in the Work by the copyright owner
+ or by an individual or Legal Entity authorized to submit on behalf of
+ the copyright owner. For the purposes of this definition, "submitted"
+ means any form of electronic, verbal, or written communication sent
+ to the Licensor or its representatives, including but not limited to
+ communication on electronic mailing lists, source code control systems,
+ and issue tracking systems that are managed by, or on behalf of, the
+ Licensor for the purpose of discussing and improving the Work, but
+ excluding communication that is conspicuously marked or otherwise
+ designated in writing by the copyright owner as "Not a Contribution."
+
+ "Contributor" shall mean Licensor and any individual or Legal Entity
+ on behalf of whom a Contribution has been received by Licensor and
+ subsequently incorporated within the Work.
+
+ 2. Grant of Copyright License. Subject to the terms and conditions of
+ this License, each Contributor hereby grants to You a perpetual,
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
+ copyright license to reproduce, prepare Derivative Works of,
+ publicly display, publicly perform, sublicense, and distribute the
+ Work and such Derivative Works in Source or Object form.
+
+ 3. Grant of Patent License. Subject to the terms and conditions of
+ this License, each Contributor hereby grants to You a perpetual,
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
+ (except as stated in this section) patent license to make, have made,
+ use, offer to sell, sell, import, and otherwise transfer the Work,
+ where such license applies only to those patent claims licensable
+ by such Contributor that are necessarily infringed by their
+ Contribution(s) alone or by combination of their Contribution(s)
+ with the Work to which such Contribution(s) was submitted. If You
+ institute patent litigation against any entity (including a
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
+ or a Contribution incorporated within the Work constitutes direct
+ or contributory patent infringement, then any patent licenses
+ granted to You under this License for that Work shall terminate
+ as of the date such litigation is filed.
+
+ 4. Redistribution. You may reproduce and distribute copies of the
+ Work or Derivative Works thereof in any medium, with or without
+ modifications, and in Source or Object form, provided that You
+ meet the following conditions:
+
+ (a) You must give any other recipients of the Work or
+ Derivative Works a copy of this License; and
+
+ (b) You must cause any modified files to carry prominent notices
+ stating that You changed the files; and
+
+ (c) You must retain, in the Source form of any Derivative Works
+ that You distribute, all copyright, patent, trademark, and
+ attribution notices from the Source form of the Work,
+ excluding those notices that do not pertain to any part of
+ the Derivative Works; and
+
+ (d) If the Work includes a "NOTICE" text file as part of its
+ distribution, then any Derivative Works that You distribute must
+ include a readable copy of the attribution notices contained
+ within such NOTICE file, excluding those notices that do not
+ pertain to any part of the Derivative Works, in at least one
+ of the following places: within a NOTICE text file distributed
+ as part of the Derivative Works; within the Source form or
+ documentation, if provided along with the Derivative Works; or,
+ within a display generated by the Derivative Works, if and
+ wherever such third-party notices normally appear. The contents
+ of the NOTICE file are for informational purposes only and
+ do not modify the License. You may add Your own attribution
+ notices within Derivative Works that You distribute, alongside
+ or as an addendum to the NOTICE text from the Work, provided
+ that such additional attribution notices cannot be construed
+ as modifying the License.
+
+ You may add Your own copyright statement to Your modifications and
+ may provide additional or different license terms and conditions
+ for use, reproduction, or distribution of Your modifications, or
+ for any such Derivative Works as a whole, provided Your use,
+ reproduction, and distribution of the Work otherwise complies with
+ the conditions stated in this License.
+
+ 5. Submission of Contributions. Unless You explicitly state otherwise,
+ any Contribution intentionally submitted for inclusion in the Work
+ by You to the Licensor shall be under the terms and conditions of
+ this License, without any additional terms or conditions.
+ Notwithstanding the above, nothing herein shall supersede or modify
+ the terms of any separate license agreement you may have executed
+ with Licensor regarding such Contributions.
+
+ 6. Trademarks. This License does not grant permission to use the trade
+ names, trademarks, service marks, or product names of the Licensor,
+ except as required for reasonable and customary use in describing the
+ origin of the Work and reproducing the content of the NOTICE file.
+
+ 7. Disclaimer of Warranty. Unless required by applicable law or
+ agreed to in writing, Licensor provides the Work (and each
+ Contributor provides its Contributions) on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
+ implied, including, without limitation, any warranties or conditions
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
+ PARTICULAR PURPOSE. You are solely responsible for determining the
+ appropriateness of using or redistributing the Work and assume any
+ risks associated with Your exercise of permissions under this License.
+
+ 8. Limitation of Liability. In no event and under no legal theory,
+ whether in tort (including negligence), contract, or otherwise,
+ unless required by applicable law (such as deliberate and grossly
+ negligent acts) or agreed to in writing, shall any Contributor be
+ liable to You for damages, including any direct, indirect, special,
+ incidental, or consequential damages of any character arising as a
+ result of this License or out of the use or inability to use the
+ Work (including but not limited to damages for loss of goodwill,
+ work stoppage, computer failure or malfunction, or any and all
+ other commercial damages or losses), even if such Contributor
+ has been advised of the possibility of such damages.
+
+ 9. Accepting Warranty or Additional Liability. While redistributing
+ the Work or Derivative Works thereof, You may choose to offer,
+ and charge a fee for, acceptance of support, warranty, indemnity,
+ or other liability obligations and/or rights consistent with this
+ License. However, in accepting such obligations, You may act only
+ on Your own behalf and on Your sole responsibility, not on behalf
+ of any other Contributor, and only if You agree to indemnify,
+ defend, and hold each Contributor harmless for any liability
+ incurred by, or claims asserted against, such Contributor by reason
+ of your accepting any such warranty or additional liability.
+
+ END OF TERMS AND CONDITIONS
+
+ APPENDIX: How to apply the Apache License to your work.
+
+ To apply the Apache License to your work, attach the following
+ boilerplate notice, with the fields enclosed by brackets "[]"
+ replaced with your own identifying information. (Don't include
+ the brackets!) The text should be enclosed in the appropriate
+ comment syntax for the file format. We also recommend that a
+ file or class name and description of purpose be included on the
+ same "printed page" as the copyright notice for easier
+ identification within third-party archives.
+
+ Copyright [yyyy] [name of copyright owner]
+
+ Licensed under the Apache License, Version 2.0 (the "License");
+ you may not use this file except in compliance with the License.
+ You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
diff --git a/README.md b/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..23ac443eedcd3f68609ee2db2f8cab7dbb2d7642
--- /dev/null
+++ b/README.md
@@ -0,0 +1,702 @@
+# Axolotl
+
+Axolotl is a tool designed to streamline the fine-tuning of various AI models, offering support for multiple configurations and architectures.
+
+Features:
+- Train various Huggingface models such as llama, pythia, falcon, mpt
+- Supports fullfinetune, lora, qlora, relora, and gptq
+- Customize configurations using a simple yaml file or CLI overwrite
+- Load different dataset formats, use custom formats, or bring your own tokenized datasets
+- Integrated with xformer, flash attention, rope scaling, and multipacking
+- Works with single GPU or multiple GPUs via FSDP or Deepspeed
+- Easily run with Docker locally or on the cloud
+- Log results and optionally checkpoints to wandb or mlflow
+- And more!
+
+
+
+
+
+
+
+
+
+## Table of Contents
+- [Introduction](#axolotl)
+- [Supported Features](#axolotl-supports)
+- [Quickstart](#quickstart-)
+- [Environment](#environment)
+ - [Docker](#docker)
+ - [Conda/Pip venv](#condapip-venv)
+ - [Cloud GPU](#cloud-gpu) - Latitude.sh, JarvisLabs, RunPod
+ - [Bare Metal Cloud GPU](#bare-metal-cloud-gpu)
+ - [Windows](#windows)
+ - [Mac](#mac)
+ - [Google Colab](#google-colab)
+ - [Launching on public clouds via SkyPilot](#launching-on-public-clouds-via-skypilot)
+ - [Launching on public clouds via dstack](#launching-on-public-clouds-via-dstack)
+- [Dataset](#dataset)
+- [Config](#config)
+ - [Train](#train)
+ - [Inference](#inference-playground)
+ - [Merge LORA to Base](#merge-lora-to-base)
+ - [Special Tokens](#special-tokens)
+ - [All Config Options](#all-config-options)
+- Advanced Topics
+ - [Multipack](./docs/multipack.qmd)
+ - [RLHF & DPO](./docs/rlhf.qmd)
+ - [Dataset Pre-Processing](./docs/dataset_preprocessing.qmd)
+- [Common Errors](#common-errors-)
+ - [Tokenization Mismatch b/w Training & Inference](#tokenization-mismatch-bw-inference--training)
+- [Debugging Axolotl](#debugging-axolotl)
+- [Need Help?](#need-help-)
+- [Badge](#badge-)
+- [Community Showcase](#community-showcase)
+- [Contributing](#contributing-)
+- [Sponsors](#sponsors-)
+
+
+
+
+
+
+
+
+ Axolotl provides a unified repository for fine-tuning a variety of AI models with ease
+
+
+ Go ahead and Axolotl questions!!
+
+
+
+
+
+
+
+
+
+
+## Axolotl supports
+
+| | fp16/fp32 | lora | qlora | gptq | gptq w/flash attn | flash attn | xformers attn |
+|-------------|:----------|:-----|-------|------|-------------------|------------|--------------|
+| llama | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
+| Mistral | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
+| Mixtral-MoE | ✅ | ✅ | ✅ | ❓ | ❓ | ❓ | ❓ |
+| Mixtral8X22 | ✅ | ✅ | ✅ | ❓ | ❓ | ❓ | ❓ |
+| Pythia | ✅ | ✅ | ✅ | ❌ | ❌ | ❌ | ❓ |
+| cerebras | ✅ | ✅ | ✅ | ❌ | ❌ | ❌ | ❓ |
+| btlm | ✅ | ✅ | ✅ | ❌ | ❌ | ❌ | ❓ |
+| mpt | ✅ | ❌ | ❓ | ❌ | ❌ | ❌ | ❓ |
+| falcon | ✅ | ✅ | ✅ | ❌ | ❌ | ❌ | ❓ |
+| gpt-j | ✅ | ✅ | ✅ | ❌ | ❌ | ❓ | ❓ |
+| XGen | ✅ | ❓ | ✅ | ❓ | ❓ | ❓ | ✅ |
+| phi | ✅ | ✅ | ✅ | ❓ | ❓ | ❓ | ❓ |
+| RWKV | ✅ | ❓ | ❓ | ❓ | ❓ | ❓ | ❓ |
+| Qwen | ✅ | ✅ | ✅ | ❓ | ❓ | ❓ | ❓ |
+| Gemma | ✅ | ✅ | ✅ | ❓ | ❓ | ✅ | ❓ |
+
+✅: supported
+❌: not supported
+❓: untested
+
+## Quickstart ⚡
+
+Get started with Axolotl in just a few steps! This quickstart guide will walk you through setting up and running a basic fine-tuning task.
+
+**Requirements**: Python >=3.10 and Pytorch >=2.1.1.
+
+```bash
+git clone https://github.com/OpenAccess-AI-Collective/axolotl
+cd axolotl
+
+pip3 install packaging ninja
+pip3 install -e '.[flash-attn,deepspeed]'
+```
+
+### Usage
+```bash
+# preprocess datasets - optional but recommended
+CUDA_VISIBLE_DEVICES="" python -m axolotl.cli.preprocess examples/openllama-3b/lora.yml
+
+# finetune lora
+accelerate launch -m axolotl.cli.train examples/openllama-3b/lora.yml
+
+# inference
+accelerate launch -m axolotl.cli.inference examples/openllama-3b/lora.yml \
+ --lora_model_dir="./lora-out"
+
+# gradio
+accelerate launch -m axolotl.cli.inference examples/openllama-3b/lora.yml \
+ --lora_model_dir="./lora-out" --gradio
+
+# remote yaml files - the yaml config can be hosted on a public URL
+# Note: the yaml config must directly link to the **raw** yaml
+accelerate launch -m axolotl.cli.train https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/examples/openllama-3b/lora.yml
+```
+
+## Advanced Setup
+
+### Environment
+
+#### Docker
+
+ ```bash
+ docker run --gpus '"all"' --rm -it winglian/axolotl:main-latest
+ ```
+
+ Or run on the current files for development:
+
+ ```sh
+ docker compose up -d
+ ```
+
+>[!Tip]
+> If you want to debug axolotl or prefer to use Docker as your development environment, see the [debugging guide's section on Docker](docs/debugging.qmd#debugging-with-docker).
+
+
+
+ Docker advanced
+
+ A more powerful Docker command to run would be this:
+
+ ```bash
+docker run --privileged --gpus '"all"' --shm-size 10g --rm -it --name axolotl --ipc=host --ulimit memlock=-1 --ulimit stack=67108864 --mount type=bind,src="${PWD}",target=/workspace/axolotl -v ${HOME}/.cache/huggingface:/root/.cache/huggingface winglian/axolotl:main-latest
+ ```
+
+ It additionally:
+ * Prevents memory issues when running e.g. deepspeed (e.g. you could hit SIGBUS/signal 7 error) through `--ipc` and `--ulimit` args.
+ * Persists the downloaded HF data (models etc.) and your modifications to axolotl code through `--mount`/`-v` args.
+ * The `--name` argument simply makes it easier to refer to the container in vscode (`Dev Containers: Attach to Running Container...`) or in your terminal.
+ * The `--privileged` flag gives all capabilities to the container.
+ * The `--shm-size 10g` argument increases the shared memory size. Use this if you see `exitcode: -7` errors using deepspeed.
+
+ [More information on nvidia website](https://docs.nvidia.com/deeplearning/frameworks/user-guide/index.html#setincshmem)
+
+
+
+#### Conda/Pip venv
+ 1. Install python >=**3.10**
+
+ 2. Install pytorch stable https://pytorch.org/get-started/locally/
+
+ 3. Install Axolotl along with python dependencies
+ ```bash
+ pip3 install packaging
+ pip3 install -e '.[flash-attn,deepspeed]'
+ ```
+ 4. (Optional) Login to Huggingface to use gated models/datasets.
+ ```bash
+ huggingface-cli login
+ ```
+ Get the token at huggingface.co/settings/tokens
+
+#### Cloud GPU
+
+For cloud GPU providers that support docker images, use [`winglian/axolotl-cloud:main-latest`](https://hub.docker.com/r/winglian/axolotl-cloud/tags)
+
+- on Latitude.sh use this [direct link](https://latitude.sh/blueprint/989e0e79-3bf6-41ea-a46b-1f246e309d5c)
+- on JarvisLabs.ai use this [direct link](https://jarvislabs.ai/templates/axolotl)
+- on RunPod use this [direct link](https://runpod.io/gsc?template=v2ickqhz9s&ref=6i7fkpdz)
+
+#### Bare Metal Cloud GPU
+
+##### LambdaLabs
+
+
+
+ Click to Expand
+
+ 1. Install python
+ ```bash
+ sudo apt update
+ sudo apt install -y python3.10
+
+ sudo update-alternatives --install /usr/bin/python python /usr/bin/python3.10 1
+ sudo update-alternatives --config python # pick 3.10 if given option
+ python -V # should be 3.10
+
+ ```
+
+ 2. Install pip
+ ```bash
+ wget https://bootstrap.pypa.io/get-pip.py
+ python get-pip.py
+ ```
+
+ 3. Install Pytorch https://pytorch.org/get-started/locally/
+
+ 4. Follow instructions on quickstart.
+
+ 5. Run
+ ```bash
+ pip3 install protobuf==3.20.3
+ pip3 install -U --ignore-installed requests Pillow psutil scipy
+ ```
+
+ 6. Set path
+ ```bash
+ export LD_LIBRARY_PATH=/usr/lib/x86_64-linux-gnu:$LD_LIBRARY_PATH
+ ```
+
+
+##### GCP
+
+
+
+Click to Expand
+
+Use a Deeplearning linux OS with cuda and pytorch installed. Then follow instructions on quickstart.
+
+Make sure to run the below to uninstall xla.
+```bash
+pip uninstall -y torch_xla[tpu]
+```
+
+
+
+#### Windows
+Please use WSL or Docker!
+
+#### Mac
+
+Use the below instead of the install method in QuickStart.
+```
+pip3 install -e '.'
+```
+More info: [mac.md](/docs/mac.qmd)
+
+#### Google Colab
+
+Please use this example [notebook](examples/colab-notebooks/colab-axolotl-example.ipynb).
+
+#### Launching on public clouds via SkyPilot
+To launch on GPU instances (both on-demand and spot instances) on 7+ clouds (GCP, AWS, Azure, OCI, and more), you can use [SkyPilot](https://skypilot.readthedocs.io/en/latest/index.html):
+
+```bash
+pip install "skypilot-nightly[gcp,aws,azure,oci,lambda,kubernetes,ibm,scp]" # choose your clouds
+sky check
+```
+
+Get the [example YAMLs](https://github.com/skypilot-org/skypilot/tree/master/llm/axolotl) of using Axolotl to finetune `mistralai/Mistral-7B-v0.1`:
+```
+git clone https://github.com/skypilot-org/skypilot.git
+cd skypilot/llm/axolotl
+```
+
+Use one command to launch:
+```bash
+# On-demand
+HF_TOKEN=xx sky launch axolotl.yaml --env HF_TOKEN
+
+# Managed spot (auto-recovery on preemption)
+HF_TOKEN=xx BUCKET= sky spot launch axolotl-spot.yaml --env HF_TOKEN --env BUCKET
+```
+
+#### Launching on public clouds via dstack
+To launch on GPU instance (both on-demand and spot instances) on public clouds (GCP, AWS, Azure, Lambda Labs, TensorDock, Vast.ai, and CUDO), you can use [dstack](https://dstack.ai/).
+
+Write a job description in YAML as below:
+
+```yaml
+# dstack.yaml
+type: task
+
+image: winglian/axolotl-cloud:main-20240429-py3.11-cu121-2.2.2
+
+env:
+ - HUGGING_FACE_HUB_TOKEN
+ - WANDB_API_KEY
+
+commands:
+ - accelerate launch -m axolotl.cli.train config.yaml
+
+ports:
+ - 6006
+
+resources:
+ gpu:
+ memory: 24GB..
+ count: 2
+```
+
+then, simply run the job with `dstack run` command. Append `--spot` option if you want spot instance. `dstack run` command will show you the instance with cheapest price across multi cloud services:
+
+```bash
+pip install dstack
+HUGGING_FACE_HUB_TOKEN=xxx WANDB_API_KEY=xxx dstack run . -f dstack.yaml # --spot
+```
+
+For further and fine-grained use cases, please refer to the official [dstack documents](https://dstack.ai/docs/) and the detailed description of [axolotl example](https://github.com/dstackai/dstack/tree/master/examples/fine-tuning/axolotl) on the official repository.
+
+### Dataset
+
+Axolotl supports a variety of dataset formats. It is recommended to use a JSONL. The schema of the JSONL depends upon the task and the prompt template you wish to use. Instead of a JSONL, you can also use a HuggingFace dataset with columns for each JSONL field.
+
+See [these docs](https://openaccess-ai-collective.github.io/axolotl/docs/dataset-formats/) for more information on how to use different dataset formats.
+
+### Config
+
+See [examples](examples) for quick start. It is recommended to duplicate and modify to your needs. The most important options are:
+
+- model
+ ```yaml
+ base_model: ./llama-7b-hf # local or huggingface repo
+ ```
+ Note: The code will load the right architecture.
+
+- dataset
+ ```yaml
+ datasets:
+ # huggingface repo
+ - path: vicgalle/alpaca-gpt4
+ type: alpaca
+
+ # huggingface repo with specific configuration/subset
+ - path: EleutherAI/pile
+ name: enron_emails
+ type: completion # format from earlier
+ field: text # Optional[str] default: text, field to use for completion data
+
+ # huggingface repo with multiple named configurations/subsets
+ - path: bigcode/commitpackft
+ name:
+ - ruby
+ - python
+ - typescript
+ type: ... # unimplemented custom format
+
+ # fastchat conversation
+ # See 'conversation' options: https://github.com/lm-sys/FastChat/blob/main/fastchat/conversation.py
+ - path: ...
+ type: sharegpt
+ conversation: chatml # default: vicuna_v1.1
+
+ # local
+ - path: data.jsonl # or json
+ ds_type: json # see other options below
+ type: alpaca
+
+ # dataset with splits, but no train split
+ - path: knowrohit07/know_sql
+ type: context_qa.load_v2
+ train_on_split: validation
+
+ # loading from s3 or gcs
+ # s3 creds will be loaded from the system default and gcs only supports public access
+ - path: s3://path_to_ds # Accepts folder with arrow/parquet or file path like above. Supports s3, gcs.
+ ...
+
+ # Loading Data From a Public URL
+ # - The file format is `json` (which includes `jsonl`) by default. For different formats, adjust the `ds_type` option accordingly.
+ - path: https://some.url.com/yourdata.jsonl # The URL should be a direct link to the file you wish to load. URLs must use HTTPS protocol, not HTTP.
+ ds_type: json # this is the default, see other options below.
+ ```
+
+- loading
+ ```yaml
+ load_in_4bit: true
+ load_in_8bit: true
+
+ bf16: auto # require >=ampere, auto will detect if your GPU supports this and choose automatically.
+ fp16: # leave empty to use fp16 when bf16 is 'auto'. set to false if you want to fallback to fp32
+ tf32: true # require >=ampere
+
+ bfloat16: true # require >=ampere, use instead of bf16 when you don't want AMP (automatic mixed precision)
+ float16: true # use instead of fp16 when you don't want AMP
+ ```
+ Note: Repo does not do 4-bit quantization.
+
+- lora
+ ```yaml
+ adapter: lora # 'qlora' or leave blank for full finetune
+ lora_r: 8
+ lora_alpha: 16
+ lora_dropout: 0.05
+ lora_target_modules:
+ - q_proj
+ - v_proj
+ ```
+
+#### All Config Options
+
+See [these docs](docs/config.qmd) for all config options.
+
+### Train
+
+Run
+```bash
+accelerate launch -m axolotl.cli.train your_config.yml
+```
+
+> [!TIP]
+> You can also reference a config file that is hosted on a public URL, for example `accelerate launch -m axolotl.cli.train https://yourdomain.com/your_config.yml`
+
+#### Preprocess dataset
+
+You can optionally pre-tokenize dataset with the following before finetuning.
+This is recommended for large datasets.
+
+- Set `dataset_prepared_path:` to a local folder for saving and loading pre-tokenized dataset.
+- (Optional): Set `push_dataset_to_hub: hf_user/repo` to push it to Huggingface.
+- (Optional): Use `--debug` to see preprocessed examples.
+
+```bash
+python -m axolotl.cli.preprocess your_config.yml
+```
+
+#### Multi-GPU
+
+Below are the options available in axolotl for training with multiple GPUs. Note that DeepSpeed
+is the recommended multi-GPU option currently because FSDP may experience
+[loss instability](https://github.com/huggingface/transformers/issues/26498).
+
+##### DeepSpeed
+
+Deepspeed is an optimization suite for multi-gpu systems allowing you to train much larger models than you
+might typically be able to fit into your GPU's VRAM. More information about the various optimization types
+for deepspeed is available at https://huggingface.co/docs/accelerate/main/en/usage_guides/deepspeed#what-is-integrated
+
+We provide several default deepspeed JSON configurations for ZeRO stage 1, 2, and 3.
+
+```yaml
+deepspeed: deepspeed_configs/zero1.json
+```
+
+```shell
+accelerate launch -m axolotl.cli.train examples/llama-2/config.yml --deepspeed deepspeed_configs/zero1.json
+```
+
+##### FSDP
+
+- llama FSDP
+```yaml
+fsdp:
+ - full_shard
+ - auto_wrap
+fsdp_config:
+ fsdp_offload_params: true
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
+```
+
+##### FSDP + QLoRA
+
+Axolotl supports training with FSDP and QLoRA, see [these docs](docs/fsdp_qlora.qmd) for more information.
+
+##### Weights & Biases Logging
+
+Make sure your `WANDB_API_KEY` environment variable is set (recommended) or you login to wandb with `wandb login`.
+
+- wandb options
+```yaml
+wandb_mode:
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+```
+
+##### Special Tokens
+
+It is important to have special tokens like delimiters, end-of-sequence, beginning-of-sequence in your tokenizer's vocabulary. This will help you avoid tokenization issues and help your model train better. You can do this in axolotl like this:
+
+```yml
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
+tokens: # these are delimiters
+ - "<|im_start|>"
+ - "<|im_end|>"
+```
+
+When you include these tokens in your axolotl config, axolotl adds these tokens to the tokenizer's vocabulary.
+
+### Inference Playground
+
+Axolotl allows you to load your model in an interactive terminal playground for quick experimentation.
+The config file is the same config file used for training.
+
+Pass the appropriate flag to the inference command, depending upon what kind of model was trained:
+
+- Pretrained LORA:
+ ```bash
+ python -m axolotl.cli.inference examples/your_config.yml --lora_model_dir="./lora-output-dir"
+ ```
+- Full weights finetune:
+ ```bash
+ python -m axolotl.cli.inference examples/your_config.yml --base_model="./completed-model"
+ ```
+- Full weights finetune w/ a prompt from a text file:
+ ```bash
+ cat /tmp/prompt.txt | python -m axolotl.cli.inference examples/your_config.yml \
+ --base_model="./completed-model" --prompter=None --load_in_8bit=True
+ ```
+-- With gradio hosting
+ ```bash
+ python -m axolotl.cli.inference examples/your_config.yml --gradio
+ ```
+
+Please use `--sample_packing False` if you have it on and receive the error similar to below:
+
+> RuntimeError: stack expects each tensor to be equal size, but got [1, 32, 1, 128] at entry 0 and [1, 32, 8, 128] at entry 1
+
+### Merge LORA to base
+
+The following command will merge your LORA adapater with your base model. You can optionally pass the argument `--lora_model_dir` to specify the directory where your LORA adapter was saved, otherwhise, this will be inferred from `output_dir` in your axolotl config file. The merged model is saved in the sub-directory `{lora_model_dir}/merged`.
+
+```bash
+python3 -m axolotl.cli.merge_lora your_config.yml --lora_model_dir="./completed-model"
+```
+
+You may need to use the `gpu_memory_limit` and/or `lora_on_cpu` config options to avoid running out of memory. If you still run out of CUDA memory, you can try to merge in system RAM with
+
+```bash
+CUDA_VISIBLE_DEVICES="" python3 -m axolotl.cli.merge_lora ...
+```
+
+although this will be very slow, and using the config options above are recommended instead.
+
+## Common Errors 🧰
+
+See also the [FAQ's](./docs/faq.qmd) and [debugging guide](docs/debugging.qmd).
+
+> If you encounter a 'Cuda out of memory' error, it means your GPU ran out of memory during the training process. Here's how to resolve it:
+
+Please reduce any below
+ - `micro_batch_size`
+ - `eval_batch_size`
+ - `gradient_accumulation_steps`
+ - `sequence_len`
+
+If it does not help, try running without deepspeed and without accelerate (replace "accelerate launch" with "python") in the command.
+
+Using adamw_bnb_8bit might also save you some memory.
+
+> `failed (exitcode: -9)`
+
+Usually means your system has run out of system memory.
+Similarly, you should consider reducing the same settings as when you run out of VRAM.
+Additionally, look into upgrading your system RAM which should be simpler than GPU upgrades.
+
+> RuntimeError: expected scalar type Float but found Half
+
+Try set `fp16: true`
+
+> NotImplementedError: No operator found for `memory_efficient_attention_forward` ...
+
+Try to turn off xformers.
+
+> accelerate config missing
+
+It's safe to ignore it.
+
+> NCCL Timeouts during training
+
+See the [NCCL](docs/nccl.qmd) guide.
+
+
+### Tokenization Mismatch b/w Inference & Training
+
+For many formats, Axolotl constructs prompts by concatenating token ids _after_ tokenizing strings. The reason for concatenating token ids rather than operating on strings is to maintain precise accounting for attention masks.
+
+If you decode a prompt constructed by axolotl, you might see spaces between tokens (or lack thereof) that you do not expect, especially around delimiters and special tokens. When you are starting out with a new format, you should always do the following:
+
+1. Materialize some data using `python -m axolotl.cli.preprocess your_config.yml --debug`, and then decode the first few rows with your model's tokenizer.
+2. During inference, right before you pass a tensor of token ids to your model, decode these tokens back into a string.
+3. Make sure the inference string from #2 looks **exactly** like the data you fine tuned on from #1, including spaces and new lines. If they aren't the same, adjust your inference server accordingly.
+4. As an additional troubleshooting step, you can look at the token ids between 1 and 2 to make sure they are identical.
+
+Having misalignment between your prompts during training and inference can cause models to perform very poorly, so it is worth checking this. See [this blog post](https://hamel.dev/notes/llm/05_tokenizer_gotchas.html) for a concrete example.
+
+## Debugging Axolotl
+
+See [this debugging guide](docs/debugging.qmd) for tips on debugging Axolotl, along with an example configuration for debugging with VSCode.
+
+## Need help? 🙋
+
+Join our [Discord server](https://discord.gg/HhrNrHJPRb) where we our community members can help you.
+
+Need dedicated support? Please contact us at [✉️wing@openaccessaicollective.org](mailto:wing@openaccessaicollective.org) for dedicated support options.
+
+## Badge ❤🏷️
+
+Building something cool with Axolotl? Consider adding a badge to your model card.
+
+```markdown
+[ ](https://github.com/OpenAccess-AI-Collective/axolotl)
+```
+
+[ ](https://github.com/OpenAccess-AI-Collective/axolotl)
+
+## Community Showcase
+
+Check out some of the projects and models that have been built using Axolotl! Have a model you'd like to add to our Community Showcase? Open a PR with your model.
+
+Open Access AI Collective
+- [Minotaur 13b](https://huggingface.co/openaccess-ai-collective/minotaur-13b-fixed)
+- [Manticore 13b](https://huggingface.co/openaccess-ai-collective/manticore-13b)
+- [Hippogriff 30b](https://huggingface.co/openaccess-ai-collective/hippogriff-30b-chat)
+
+PocketDoc Labs
+- [Dan's PersonalityEngine 13b LoRA](https://huggingface.co/PocketDoc/Dans-PersonalityEngine-13b-LoRA)
+
+## Contributing 🤝
+
+Please read the [contributing guide](./.github/CONTRIBUTING.md)
+
+Bugs? Please check the [open issues](https://github.com/OpenAccess-AI-Collective/axolotl/issues/bug) else create a new Issue.
+
+PRs are **greatly welcome**!
+
+Please run the quickstart instructions followed by the below to setup env:
+```bash
+pip3 install -r requirements-dev.txt -r requirements-tests.txt
+pre-commit install
+
+# test
+pytest tests/
+
+# optional: run against all files
+pre-commit run --all-files
+```
+
+Thanks to all of our contributors to date. Help drive open source AI progress forward by contributing to Axolotl.
+
+
+
+
+
+## Sponsors 🤝❤
+
+OpenAccess AI Collective is run by volunteer contributors such as [winglian](https://github.com/winglian),
+[NanoCode012](https://github.com/NanoCode012), [tmm1](https://github.com/tmm1),
+[mhenrichsen](https://github.com/mhenrichsen), [casper-hansen](https://github.com/casper-hansen),
+[hamelsmu](https://github.com/hamelsmu) and many more who help us accelerate forward by fixing bugs, answering
+community questions and implementing new features. Axolotl needs donations from sponsors for the compute needed to
+run our unit & integration tests, troubleshooting community issues, and providing bounties. If you love axolotl,
+consider sponsoring the project via [GitHub Sponsors](https://github.com/sponsors/OpenAccess-AI-Collective),
+[Ko-fi](https://ko-fi.com/axolotl_ai) or reach out directly to
+[wing@openaccessaicollective.org](mailto:wing@openaccessaicollective.org).
+
+---
+
+#### 💎 Diamond Sponsors - [Contact directly](mailto:wing@openaccessaicollective.org)
+
+---
+
+#### 🥇 Gold Sponsors - $5000/mo
+
+---
+
+#### 🥈 Silver Sponsors - $1000/mo
+
+---
+
+#### 🥉 Bronze Sponsors - $500/mo
+
+ - [JarvisLabs.ai](https://jarvislabs.ai)
+
+---
diff --git a/TODO.md b/TODO.md
new file mode 100644
index 0000000000000000000000000000000000000000..2002bbbaf1b8781e9f632856c032702151895bb2
--- /dev/null
+++ b/TODO.md
@@ -0,0 +1,10 @@
+# todo list
+
+- [] Validation of parameters for combinations that won't work
+
+
+
+## things that are known not to work
+
+- FSDP offload and gradient_checkpointing - https://github.com/pytorch/pytorch/issues/82203
+- adamw_bnb_8bit doesn't play well with FSDP offload
diff --git a/_quarto.yml b/_quarto.yml
new file mode 100644
index 0000000000000000000000000000000000000000..749f68cce6b5a145c4352c9ff11e79fc9d0f5b7e
--- /dev/null
+++ b/_quarto.yml
@@ -0,0 +1,51 @@
+project:
+ type: website
+
+website:
+ title: "Axolotl"
+ description: "Fine-tuning"
+ favicon: favicon.jpg
+ navbar:
+ title: Axolotl
+ background: dark
+ pinned: false
+ collapse: false
+ tools:
+ - icon: twitter
+ href: https://twitter.com/axolotl_ai
+ - icon: github
+ href: https://github.com/OpenAccess-AI-Collective/axolotl/
+ - icon: discord
+ href: https://discord.gg/7m9sfhzaf3
+
+ sidebar:
+ pinned: true
+ collapse-level: 2
+ style: docked
+ contents:
+ - text: Home
+ href: index.qmd
+ - section: "How-To Guides"
+ contents:
+ # TODO Edit folder structure after we have more docs.
+ - docs/debugging.qmd
+ - docs/multipack.qmd
+ - docs/fsdp_qlora.qmd
+ - docs/input_output.qmd
+ - docs/rlhf.qmd
+ - docs/nccl.qmd
+ - docs/mac.qmd
+ - docs/multi-node.qmd
+ - section: "Dataset Formats"
+ contents: docs/dataset-formats/*
+ - section: "Reference"
+ contents:
+ - docs/config.qmd
+ - docs/faq.qmd
+
+
+format:
+ html:
+ theme: materia
+ css: styles.css
+ toc: true
diff --git a/cicd/Dockerfile.jinja b/cicd/Dockerfile.jinja
new file mode 100644
index 0000000000000000000000000000000000000000..ce03e08c2f6b8955ba200f2fa885e96ea776eb59
--- /dev/null
+++ b/cicd/Dockerfile.jinja
@@ -0,0 +1,40 @@
+FROM winglian/axolotl-base:{{ BASE_TAG }}
+
+ENV TORCH_CUDA_ARCH_LIST="7.0 7.5 8.0 8.6+PTX"
+ENV AXOLOTL_EXTRAS="{{ AXOLOTL_EXTRAS }}"
+ENV AXOLOTL_ARGS="{{ AXOLOTL_ARGS }}"
+ENV CUDA="{{ CUDA }}"
+ENV BNB_CUDA_VERSION="{{ CUDA }}"
+ENV PYTORCH_VERSION="{{ PYTORCH_VERSION }}"
+ENV GITHUB_REF="{{ GITHUB_REF }}"
+ENV GITHUB_SHA="{{ GITHUB_SHA }}"
+
+RUN apt-get update && \
+ apt-get install -y --allow-change-held-packages vim curl nano libnccl2 libnccl-dev
+
+WORKDIR /workspace
+
+RUN git clone --depth=1 https://github.com/OpenAccess-AI-Collective/axolotl.git
+
+WORKDIR /workspace/axolotl
+
+RUN git fetch origin +$GITHUB_REF && \
+ git checkout FETCH_HEAD
+
+# If AXOLOTL_EXTRAS is set, append it in brackets
+RUN pip install causal_conv1d
+RUN if [ "$AXOLOTL_EXTRAS" != "" ] ; then \
+ pip install -e .[deepspeed,flash-attn,mamba-ssm,galore,$AXOLOTL_EXTRAS] $AXOLOTL_ARGS; \
+ else \
+ pip install -e .[deepspeed,flash-attn,mamba-ssm,galore] $AXOLOTL_ARGS; \
+ fi
+
+# So we can test the Docker image
+RUN pip install pytest
+
+# fix so that git fetch/pull from remote works
+RUN git config remote.origin.fetch "+refs/heads/*:refs/remotes/origin/*" && \
+ git config --get remote.origin.fetch
+
+# helper for huggingface-login cli
+RUN git config --global credential.helper store
diff --git a/cicd/cicd.sh b/cicd/cicd.sh
new file mode 100644
index 0000000000000000000000000000000000000000..fa2049b6bd685e4fa59c29e49bf8699f80c4acc3
--- /dev/null
+++ b/cicd/cicd.sh
@@ -0,0 +1,5 @@
+#!/bin/bash
+
+pytest --ignore=tests/e2e/ /workspace/axolotl/tests/
+pytest /workspace/axolotl/tests/e2e/patched/
+pytest --ignore=tests/e2e/patched/ /workspace/axolotl/tests/e2e/
diff --git a/cicd/tests.py b/cicd/tests.py
new file mode 100644
index 0000000000000000000000000000000000000000..bfbdb7b90acbe203dfdf415f29df794996f00eff
--- /dev/null
+++ b/cicd/tests.py
@@ -0,0 +1,75 @@
+"""
+ modal application to run axolotl gpu tests in Modal
+ """
+import os
+import pathlib
+import tempfile
+
+import jinja2
+import modal
+from jinja2 import select_autoescape
+from modal import Image, Stub
+
+cicd_path = pathlib.Path(__file__).parent.resolve()
+
+template_loader = jinja2.FileSystemLoader(searchpath=cicd_path)
+template_env = jinja2.Environment(
+ loader=template_loader, autoescape=select_autoescape()
+)
+df_template = template_env.get_template("Dockerfile.jinja")
+
+df_args = {
+ "AXOLOTL_EXTRAS": os.environ.get("AXOLOTL_EXTRAS", ""),
+ "AXOLOTL_ARGS": os.environ.get("AXOLOTL_ARGS", ""),
+ "PYTORCH_VERSION": os.environ.get("PYTORCH_VERSION", "2.0.1"),
+ "BASE_TAG": os.environ.get("BASE_TAG", "main-base-py3.10-cu118-2.0.1"),
+ "CUDA": os.environ.get("CUDA", "118"),
+ "GITHUB_REF": os.environ.get("GITHUB_REF", "refs/heads/main"),
+ "GITHUB_SHA": os.environ.get("GITHUB_SHA", ""),
+}
+
+dockerfile_contents = df_template.render(**df_args)
+
+temp_dir = tempfile.mkdtemp()
+with open(pathlib.Path(temp_dir) / "Dockerfile", "w", encoding="utf-8") as f:
+ f.write(dockerfile_contents)
+
+cicd_image = (
+ Image.from_dockerfile(
+ pathlib.Path(temp_dir) / "Dockerfile",
+ force_build=True,
+ gpu="A10G",
+ )
+ .env(df_args)
+ .pip_install("fastapi==0.110.0", "pydantic==2.6.3")
+)
+
+stub = Stub("Axolotl CI/CD", secrets=[])
+
+
+N_GPUS = int(os.environ.get("N_GPUS", 1))
+GPU_CONFIG = modal.gpu.A10G(count=N_GPUS)
+
+
+def run_cmd(cmd: str, run_folder: str):
+ import subprocess # nosec
+
+ # Propagate errors from subprocess.
+ if exit_code := subprocess.call(cmd.split(), cwd=run_folder): # nosec
+ exit(exit_code) # pylint: disable=consider-using-sys-exit
+
+
+@stub.function(
+ image=cicd_image,
+ gpu=GPU_CONFIG,
+ timeout=45 * 60,
+ cpu=8.0,
+ memory=131072,
+)
+def cicd_pytest():
+ run_cmd("./cicd/cicd.sh", "/workspace/axolotl")
+
+
+@stub.local_entrypoint()
+def main():
+ cicd_pytest.remote()
diff --git a/config.yml b/config.yml
new file mode 100644
index 0000000000000000000000000000000000000000..0257223690de3c4af6b1fcbb8da42cb73368ffc5
--- /dev/null
+++ b/config.yml
@@ -0,0 +1,66 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+push_to_hub: true
+hf_use_auth_token: true
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 1
+gradient_accumulation_steps: 1
+num_epochs: 3
+learning_rate: 0.0001
+optimizer: adamw_bnb_8bit
+optim_args:
+ amsgrad: true
+lr_scheduler: cosine
+train_on_inputs: true
+group_by_length: false
+bfloat16: false
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+unsloth_cross_entropy_loss: true
+#unsloth_lora_mlp: true
+#unsloth_lora_qkv: true
+#unsloth_lora_o: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.5
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero1.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/deepspeed_configs/zero1.json b/deepspeed_configs/zero1.json
new file mode 100644
index 0000000000000000000000000000000000000000..8a57b8605a5cb2628932b4c38dba02f7743ddc2d
--- /dev/null
+++ b/deepspeed_configs/zero1.json
@@ -0,0 +1,23 @@
+{
+ "zero_optimization": {
+ "stage": 1,
+ "overlap_comm": true
+ },
+ "bf16": {
+ "enabled": "auto"
+ },
+ "fp16": {
+ "enabled": "auto",
+ "auto_cast": false,
+ "loss_scale": 0,
+ "initial_scale_power": 32,
+ "loss_scale_window": 1000,
+ "hysteresis": 2,
+ "min_loss_scale": 1
+ },
+ "gradient_accumulation_steps": "auto",
+ "gradient_clipping": "auto",
+ "train_batch_size": "auto",
+ "train_micro_batch_size_per_gpu": "auto",
+ "wall_clock_breakdown": false
+}
diff --git a/deepspeed_configs/zero2.json b/deepspeed_configs/zero2.json
new file mode 100644
index 0000000000000000000000000000000000000000..153ac02802e9533e45a63dcbc4d14bda7e7feba1
--- /dev/null
+++ b/deepspeed_configs/zero2.json
@@ -0,0 +1,27 @@
+{
+ "zero_optimization": {
+ "stage": 2,
+ "offload_optimizer": {
+ "device": "cpu"
+ },
+ "contiguous_gradients": true,
+ "overlap_comm": true
+ },
+ "bf16": {
+ "enabled": "auto"
+ },
+ "fp16": {
+ "enabled": "auto",
+ "auto_cast": false,
+ "loss_scale": 0,
+ "initial_scale_power": 32,
+ "loss_scale_window": 1000,
+ "hysteresis": 2,
+ "min_loss_scale": 1
+ },
+ "gradient_accumulation_steps": "auto",
+ "gradient_clipping": "auto",
+ "train_batch_size": "auto",
+ "train_micro_batch_size_per_gpu": "auto",
+ "wall_clock_breakdown": false
+}
diff --git a/deepspeed_configs/zero3.json b/deepspeed_configs/zero3.json
new file mode 100644
index 0000000000000000000000000000000000000000..90ec3677ea6a7ce4f34d5d07e83d4a3393156823
--- /dev/null
+++ b/deepspeed_configs/zero3.json
@@ -0,0 +1,31 @@
+{
+ "zero_optimization": {
+ "stage": 3,
+ "overlap_comm": true,
+ "contiguous_gradients": true,
+ "sub_group_size": 0,
+ "reduce_bucket_size": "auto",
+ "stage3_prefetch_bucket_size": "auto",
+ "stage3_param_persistence_threshold": "auto",
+ "stage3_max_live_parameters": 0,
+ "stage3_max_reuse_distance": 0,
+ "stage3_gather_16bit_weights_on_model_save": true
+ },
+ "bf16": {
+ "enabled": "auto"
+ },
+ "fp16": {
+ "enabled": "auto",
+ "auto_cast": false,
+ "loss_scale": 0,
+ "initial_scale_power": 32,
+ "loss_scale_window": 1000,
+ "hysteresis": 2,
+ "min_loss_scale": 1
+ },
+ "gradient_accumulation_steps": "auto",
+ "gradient_clipping": "auto",
+ "train_batch_size": "auto",
+ "train_micro_batch_size_per_gpu": "auto",
+ "wall_clock_breakdown": false
+}
diff --git a/deepspeed_configs/zero3_bf16.json b/deepspeed_configs/zero3_bf16.json
new file mode 100644
index 0000000000000000000000000000000000000000..16e64d76b4d680a759ebdfbbda294cb47b52a55f
--- /dev/null
+++ b/deepspeed_configs/zero3_bf16.json
@@ -0,0 +1,31 @@
+{
+ "zero_optimization": {
+ "stage": 3,
+ "overlap_comm": true,
+ "contiguous_gradients": true,
+ "sub_group_size": 0,
+ "reduce_bucket_size": "auto",
+ "stage3_prefetch_bucket_size": "auto",
+ "stage3_param_persistence_threshold": "auto",
+ "stage3_max_live_parameters": 0,
+ "stage3_max_reuse_distance": 0,
+ "stage3_gather_16bit_weights_on_model_save": true
+ },
+ "bf16": {
+ "enabled": true
+ },
+ "fp16": {
+ "enabled": "auto",
+ "auto_cast": false,
+ "loss_scale": 0,
+ "initial_scale_power": 32,
+ "loss_scale_window": 1000,
+ "hysteresis": 2,
+ "min_loss_scale": 1
+ },
+ "gradient_accumulation_steps": "auto",
+ "gradient_clipping": "auto",
+ "train_batch_size": "auto",
+ "train_micro_batch_size_per_gpu": "auto",
+ "wall_clock_breakdown": false
+}
diff --git a/deepspeed_configs/zero3_bf16_cpuoffload_all.json b/deepspeed_configs/zero3_bf16_cpuoffload_all.json
new file mode 100644
index 0000000000000000000000000000000000000000..09ca6785b2e05614092a34245f3cd007f2872eba
--- /dev/null
+++ b/deepspeed_configs/zero3_bf16_cpuoffload_all.json
@@ -0,0 +1,41 @@
+{
+ "zero_force_ds_cpu_optimizer": false,
+ "zero_allow_untested_optimizer": true,
+ "zero_optimization": {
+ "stage": 3,
+ "offload_optimizer": {
+ "device": "cpu",
+ "pin_memory": true
+ },
+ "offload_param": {
+ "device": "cpu",
+ "pin_memory": true
+ },
+ "overlap_comm": true,
+ "contiguous_gradients": true,
+ "sub_group_size": 0,
+ "reduce_bucket_size": "auto",
+ "stage3_prefetch_bucket_size": "auto",
+ "stage3_param_persistence_threshold": "auto",
+ "stage3_max_live_parameters": 0,
+ "stage3_max_reuse_distance": 0,
+ "stage3_gather_16bit_weights_on_model_save": true
+ },
+ "bf16": {
+ "enabled": true
+ },
+ "fp16": {
+ "enabled": "auto",
+ "auto_cast": false,
+ "loss_scale": 0,
+ "initial_scale_power": 32,
+ "loss_scale_window": 1000,
+ "hysteresis": 2,
+ "min_loss_scale": 1
+ },
+ "gradient_accumulation_steps": "auto",
+ "gradient_clipping": "auto",
+ "train_batch_size": "auto",
+ "train_micro_batch_size_per_gpu": "auto",
+ "wall_clock_breakdown": false
+}
diff --git a/deepspeed_configs/zero3_bf16_cpuoffload_params.json b/deepspeed_configs/zero3_bf16_cpuoffload_params.json
new file mode 100644
index 0000000000000000000000000000000000000000..41d4a21323414f3192ca544d10099dda479c3b4d
--- /dev/null
+++ b/deepspeed_configs/zero3_bf16_cpuoffload_params.json
@@ -0,0 +1,37 @@
+{
+ "zero_force_ds_cpu_optimizer": false,
+ "zero_allow_untested_optimizer": true,
+ "zero_optimization": {
+ "stage": 3,
+ "offload_param": {
+ "device": "cpu",
+ "pin_memory": true
+ },
+ "overlap_comm": true,
+ "contiguous_gradients": true,
+ "sub_group_size": 0,
+ "reduce_bucket_size": "auto",
+ "stage3_prefetch_bucket_size": "auto",
+ "stage3_param_persistence_threshold": "auto",
+ "stage3_max_live_parameters": 0,
+ "stage3_max_reuse_distance": 0,
+ "stage3_gather_16bit_weights_on_model_save": true
+ },
+ "bf16": {
+ "enabled": true
+ },
+ "fp16": {
+ "enabled": "auto",
+ "auto_cast": false,
+ "loss_scale": 0,
+ "initial_scale_power": 32,
+ "loss_scale_window": 1000,
+ "hysteresis": 2,
+ "min_loss_scale": 1
+ },
+ "gradient_accumulation_steps": "auto",
+ "gradient_clipping": "auto",
+ "train_batch_size": "auto",
+ "train_micro_batch_size_per_gpu": "auto",
+ "wall_clock_breakdown": false
+}
diff --git a/devtools/README.md b/devtools/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..0114ee3a808f91feaf9972e12bf76c9aeb709713
--- /dev/null
+++ b/devtools/README.md
@@ -0,0 +1 @@
+This directory contains example config files that might be useful for debugging. Please see [docs/debugging.qmd](../docs/debugging.qmd) for more information.
diff --git a/devtools/dev_sharegpt.yml b/devtools/dev_sharegpt.yml
new file mode 100644
index 0000000000000000000000000000000000000000..9c65b49dcd26a83319b29b6b43bbfa2d77893b3b
--- /dev/null
+++ b/devtools/dev_sharegpt.yml
@@ -0,0 +1,48 @@
+# Example config for debugging the sharegpt prompt format
+base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+
+datasets:
+ - path: philschmid/guanaco-sharegpt-style
+ type: sharegpt
+ shards: 10
+val_set_size: 0
+output_dir: temp_debug/axolotl_outputs/model
+dataset_prepared_path: temp_debug/axolotl_outputs/data
+dataset_processes: 1
+
+sequence_len: 4096
+sample_packing: false
+pad_to_sequence_len: true
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+micro_batch_size: 1
+num_epochs: 1
+max_steps: 10
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: false
+fp16: true
+tf32: false
+
+gradient_checkpointing: true
+logging_steps: 1
+flash_attention: true
+
+warmup_steps: 10
+weight_decay: 0.0
diff --git a/docker-compose.yaml b/docker-compose.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..a16be726cfaa8c85e78f5abccdd33faba59028ba
--- /dev/null
+++ b/docker-compose.yaml
@@ -0,0 +1,25 @@
+# version: '3.8'
+services:
+ axolotl:
+ build:
+ context: .
+ dockerfile: ./docker/Dockerfile
+ volumes:
+ - .:/workspace/axolotl
+ - ~/.cache/huggingface/:/root/.cache/huggingface/
+ # set environment variables
+ environment:
+ # Set environment variables
+ - GIT_AUTHOR_NAME=${GIT_AUTHOR_NAME}
+ - GIT_AUTHOR_EMAIL=${GIT_AUTHOR_EMAIL}
+ - GIT_COMMITTER_NAME=${GIT_COMMITTER_NAME}
+ - GIT_COMMITTER_EMAIL=${GIT_COMMITTER_EMAIL}
+ - WANDB_API_KEY=${WANDB_API_KEY}
+ deploy:
+ resources:
+ reservations:
+ devices:
+ - driver: nvidia
+ # count: 1
+ capabilities: [gpu]
+ command: tail -f /dev/null
diff --git a/docker/Dockerfile b/docker/Dockerfile
new file mode 100644
index 0000000000000000000000000000000000000000..6b9cf7d4c42537091b97cac7b085f96daf35b712
--- /dev/null
+++ b/docker/Dockerfile
@@ -0,0 +1,38 @@
+ARG BASE_TAG=main-base
+FROM winglian/axolotl-base:$BASE_TAG
+
+ARG TORCH_CUDA_ARCH_LIST="7.0 7.5 8.0 8.6+PTX"
+ARG AXOLOTL_EXTRAS=""
+ARG AXOLOTL_ARGS=""
+ARG CUDA="118"
+ENV BNB_CUDA_VERSION=$CUDA
+ARG PYTORCH_VERSION="2.1.2"
+
+ENV PYTORCH_VERSION=$PYTORCH_VERSION
+
+RUN apt-get update && \
+ apt-get install -y --allow-change-held-packages vim curl nano libnccl2 libnccl-dev rsync s3fs
+
+WORKDIR /workspace
+
+RUN git clone --depth=1 https://github.com/OpenAccess-AI-Collective/axolotl.git
+
+WORKDIR /workspace/axolotl
+
+# If AXOLOTL_EXTRAS is set, append it in brackets
+RUN pip install causal_conv1d
+RUN if [ "$AXOLOTL_EXTRAS" != "" ] ; then \
+ pip install -e .[deepspeed,flash-attn,mamba-ssm,galore,$AXOLOTL_EXTRAS] $AXOLOTL_ARGS; \
+ else \
+ pip install -e .[deepspeed,flash-attn,mamba-ssm,galore] $AXOLOTL_ARGS; \
+ fi
+
+# So we can test the Docker image
+RUN pip install pytest
+
+# fix so that git fetch/pull from remote works
+RUN git config remote.origin.fetch "+refs/heads/*:refs/remotes/origin/*" && \
+ git config --get remote.origin.fetch
+
+# helper for huggingface-login cli
+RUN git config --global credential.helper store
diff --git a/docker/Dockerfile-base b/docker/Dockerfile-base
new file mode 100644
index 0000000000000000000000000000000000000000..1de5537dace76ca6b9d2dc2d7f98b32f6d1c667b
--- /dev/null
+++ b/docker/Dockerfile-base
@@ -0,0 +1,37 @@
+ARG CUDA_VERSION="11.8.0"
+ARG CUDNN_VERSION="8"
+ARG UBUNTU_VERSION="22.04"
+ARG MAX_JOBS=4
+
+FROM nvidia/cuda:$CUDA_VERSION-cudnn$CUDNN_VERSION-devel-ubuntu$UBUNTU_VERSION as base-builder
+
+ENV PATH="/root/miniconda3/bin:${PATH}"
+
+ARG PYTHON_VERSION="3.10"
+ARG PYTORCH_VERSION="2.1.2"
+ARG CUDA="118"
+ARG TORCH_CUDA_ARCH_LIST="7.0 7.5 8.0 8.6 9.0+PTX"
+
+ENV PYTHON_VERSION=$PYTHON_VERSION
+ENV TORCH_CUDA_ARCH_LIST=$TORCH_CUDA_ARCH_LIST
+
+RUN apt-get update \
+ && apt-get install -y wget git build-essential ninja-build git-lfs libaio-dev && rm -rf /var/lib/apt/lists/* \
+ && wget \
+ https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh \
+ && mkdir /root/.conda \
+ && bash Miniconda3-latest-Linux-x86_64.sh -b \
+ && rm -f Miniconda3-latest-Linux-x86_64.sh \
+ && conda create -n "py${PYTHON_VERSION}" python="${PYTHON_VERSION}"
+
+ENV PATH="/root/miniconda3/envs/py${PYTHON_VERSION}/bin:${PATH}"
+
+WORKDIR /workspace
+
+RUN python3 -m pip install --upgrade pip && pip3 install packaging && \
+ python3 -m pip install --no-cache-dir -U torch==${PYTORCH_VERSION}+cu${CUDA} --extra-index-url https://download.pytorch.org/whl/cu$CUDA
+
+RUN git lfs install --skip-repo && \
+ pip3 install awscli && \
+ # The base image ships with `pydantic==1.8.2` which is not working
+ pip3 install -U --no-cache-dir pydantic==1.10.10
diff --git a/docker/Dockerfile-cloud b/docker/Dockerfile-cloud
new file mode 100644
index 0000000000000000000000000000000000000000..69ce143bb22118284eb5ae3cc68e5601fffb50ee
--- /dev/null
+++ b/docker/Dockerfile-cloud
@@ -0,0 +1,27 @@
+ARG BASE_TAG=main
+FROM winglian/axolotl:$BASE_TAG
+
+ENV HF_DATASETS_CACHE="/workspace/data/huggingface-cache/datasets"
+ENV HUGGINGFACE_HUB_CACHE="/workspace/data/huggingface-cache/hub"
+ENV TRANSFORMERS_CACHE="/workspace/data/huggingface-cache/hub"
+ENV HF_HOME="/workspace/data/huggingface-cache/hub"
+ENV HF_HUB_ENABLE_HF_TRANSFER="1"
+
+EXPOSE 8888
+EXPOSE 22
+
+COPY scripts/cloud-entrypoint.sh /root/cloud-entrypoint.sh
+COPY scripts/motd /etc/motd
+
+RUN pip install jupyterlab notebook ipywidgets && \
+ jupyter lab clean
+RUN apt install --yes --no-install-recommends openssh-server tmux && \
+ mkdir -p ~/.ssh && \
+ chmod 700 ~/.ssh && \
+ printf "\n[[ -z \"\$TMUX\" ]] && { tmux attach-session -t ssh_tmux || tmux new-session -s ssh_tmux; exit; }\n" >> ~/.bashrc && \
+ printf "[ ! -z \"\$TERM\" -a -r /etc/motd ] && cat /etc/motd\n" >> ~/.bashrc && \
+ chmod +x /workspace/axolotl/scripts/cloud-entrypoint.sh && \
+ chmod +x /root/cloud-entrypoint.sh
+
+ENTRYPOINT ["/root/cloud-entrypoint.sh"]
+CMD ["sleep", "infinity"]
diff --git a/docker/Dockerfile-cloud-no-tmux b/docker/Dockerfile-cloud-no-tmux
new file mode 100644
index 0000000000000000000000000000000000000000..8af194de0242bb8a67908469b4af355641533c3a
--- /dev/null
+++ b/docker/Dockerfile-cloud-no-tmux
@@ -0,0 +1,26 @@
+ARG BASE_TAG=main
+FROM winglian/axolotl:$BASE_TAG
+
+ENV HF_DATASETS_CACHE="/workspace/data/huggingface-cache/datasets"
+ENV HUGGINGFACE_HUB_CACHE="/workspace/data/huggingface-cache/hub"
+ENV TRANSFORMERS_CACHE="/workspace/data/huggingface-cache/hub"
+ENV HF_HOME="/workspace/data/huggingface-cache/hub"
+ENV HF_HUB_ENABLE_HF_TRANSFER="1"
+
+EXPOSE 8888
+EXPOSE 22
+
+COPY scripts/cloud-entrypoint.sh /root/cloud-entrypoint.sh
+COPY scripts/motd /etc/motd
+
+RUN pip install jupyterlab notebook ipywidgets && \
+ jupyter lab clean
+RUN apt install --yes --no-install-recommends openssh-server tmux && \
+ mkdir -p ~/.ssh && \
+ chmod 700 ~/.ssh && \
+ printf "[ ! -z \"\$TERM\" -a -r /etc/motd ] && cat /etc/motd\n" >> ~/.bashrc && \
+ chmod +x /workspace/axolotl/scripts/cloud-entrypoint.sh && \
+ chmod +x /root/cloud-entrypoint.sh
+
+ENTRYPOINT ["/root/cloud-entrypoint.sh"]
+CMD ["sleep", "infinity"]
diff --git a/docker/Dockerfile-tests b/docker/Dockerfile-tests
new file mode 100644
index 0000000000000000000000000000000000000000..d5c0595a8771424ceb1e3cab88bd74314a1e3abe
--- /dev/null
+++ b/docker/Dockerfile-tests
@@ -0,0 +1,41 @@
+ARG BASE_TAG=main-base
+FROM winglian/axolotl-base:$BASE_TAG
+
+ARG TORCH_CUDA_ARCH_LIST="7.0 7.5 8.0 8.6+PTX"
+ARG AXOLOTL_EXTRAS=""
+ARG AXOLOTL_ARGS=""
+ARG CUDA="118"
+ENV BNB_CUDA_VERSION=$CUDA
+ARG PYTORCH_VERSION="2.1.2"
+ARG GITHUB_REF="main"
+
+ENV PYTORCH_VERSION=$PYTORCH_VERSION
+
+RUN apt-get update && \
+ apt-get install -y --allow-change-held-packages vim curl nano libnccl2 libnccl-dev
+
+WORKDIR /workspace
+
+RUN git clone --depth=1 https://github.com/OpenAccess-AI-Collective/axolotl.git
+
+WORKDIR /workspace/axolotl
+
+RUN git fetch origin +$GITHUB_REF && \
+ git checkout FETCH_HEAD
+
+# If AXOLOTL_EXTRAS is set, append it in brackets
+RUN if [ "$AXOLOTL_EXTRAS" != "" ] ; then \
+ pip install -e .[deepspeed,flash-attn,mamba-ssm,$AXOLOTL_EXTRAS] $AXOLOTL_ARGS; \
+ else \
+ pip install -e .[deepspeed,flash-attn,mamba-ssm] $AXOLOTL_ARGS; \
+ fi
+
+# So we can test the Docker image
+RUN pip install pytest
+
+# fix so that git fetch/pull from remote works
+RUN git config remote.origin.fetch "+refs/heads/*:refs/remotes/origin/*" && \
+ git config --get remote.origin.fetch
+
+# helper for huggingface-login cli
+RUN git config --global credential.helper store
diff --git a/docs/.gitignore b/docs/.gitignore
new file mode 100644
index 0000000000000000000000000000000000000000..4c23a061fa46d2db78ad1dc344484d953bd1a711
--- /dev/null
+++ b/docs/.gitignore
@@ -0,0 +1,2 @@
+/.quarto/
+_site/
diff --git a/docs/batch_vs_grad.qmd b/docs/batch_vs_grad.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..e7b3b7d818310da435d736feb16c70106295f87f
--- /dev/null
+++ b/docs/batch_vs_grad.qmd
@@ -0,0 +1,59 @@
+---
+title: Batch size vs Gradient accumulation
+description: Understanding of batch size and gradient accumulation steps
+---
+
+Gradient accumulation means accumulating gradients over several mini-batches and updating the model weights afterward. When the samples in each batch are diverse, this technique doesn't significantly impact learning.
+
+This method allows for effective training with larger effective batch sizes without needing proportionally larger memory. Here's why:
+
+1. **Memory Consumption with Batch Size**: The primary reason increasing the batch size impacts memory is due to the storage requirements for intermediate activations. When you forward propagate a batch through a network, you have to store the activations at each layer for each sample in the batch, because these activations are used during backpropagation to compute gradients. Therefore, larger batches mean more activations, leading to greater GPU memory consumption.
+
+2. **Gradient Accumulation**: With gradient accumulation, you're effectively simulating a larger batch size by accumulating gradients over several smaller batches (or micro-batches). However, at any given time, you're only forward and backward propagating a micro-batch. This means you only store activations for the micro-batch, not the full accumulated batch. As a result, you can simulate the effect of a larger batch size without the memory cost of storing activations for a large batch.
+
+**Example 1:**
+Micro batch size: 3
+Gradient accumulation steps: 2
+Number of GPUs: 3
+Total batch size = 3 * 2 * 3 = 18
+
+```
+| GPU 1 | GPU 2 | GPU 3 |
+|----------------|----------------|----------------|
+| S1, S2, S3 | S4, S5, S6 | S7, S8, S9 |
+| e1, e2, e3 | e4, e5, e6 | e7, e8, e9 |
+|----------------|----------------|----------------|
+| → (accumulate) | → (accumulate) | → (accumulate) |
+|----------------|----------------|----------------|
+| S10, S11, S12 | S13, S14, S15 | S16, S17, S18 |
+| e10, e11, e12 | e13, e14, e15 | e16, e17, e18 |
+|----------------|----------------|----------------|
+| → (apply) | → (apply) | → (apply) |
+
+Accumulated gradient for the weight w1 after the second iteration (considering all GPUs):
+Total gradient for w1 = e1 + e2 + e3 + e4 + e5 + e6 + e7 + e8 + e9 + e10 + e11 + e12 + e13 + e14 + e15 + e16 + e17 + e18
+
+Weight update for w1:
+w1_new = w1_old - learning rate x (Total gradient for w1 / 18)
+```
+
+**Example 2:**
+Micro batch size: 2
+Gradient accumulation steps: 1
+Number of GPUs: 3
+Total batch size = 2 * 1 * 3 = 6
+
+```
+| GPU 1 | GPU 2 | GPU 3 |
+|-----------|-----------|-----------|
+| S1, S2 | S3, S4 | S5, S6 |
+| e1, e2 | e3, e4 | e5, e6 |
+|-----------|-----------|-----------|
+| → (apply) | → (apply) | → (apply) |
+
+Accumulated gradient for the weight w1 (considering all GPUs):
+Total gradient for w1 = e1 + e2 + e3 + e4 + e5 + e6
+
+Weight update for w1:
+w1_new = w1_old - learning rate × (Total gradient for w1 / 6)
+```
diff --git a/docs/config.qmd b/docs/config.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..570a173f9a7875484492486a89bd3179389c99af
--- /dev/null
+++ b/docs/config.qmd
@@ -0,0 +1,453 @@
+---
+title: Config options
+description: A complete list of all configuration options.
+---
+
+```yaml
+# This is the huggingface model that contains *.pt, *.safetensors, or *.bin files
+# This can also be a relative path to a model on disk
+base_model: ./llama-7b-hf
+# You can specify an ignore pattern if the model repo contains more than 1 model type (*.pt, etc)
+base_model_ignore_patterns:
+# If the base_model repo on hf hub doesn't include configuration .json files,
+# You can set that here, or leave this empty to default to base_model
+base_model_config: ./llama-7b-hf
+# You can specify to choose a specific model revision from huggingface hub
+revision_of_model:
+# Optional tokenizer configuration path in case you want to use a different tokenizer
+# than the one defined in the base model
+tokenizer_config:
+# If you want to specify the type of model to load, AutoModelForCausalLM is a good choice too
+model_type: AutoModelForCausalLM
+# Corresponding tokenizer for the model AutoTokenizer is a good choice
+tokenizer_type: AutoTokenizer
+# Trust remote code for untrusted source
+trust_remote_code:
+# use_fast option for tokenizer loading from_pretrained, default to True
+tokenizer_use_fast:
+# Whether to use the legacy tokenizer setting, defaults to True
+tokenizer_legacy:
+# Resize the model embeddings when new tokens are added to multiples of 32
+# This is reported to improve training speed on some models
+resize_token_embeddings_to_32x:
+
+# (Internal use only)
+# Used to identify which the model is based on
+is_falcon_derived_model:
+is_llama_derived_model:
+is_qwen_derived_model:
+# Please note that if you set this to true, `padding_side` will be set to "left" by default
+is_mistral_derived_model:
+
+# optional overrides to the base model configuration
+overrides_of_model_config:
+ # RoPE Scaling https://github.com/huggingface/transformers/pull/24653
+ rope_scaling:
+ type: # linear | dynamic
+ factor: # float
+
+# optional overrides to the bnb 4bit quantization configuration
+# https://huggingface.co/docs/transformers/main/main_classes/quantization#transformers.BitsAndBytesConfig
+bnb_config_kwargs:
+ # These are default values
+ llm_int8_has_fp16_weight: false
+ bnb_4bit_quant_type: nf4
+ bnb_4bit_use_double_quant: true
+
+
+# Whether you are training a 4-bit GPTQ quantized model
+gptq: true
+
+# This will attempt to quantize the model down to 8 bits and use adam 8 bit optimizer
+load_in_8bit: true
+# Use bitsandbytes 4 bit
+load_in_4bit:
+
+# Use CUDA bf16
+bf16: true # bool or 'full' for `bf16_full_eval`. require >=ampere
+# Use CUDA fp16
+fp16: true
+# Use CUDA tf32
+tf32: true # require >=ampere
+
+# No AMP (automatic mixed precision)
+bfloat16: true # require >=ampere
+float16: true
+
+# Limit the memory for all available GPUs to this amount (if an integer, expressed in gigabytes); default: unset
+gpu_memory_limit: 20GiB
+# Do the LoRA/PEFT loading on CPU -- this is required if the base model is so large it takes up most or all of the available GPU VRAM, e.g. during a model and LoRA merge
+lora_on_cpu: true
+
+# A list of one or more datasets to finetune the model with
+datasets:
+ # HuggingFace dataset repo | s3://,gs:// path | "json" for local dataset, make sure to fill data_files
+ - path: vicgalle/alpaca-gpt4
+ # The type of prompt to use for training. [alpaca, sharegpt, gpteacher, oasst, reflection]
+ type: alpaca # format | format: (chat/instruct) | .load_
+ ds_type: # Optional[str] (json|arrow|parquet|text|csv) defines the datatype when path is a file
+ data_files: # Optional[str] path to source data files
+ shards: # Optional[int] number of shards to split data into
+ name: # Optional[str] name of dataset configuration to load
+ train_on_split: train # Optional[str] name of dataset split to load from
+
+ # Optional[str] fastchat conversation type, only used with type: sharegpt
+ conversation: # Options (see Conversation 'name'): https://github.com/lm-sys/FastChat/blob/main/fastchat/conversation.py
+ field_human: # Optional[str]. Human key to use for conversation.
+ field_model: # Optional[str]. Assistant key to use for conversation.
+ # Add additional keys from your dataset as input or output roles
+ roles:
+ input: # Optional[List[str]]. These will be masked based on train_on_input
+ output: # Optional[List[str]].
+
+ # Custom user instruction prompt
+ - path: repo
+ type:
+ # The below are defaults. only set what's needed if you use a different column name.
+ system_prompt: ""
+ system_format: "{system}"
+ field_system: system
+ field_instruction: instruction
+ field_input: input
+ field_output: output
+
+ # Customizable to be single line or multi-line
+ # Use {instruction}/{input} as key to be replaced
+ # 'format' can include {input}
+ format: |-
+ User: {instruction} {input}
+ Assistant:
+ # 'no_input_format' cannot include {input}
+ no_input_format: "{instruction} "
+
+ # For `completion` datsets only, uses the provided field instead of `text` column
+ field:
+
+# If false, the datasets will not be shuffled and will keep their original order in `datasets`.
+# The same applies to the `test_datasets` option and the `pretraining_dataset` option. Default is true.
+shuffle_merged_datasets: true
+
+# A list of one or more datasets to eval the model with.
+# You can use either test_datasets, or val_set_size, but not both.
+test_datasets:
+ - path: /workspace/data/eval.jsonl
+ ds_type: json
+ # You need to specify a split. For "json" datasets the default split is called "train".
+ split: train
+ type: completion
+ data_files:
+ - /workspace/data/eval.jsonl
+
+# use RL training: 'dpo', 'ipo', 'kto_pair'
+rl:
+
+# Saves the desired chat template to the tokenizer_config.json for easier inferencing
+# Currently supports chatml and inst (mistral/mixtral)
+chat_template: chatml
+# Changes the default system message
+default_system_message: You are a helpful assistant. Please give a long and detailed answer. # Currently only supports chatml.
+# Axolotl attempts to save the dataset as an arrow after packing the data together so
+# subsequent training attempts load faster, relative path
+dataset_prepared_path: data/last_run_prepared
+# Push prepared dataset to hub
+push_dataset_to_hub: # repo path
+# The maximum number of processes to use while preprocessing your input dataset. This defaults to `os.cpu_count()`
+# if not set.
+dataset_processes: # defaults to os.cpu_count() if not set
+# Keep dataset in memory while preprocessing
+# Only needed if cached dataset is taking too much storage
+dataset_keep_in_memory:
+# push checkpoints to hub
+hub_model_id: # private repo path to push finetuned model
+# how to push checkpoints to hub
+# https://huggingface.co/docs/transformers/v4.31.0/en/main_classes/trainer#transformers.TrainingArguments.hub_strategy
+hub_strategy:
+# Whether to use hf `use_auth_token` for loading datasets. Useful for fetching private datasets
+# Required to be true when used in combination with `push_dataset_to_hub`
+hf_use_auth_token: # boolean
+# How much of the dataset to set aside as evaluation. 1 = 100%, 0.50 = 50%, etc. 0 for no eval.
+val_set_size: 0.04
+# Num shards for whole dataset
+dataset_shard_num:
+# Index of shard to use for whole dataset
+dataset_shard_idx:
+
+# The maximum length of an input to train with, this should typically be less than 2048
+# as most models have a token/context limit of 2048
+sequence_len: 2048
+# Pad inputs so each step uses constant sized buffers
+# This will reduce memory fragmentation and may prevent OOMs, by re-using memory more efficiently
+pad_to_sequence_len:
+# Use efficient multi-packing with block diagonal attention and per sequence position_ids. Recommend set to 'true'
+sample_packing:
+# Set to 'false' if getting errors during eval with sample_packing on.
+eval_sample_packing:
+# You can set these packing optimizations AFTER starting a training at least once.
+# The trainer will provide recommended values for these values.
+sample_packing_eff_est:
+total_num_tokens:
+
+# Passed through to transformers when loading the model when launched without accelerate
+# Use `sequential` when training w/ model parallelism to limit memory
+device_map:
+# Defines the max memory usage per gpu on the system. Passed through to transformers when loading the model.
+max_memory:
+
+# If you want to use 'lora' or 'qlora' or leave blank to train all parameters in original model
+adapter: lora
+# If you already have a lora model trained that you want to load, put that here.
+# This means after training, if you want to test the model, you should set this to the value of `output_dir`.
+# Note that if you merge an adapter to the base model, a new subdirectory `merged` will be created under the `output_dir`.
+lora_model_dir:
+
+# LoRA hyperparameters
+# For more details about the following options, see:
+# https://www.anyscale.com/blog/fine-tuning-llms-lora-or-full-parameter-an-in-depth-analysis-with-llama-2
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+ - q_proj
+ - v_proj
+# - k_proj
+# - o_proj
+# - gate_proj
+# - down_proj
+# - up_proj
+lora_target_linear: # If true, will target all linear modules
+peft_layers_to_transform: # The layer indices to transform, otherwise, apply to all layers
+
+# If you added new tokens to the tokenizer, you may need to save some LoRA modules because they need to know the new tokens.
+# For LLaMA and Mistral, you need to save `embed_tokens` and `lm_head`. It may vary for other models.
+# `embed_tokens` converts tokens to embeddings, and `lm_head` converts embeddings to token probabilities.
+# https://github.com/huggingface/peft/issues/334#issuecomment-1561727994
+lora_modules_to_save:
+# - embed_tokens
+# - lm_head
+
+lora_fan_in_fan_out: false
+
+# LoRA+ hyperparameters
+# For more details about the following options, see:
+# https://arxiv.org/abs/2402.12354 and `src/axolotl/core/train_builder.py`
+loraplus_lr_ratio: # loraplus learning rate ratio lr_B / lr_A. Recommended value is 2^4.
+loraplus_lr_embedding: # loraplus learning rate for lora embedding layers. Default value is 1e-6.
+
+peft:
+ # Configuration options for loftq initialization for LoRA
+ # https://huggingface.co/docs/peft/developer_guides/quantization#loftq-initialization
+ loftq_config:
+ loftq_bits: # typically 4 bits
+
+# ReLoRA configuration
+# Must use either 'lora' or 'qlora' adapter, and does not support fsdp or deepspeed
+relora_steps: # Number of steps per ReLoRA restart
+relora_warmup_steps: # Number of per-restart warmup steps
+relora_anneal_steps: # Number of anneal steps for each relora cycle
+relora_prune_ratio: # threshold for optimizer magnitude when pruning
+relora_cpu_offload: # True to perform lora weight merges on cpu during restarts, for modest gpu memory savings
+
+# wandb configuration if you're using it
+# Make sure your `WANDB_API_KEY` environment variable is set (recommended) or you login to wandb with `wandb login`.
+wandb_mode: # "offline" to save run metadata locally and not sync to the server, "disabled" to turn off wandb
+wandb_project: # Your wandb project name
+wandb_entity: # A wandb Team name if using a Team
+wandb_watch:
+wandb_name: # Set the name of your wandb run
+wandb_run_id: # Set the ID of your wandb run
+wandb_log_model: # "checkpoint" to log model to wandb Artifacts every `save_steps` or "end" to log only at the end of training
+
+# mlflow configuration if you're using it
+mlflow_tracking_uri: # URI to mlflow
+mlflow_experiment_name: # Your experiment name
+hf_mlflow_log_artifacts: # set to true to copy each saved checkpoint on each save to mlflow artifact registry
+
+# Where to save the full-finetuned model to
+output_dir: ./completed-model
+
+# Whether to use torch.compile and which backend to use
+torch_compile: # bool
+torch_compile_backend: # Optional[str]
+
+# Training hyperparameters
+
+# If greater than 1, backpropagation will be skipped and the gradients will be accumulated for the given number of steps.
+gradient_accumulation_steps: 1
+# The number of samples to include in each batch. This is the number of samples sent to each GPU.
+# Batch size per gpu = micro_batch_size * gradient_accumulation_steps
+micro_batch_size: 2
+eval_batch_size:
+num_epochs: 4
+warmup_steps: 100 # cannot use with warmup_ratio
+warmup_ratio: 0.05 # cannot use with warmup_steps
+learning_rate: 0.00003
+lr_quadratic_warmup:
+logging_steps:
+eval_steps: # Leave empty to eval at each epoch, integers for every N steps. decimal for fraction of total steps
+evals_per_epoch: # number of times per epoch to run evals, mutually exclusive with eval_steps
+save_strategy: # Set to `no` to skip checkpoint saves
+save_steps: # Leave empty to save at each epoch
+saves_per_epoch: # number of times per epoch to save a checkpoint, mutually exclusive with save_steps
+save_total_limit: # Checkpoints saved at a time
+# Maximum number of iterations to train for. It precedes num_epochs which means that
+# if both are set, num_epochs will not be guaranteed.
+# e.g., when 1 epoch is 1000 steps => `num_epochs: 2` and `max_steps: 100` will train for 100 steps
+max_steps:
+
+eval_table_size: # Approximate number of predictions sent to wandb depending on batch size. Enabled above 0. Default is 0
+eval_max_new_tokens: # Total number of tokens generated for predictions sent to wandb. Default is 128
+eval_causal_lm_metrics: # HF evaluate metrics used during evaluation. Default is ["sacrebleu", "comet", "ter", chrf]
+
+loss_watchdog_threshold: # High loss value, indicating the learning has broken down (a good estimate is ~2 times the loss at the start of training)
+loss_watchdog_patience: # Number of high-loss steps in a row before the trainer aborts (default: 3)
+
+# Save model as safetensors (require safetensors package)
+save_safetensors:
+
+# Whether to mask out or include the human's prompt from the training labels
+train_on_inputs: false
+# Group similarly sized data to minimize padding.
+# May be slower to start, as it must download and sort the entire dataset.
+# Note that training loss may have an oscillating pattern with this enabled.
+group_by_length: false
+
+# Whether to use gradient checkpointing https://huggingface.co/docs/transformers/v4.18.0/en/performance#gradient-checkpointing
+gradient_checkpointing: false
+# additional kwargs to pass to the trainer for gradient checkpointing
+# gradient_checkpointing_kwargs:
+# use_reentrant: true
+
+# Stop training after this many evaluation losses have increased in a row
+# https://huggingface.co/transformers/v4.2.2/_modules/transformers/trainer_callback.html#EarlyStoppingCallback
+early_stopping_patience: 3
+
+# Specify a scheduler and kwargs to use with the optimizer
+lr_scheduler: # 'one_cycle' | 'log_sweep' | empty for cosine
+lr_scheduler_kwargs:
+cosine_min_lr_ratio: # decay lr to some percentage of the peak lr, e.g. cosine_min_lr_ratio=0.1 for 10% of peak lr
+cosine_constant_lr_ratio: # freeze lr at some percentage of the step, e.g. cosine_constant_lr_ratio=0.8 means start cosine_min_lr at 80% of training step (https://arxiv.org/pdf/2308.04014.pdf)
+
+# For one_cycle optim
+lr_div_factor: # Learning rate div factor
+
+# Specify optimizer
+# Valid values are driven by the Transformers OptimizerNames class, see:
+# https://github.com/huggingface/transformers/blob/95b374952dc27d8511541d6f5a4e22c9ec11fb24/src/transformers/training_args.py#L134
+#
+# Note that not all optimizers may be available in your environment, ex: 'adamw_anyprecision' is part of
+# torchdistx, 'adamw_bnb_8bit' is part of bnb.optim.Adam8bit, etc. When in doubt, it is recommended to start with the optimizer used
+# in the examples/ for your model and fine-tuning use case.
+#
+# Valid values for 'optimizer' include:
+# - adamw_hf
+# - adamw_torch
+# - adamw_torch_fused
+# - adamw_torch_xla
+# - adamw_apex_fused
+# - adafactor
+# - adamw_anyprecision
+# - sgd
+# - adagrad
+# - adamw_bnb_8bit
+# - lion_8bit
+# - lion_32bit
+# - paged_adamw_32bit
+# - paged_adamw_8bit
+# - paged_lion_32bit
+# - paged_lion_8bit
+# - galore_adamw
+# - galore_adamw_8bit
+# - galore_adafactor
+# - galore_adamw_layerwise
+# - galore_adamw_8bit_layerwise
+# - galore_adafactor_layerwise
+optimizer:
+# Dictionary of arguments to pass to the optimizer
+optim_args:
+# For Galore Optimizers the following optim_args are available
+# rank: # type: int
+# update_proj_gap # type: int
+# scale # type: float
+# proj_type: # type: str, default = std
+
+# The target modules to optimize, i.e. the module names that you would like to train, right now this is used only for GaLore algorithm
+optim_target_modules:
+# - self_attn # for llama
+# - mlp
+
+# Specify weight decay
+weight_decay:
+# adamw hyperparams
+adam_beta1:
+adam_beta2:
+adam_epsilon:
+# Gradient clipping max norm
+max_grad_norm:
+
+# Augmentation techniques
+# NEFT https://arxiv.org/abs/2310.05914, set this to a number (paper default is 5) to add noise to embeddings
+# currently only supported on Llama and Mistral
+neftune_noise_alpha:
+
+# Whether to bettertransformers
+flash_optimum:
+# Whether to use xformers attention patch https://github.com/facebookresearch/xformers:
+xformers_attention:
+# Whether to use flash attention patch https://github.com/Dao-AILab/flash-attention:
+flash_attention:
+flash_attn_cross_entropy: # Whether to use flash-attention cross entropy implementation - advanced use only
+flash_attn_rms_norm: # Whether to use flash-attention rms norm implementation - advanced use only
+flash_attn_fuse_qkv: # Whether to fuse QKV into a single operation
+flash_attn_fuse_mlp: # Whether to fuse part of the MLP into a single operation
+# Whether to use scaled-dot-product attention
+# https://pytorch.org/docs/stable/generated/torch.nn.functional.scaled_dot_product_attention.html
+sdp_attention:
+# Shifted-sparse attention (only llama) - https://arxiv.org/pdf/2309.12307.pdf
+s2_attention:
+# Resume from a specific checkpoint dir
+resume_from_checkpoint:
+# If resume_from_checkpoint isn't set and you simply want it to start where it left off.
+# Be careful with this being turned on between different models.
+auto_resume_from_checkpoints: false
+
+# Don't mess with this, it's here for accelerate and torchrun
+local_rank:
+
+# Add or change special tokens.
+# If you add tokens here, you don't need to add them to the `tokens` list.
+special_tokens:
+ # bos_token: ""
+ # eos_token: " "
+ # unk_token: ""
+ # pad_token: "[PAD]"
+
+# Add extra tokens.
+tokens:
+
+# FSDP
+fsdp:
+fsdp_config:
+
+# Deepspeed config path. e.g., deepspeed_configs/zero3.json
+deepspeed:
+
+# Advanced DDP Arguments
+ddp_timeout:
+ddp_bucket_cap_mb:
+ddp_broadcast_buffers:
+
+# Path to torch distx for optim 'adamw_anyprecision'
+torchdistx_path:
+
+# Set to HF dataset for type: 'completion' for streaming instead of pre-tokenize
+pretraining_dataset:
+
+# Debug mode
+debug:
+
+# Seed
+seed:
+
+# Allow overwrite yml config using from cli
+strict:
+```
diff --git a/docs/dataset-formats/conversation.qmd b/docs/dataset-formats/conversation.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..f7d0cac826cb9383aa611c3207334379382ca7c3
--- /dev/null
+++ b/docs/dataset-formats/conversation.qmd
@@ -0,0 +1,63 @@
+---
+title: Conversation
+description: Conversation format for supervised fine-tuning.
+order: 3
+---
+
+## sharegpt
+
+conversations where `from` is `human`/`gpt`. (optional: first row with role `system` to override default system prompt)
+
+```{.json filename="data.jsonl"}
+{"conversations": [{"from": "...", "value": "..."}]}
+```
+
+Note: `type: sharegpt` opens special configs:
+- `conversation`: enables conversions to many Conversation types. Refer to the 'name' [here](https://github.com/lm-sys/FastChat/blob/main/fastchat/conversation.py) for options.
+- `roles`: allows you to specify the roles for input and output. This is useful for datasets with custom roles such as `tool` etc to support masking.
+- `field_human`: specify the key to use instead of `human` in the conversation.
+- `field_model`: specify the key to use instead of `gpt` in the conversation.
+
+```yaml
+datasets:
+ path: ...
+ type: sharegpt
+
+ conversation: # Options (see Conversation 'name'): https://github.com/lm-sys/FastChat/blob/main/fastchat/conversation.py
+ field_human: # Optional[str]. Human key to use for conversation.
+ field_model: # Optional[str]. Assistant key to use for conversation.
+ # Add additional keys from your dataset as input or output roles
+ roles:
+ input: # Optional[List[str]]. These will be masked based on train_on_input
+ output: # Optional[List[str]].
+```
+
+## pygmalion
+
+```{.json filename="data.jsonl"}
+{"conversations": [{"role": "...", "value": "..."}]}
+```
+
+## sharegpt.load_role
+
+conversations where `role` is used instead of `from`
+
+```{.json filename="data.jsonl"}
+{"conversations": [{"role": "...", "value": "..."}]}
+```
+
+## sharegpt.load_guanaco
+
+conversations where `from` is `prompter` `assistant` instead of default sharegpt
+
+```{.json filename="data.jsonl"}
+{"conversations": [{"from": "...", "value": "..."}]}
+```
+
+## sharegpt_jokes
+
+creates a chat where bot is asked to tell a joke, then explain why the joke is funny
+
+```{.json filename="data.jsonl"}
+{"conversations": [{"title": "...", "text": "...", "explanation": "..."}]}
+```
diff --git a/docs/dataset-formats/index.qmd b/docs/dataset-formats/index.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..91873a4c19a634cfe8dc3693472a8a95c2a5276f
--- /dev/null
+++ b/docs/dataset-formats/index.qmd
@@ -0,0 +1,14 @@
+---
+title: Dataset Formats
+description: Supported dataset formats.
+listing:
+ fields: [title, description]
+ type: table
+ sort-ui: false
+ filter-ui: false
+ max-description-length: 250
+---
+
+Axolotl supports a variety of dataset formats. It is recommended to use a JSONL format. The schema of the JSONL depends upon the task and the prompt template you wish to use. Instead of a JSONL, you can also use a HuggingFace dataset with columns for each JSONL field.
+
+Below are these various formats organized by task:
diff --git a/docs/dataset-formats/inst_tune.qmd b/docs/dataset-formats/inst_tune.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..d89c6adaf51fe31158e8bff8fd9442fe368023eb
--- /dev/null
+++ b/docs/dataset-formats/inst_tune.qmd
@@ -0,0 +1,189 @@
+---
+title: Instruction Tuning
+description: Instruction tuning formats for supervised fine-tuning.
+order: 2
+---
+
+## alpaca
+
+instruction; input(optional)
+
+```{.json filename="data.jsonl"}
+{"instruction": "...", "input": "...", "output": "..."}
+```
+
+## jeopardy
+
+question and answer
+
+```{.json filename="data.jsonl"}
+{"question": "...", "category": "...", "answer": "..."}
+```
+
+## oasst
+
+instruction
+
+```{.json filename="data.jsonl"}
+{"INSTRUCTION": "...", "RESPONSE": "..."}
+```
+
+## gpteacher
+
+instruction; input(optional)
+
+```{.json filename="data.jsonl"}
+{"instruction": "...", "input": "...", "response": "..."}
+```
+
+## reflection
+
+instruction with reflect; input(optional)
+
+```{.json filename="data.jsonl"}
+{"instruction": "...", "input": "...", "output": "...", "reflection": "...", "corrected": "..."}
+```
+
+## explainchoice
+
+question, choices, (solution OR explanation)
+
+```{.json filename="data.jsonl"}
+{"question": "...", "choices": ["..."], "solution": "...", "explanation": "..."}
+```
+
+## concisechoice
+
+question, choices, (solution OR explanation)
+
+```{.json filename="data.jsonl"}
+{"question": "...", "choices": ["..."], "solution": "...", "explanation": "..."}
+```
+
+## summarizetldr
+
+article and summary
+
+```{.json filename="data.jsonl"}
+{"article": "...", "summary": "..."}
+```
+
+## alpaca_chat
+
+basic instruct for alpaca chat
+
+```{.json filename="data.jsonl"}
+{"instruction": "...", "input": "...", "response": "..."}
+```
+
+## alpaca_chat.load_qa
+
+question and answer for alpaca chat
+
+```{.json filename="data.jsonl"}
+{"question": "...", "answer": "..."}
+```
+
+## alpaca_chat.load_concise
+
+question and answer for alpaca chat, for concise answers
+
+```{.json filename="data.jsonl"}
+{"instruction": "...", "input": "...", "response": "..."}
+```
+
+## alpaca_chat.load_camel_ai
+
+question and answer for alpaca chat, for load_camel_ai
+
+```{.json filename="data.jsonl"}
+{"message_1": "...", "message_2": "..."}
+```
+
+## alpaca_w_system.load_open_orca
+
+support for open orca datasets with included system prompts, instruct
+
+```{.json filename="data.jsonl"}
+{"system_prompt": "...", "question": "...", "response": "..."}
+```
+
+## context_qa
+
+in context question answering from an article
+
+```{.json filename="data.jsonl"}
+{"article": "...", "question": "...", "answer": "..."}
+```
+
+## context_qa.load_v2
+
+in context question answering (alternate)
+
+```{.json filename="data.jsonl"}
+{"context": "...", "question": "...", "answer": "..."}
+```
+
+## context_qa.load_404
+
+in context question answering from an article, with default response for no answer from context
+
+```{.json filename="data.jsonl"}
+{"article": "...", "unanswerable_question": "..."}
+```
+
+## creative_acr.load_answer
+
+instruction and revision
+
+```{.json filename="data.jsonl"}
+{"instruction": "...", "revision": "..."}
+```
+
+## creative_acr.load_critique
+
+critique
+
+```{.json filename="data.jsonl"}
+{"scores": "...", "critiques": "...", "instruction": "...", "answer": "..."}
+```
+
+## creative_acr.load_revise
+
+critique and revise
+
+```{.json filename="data.jsonl"}
+{"scores": "...", "critiques": "...", "instruction": "...", "answer": "...", "revision": "..."}
+```
+
+## metharme
+
+instruction, adds additional eos tokens
+
+```{.json filename="data.jsonl"}
+{"prompt": "...", "generation": "..."}
+```
+
+## How to add custom prompt format
+
+For a dataset that is preprocessed for instruction purposes:
+
+```{.json filename="data.jsonl"}
+{"input": "...", "output": "..."}
+```
+
+You can use this example in your YAML config:
+
+```{.yaml filename="config.yaml"}
+datasets:
+ - path: repo
+ type:
+ system_prompt: ""
+ field_system: system
+ field_instruction: input
+ field_output: output
+ format: "[INST] {instruction} [/INST]"
+ no_input_format: "[INST] {instruction} [/INST]"
+```
+
+See full config options under [here](../config.qmd).
diff --git a/docs/dataset-formats/pretraining.qmd b/docs/dataset-formats/pretraining.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..bb591328e247ca08a31745357f6e4999a6cc9882
--- /dev/null
+++ b/docs/dataset-formats/pretraining.qmd
@@ -0,0 +1,26 @@
+---
+title: Pre-training
+description: Data format for a pre-training completion task.
+order: 1
+---
+
+For pretraining, there is no prompt template or roles. The only required field is `text`:
+
+```{.json filename="data.jsonl"}
+{"text": "first row"}
+{"text": "second row"}
+...
+```
+
+:::{.callout-note}
+
+### Streaming is recommended for large datasets
+
+Axolotl usually loads the entire dataset into memory. This will be challenging for large datasets. Use the following config to enable streaming:
+
+```{.yaml filename="config.yaml"}
+pretraining_dataset: # hf path only
+...
+```
+
+:::
diff --git a/docs/dataset-formats/template_free.qmd b/docs/dataset-formats/template_free.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..5087d6a01396b3fde34897ee1f04acc00798f42f
--- /dev/null
+++ b/docs/dataset-formats/template_free.qmd
@@ -0,0 +1,7 @@
+---
+title: Template-Free
+description: Construct prompts without a template.
+order: 4
+---
+
+See [these docs](../input_output.qmd).
diff --git a/docs/dataset-formats/tokenized.qmd b/docs/dataset-formats/tokenized.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..8991a2110948dc461aefcb86bcf462fe085ec926
--- /dev/null
+++ b/docs/dataset-formats/tokenized.qmd
@@ -0,0 +1,12 @@
+---
+title: Custom Pre-Tokenized Dataset
+description: How to use a custom pre-tokenized dataset.
+order: 5
+---
+
+- Do not pass a `type:` in your axolotl config.
+- Columns in Dataset must be exactly `input_ids`, `attention_mask`, `labels`
+
+```{.yaml filename="config.yml"}
+- path: ...
+```
diff --git a/docs/dataset_preprocessing.qmd b/docs/dataset_preprocessing.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..c99fce444e645be894bfe0090e19fb027ebff017
--- /dev/null
+++ b/docs/dataset_preprocessing.qmd
@@ -0,0 +1,35 @@
+---
+title: Dataset Preprocessing
+description: How datasets are processed
+---
+
+Dataset pre-processing is the step where Axolotl takes each dataset you've configured alongside
+the (dataset format)[../dataset-formats/] and prompt strategies to:
+ - parse the dataset based on the *dataset format*
+ - transform the dataset to how you would interact with the model based on the *prompt strategy*
+ - tokenize the dataset based on the configured model & tokenizer
+ - shuffle and merge multiple datasets together if using more than one
+
+The processing of the datasets can happen one of two ways:
+
+1. Before kicking off training by calling `python -m axolotl.cli.preprocess /path/to/your.yaml --debug`
+2. When training is started
+
+What are the benefits of pre-processing? When training interactively or for sweeps
+(e.g. you are restarting the trainer often), processing the datasets can oftentimes be frustratingly
+slow. Pre-processing will cache the tokenized/formatted datasets according to a hash of dependent
+training parameters so that it will intelligently pull from its cache when possible.
+
+The path of the cache is controlled by `dataset_prepared_path:` and is often left blank in example
+YAMLs as this leads to a more robust solution that prevents unexpectedly reusing cached data.
+
+If `dataset_prepared_path:` is left empty, when training, the processed dataset will be cached in a
+default path of `./last_run_prepared/`, but will ignore anything already cached there. By explicitly
+setting `dataset_prepared_path: ./last_run_prepared`, the trainer will use whatever pre-processed
+data is in the cache.
+
+What are the edge cases? Let's say you are writing a custom prompt strategy or using a user-defined
+prompt template. Because the trainer cannot readily detect these changes, we cannot change the
+calculated hash value for the pre-processed dataset. If you have `dataset_prepared_path: ...` set
+and change your prompt templating logic, it may not pick up the changes you made and you will be
+training over the old prompt.
diff --git a/docs/debugging.qmd b/docs/debugging.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..7237fbd6f22f592f0ff58d9941685926817fe738
--- /dev/null
+++ b/docs/debugging.qmd
@@ -0,0 +1,245 @@
+---
+title: Debugging
+description: How to debug Axolotl
+---
+
+
+This document provides some tips and tricks for debugging Axolotl. It also provides an example configuration for debugging with VSCode. A good debugging setup is essential to understanding how Axolotl code works behind the scenes.
+
+## Table of Contents
+
+- [General Tips](#general-tips)
+- [Debugging with VSCode](#debugging-with-vscode)
+ - [Background](#background)
+ - [Configuration](#configuration)
+ - [Customizing your debugger](#customizing-your-debugger)
+ - [Video Tutorial](#video-tutorial)
+- [Debugging With Docker](#debugging-with-docker)
+ - [Setup](#setup)
+ - [Attach To Container](#attach-to-container)
+ - [Video - Attaching To Docker On Remote Host](#video---attaching-to-docker-on-remote-host)
+
+## General Tips
+
+While debugging it's helpful to simplify your test scenario as much as possible. Here are some tips for doing so:
+
+> [!Important]
+> All of these tips are incorporated into the [example configuration](#configuration) for debugging with VSCode below.
+
+1. **Make sure you are using the latest version of axolotl**: This project changes often and bugs get fixed fast. Check your git branch and make sure you have pulled the latest changes from `main`.
+1. **Eliminate concurrency**: Restrict the number of processes to 1 for both training and data preprocessing:
+ - Set `CUDA_VISIBLE_DEVICES` to a single GPU, ex: `export CUDA_VISIBLE_DEVICES=0`.
+ - Set `dataset_processes: 1` in your axolotl config or run the training command with `--dataset_processes=1`.
+2. **Use a small dataset**: Construct or use a small dataset from HF Hub. When using a small dataset, you will often have to make sure `sample_packing: False` and `eval_sample_packing: False` to avoid errors. If you are in a pinch and don't have time to construct a small dataset but want to use from the HF Hub, you can shard the data (this will still tokenize the entire dataset, but will only use a fraction of the data for training. For example, to shard the dataset into 20 pieces, add the following to your axolotl config):
+ ```yaml
+ dataset:
+ ...
+ shards: 20
+ ```
+3. **Use a small model**: A good example of a small model is [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0).
+4. **Minimize iteration time**: Make sure the training loop finishes as fast as possible, with these settings.
+ - `micro_batch_size: 1`
+ - `max_steps: 1`
+ - `val_set_size: 0`
+5. **Clear Caches:** Axolotl caches certain steps and so does the underlying HuggingFace trainer. You may want to clear some of these caches when debugging.
+ - Data preprocessing: When debugging data preprocessing, which includes prompt template formation, you may want to delete the directory set in `dataset_prepared_path:` in your axolotl config. If you didn't set this value, the default is `last_run_prepared`.
+ - HF Hub: If you are debugging data preprocessing, you should clear the relevant HF cache [HuggingFace cache](https://huggingface.co/docs/datasets/cache), by deleting the appropriate `~/.cache/huggingface/datasets/...` folder(s).
+ - **The recommended approach is to redirect all outputs and caches to a temporary folder and delete selected subfolders before each run. This is demonstrated in the example configuration below.**
+
+
+## Debugging with VSCode
+
+### Background
+
+The below example shows how to configure VSCode to debug data preprocessing of the `sharegpt` format. This is the format used when you have the following in your axolotl config:
+
+```yaml
+datasets:
+ - path: # example on HF Hub: philschmid/guanaco-sharegpt-style
+ type: sharegpt
+```
+
+>[!Important]
+> If you are already familiar with advanced VSCode debugging, you can skip the below explanation and look at the files [.vscode/launch.json](../.vscode/launch.json) and [.vscode/tasks.json](../.vscode/tasks.json) for an example configuration.
+
+>[!Tip]
+> If you prefer to watch a video, rather than read, you can skip to the [video tutorial](#video-tutorial) below (but doing both is recommended).
+
+### Setup
+
+Make sure you have an [editable install](https://setuptools.pypa.io/en/latest/userguide/development_mode.html) of Axolotl, which ensures that changes you make to the code are reflected at runtime. Run the following commands from the root of this project:
+
+```bash
+pip3 install packaging
+pip3 install -e '.[flash-attn,deepspeed]'
+```
+
+#### Remote Hosts
+
+If you developing on a remote host, you can easily use VSCode to debug remotely. To do so, you will need to follow this [remote - SSH guide](https://code.visualstudio.com/docs/remote/ssh). You can also see the video below on [Docker and Remote SSH debugging](#video---attaching-to-docker-on-remote-host).
+
+
+### Configuration
+
+The easiest way to get started is to modify the [.vscode/launch.json](../.vscode/launch.json) file in this project. This is just an example configuration, so you may need to modify or copy it to suit your needs.
+
+For example, to mimic the command `cd devtools && CUDA_VISIBLE_DEVICES=0 accelerate launch -m axolotl.cli.train dev_sharegpt.yml`, you would use the below configuration[^1]. Note that we add additional flags that override the axolotl config and incorporate the tips above (see the comments). We also set the working directory to `devtools` and set the `env` variable `HF_HOME` to a temporary folder that is later partially deleted. This is because we want to delete the HF dataset cache before each run in order to ensure that the data preprocessing code is run from scratch.
+
+```jsonc
+// .vscode/launch.json
+{
+ "version": "0.2.0",
+ "configurations": [
+ {
+ "name": "Debug axolotl prompt - sharegpt",
+ "type": "python",
+ "module": "accelerate.commands.launch",
+ "request": "launch",
+ "args": [
+ "-m", "axolotl.cli.train", "dev_sharegpt.yml",
+ // The flags below simplify debugging by overriding the axolotl config
+ // with the debugging tips above. Modify as needed.
+ "--dataset_processes=1", // limits data preprocessing to one process
+ "--max_steps=1", // limits training to just one step
+ "--batch_size=1", // minimizes batch size
+ "--micro_batch_size=1", // minimizes batch size
+ "--val_set_size=0", // disables validation
+ "--sample_packing=False", // disables sample packing which is necessary for small datasets
+ "--eval_sample_packing=False",// disables sample packing on eval set
+ "--dataset_prepared_path=temp_debug/axolotl_outputs/data", // send data outputs to a temp folder
+ "--output_dir=temp_debug/axolotl_outputs/model" // send model outputs to a temp folder
+ ],
+ "console": "integratedTerminal", // show output in the integrated terminal
+ "cwd": "${workspaceFolder}/devtools", // set working directory to devtools from the root of the project
+ "justMyCode": true, // step through only axolotl code
+ "env": {"CUDA_VISIBLE_DEVICES": "0", // Since we aren't doing distributed training, we need to limit to one GPU
+ "HF_HOME": "${workspaceFolder}/devtools/temp_debug/.hf-cache"}, // send HF cache to a temp folder
+ "preLaunchTask": "cleanup-for-dataprep", // delete temp folders (see below)
+ }
+ ]
+}
+```
+
+**Additional notes about this configuration:**
+
+- The argument `justMyCode` is set to `true` such that you step through only the axolotl code. If you want to step into dependencies, set this to `false`.
+- The `preLaunchTask`: `cleanup-for-dataprep` is defined in [.vscode/tasks.json](../.vscode/tasks.json) and is used to delete the following folders before debugging, which is essential to ensure that the data pre-processing code is run from scratch:
+ - `./devtools/temp_debug/axolotl_outputs`
+ - `./devtools/temp_debug/.hf-cache/datasets`
+
+>[!Tip]
+> You may not want to delete these folders. For example, if you are debugging model training instead of data pre-processing, you may NOT want to delete the cache or output folders. You may also need to add additional tasks to the `tasks.json` file depending on your use case.
+
+Below is the [./vscode/tasks.json](../.vscode/tasks.json) file that defines the `cleanup-for-dataprep` task. This task is run before each debugging session when you use the above configuration. Note how there are two tasks that delete the two folders mentioned above. The third task `cleanup-for-dataprep` is a composite task that combines the two tasks. A composite task is necessary because VSCode does not allow you to specify multiple tasks in the `preLaunchTask` argument of the `launch.json` file.
+
+```jsonc
+// .vscode/tasks.json
+// this file is used by launch.json
+{
+ "version": "2.0.0",
+ "tasks": [
+ // this task changes into the devtools directory and deletes the temp_debug/axolotl_outputs folder
+ {
+ "label": "delete-outputs",
+ "type": "shell",
+ "command": "rm -rf temp_debug/axolotl_outputs",
+ "options":{ "cwd": "${workspaceFolder}/devtools"},
+ "problemMatcher": []
+ },
+ // this task changes into the devtools directory and deletes the `temp_debug/.hf-cache/datasets` folder
+ {
+ "label": "delete-temp-hf-dataset-cache",
+ "type": "shell",
+ "command": "rm -rf temp_debug/.hf-cache/datasets",
+ "options":{ "cwd": "${workspaceFolder}/devtools"},
+ "problemMatcher": []
+ },
+ // this task combines the two tasks above
+ {
+ "label": "cleanup-for-dataprep",
+ "dependsOn": ["delete-outputs", "delete-temp-hf-dataset-cache"],
+ }
+ ]
+}
+```
+
+### Customizing your debugger
+
+Your debugging use case may differ from the example above. The easiest thing to do is to put your own axolotl config in the `devtools` folder and modify the `launch.json` file to use your config. You may also want to modify the `preLaunchTask` to delete different folders or not delete anything at all.
+
+### Video Tutorial
+
+The following video tutorial walks through the above configuration and demonstrates how to debug with VSCode, (click the image below to watch):
+
+
+
+
+## Debugging With Docker
+
+Using [official Axolotl Docker images](https://hub.docker.com/r/winglian/axolotl/tags) is a great way to debug your code, and is a very popular way to use Axolotl. Attaching VSCode to Docker takes a few more steps.
+
+### Setup
+
+On the host that is running axolotl (ex: if you are using a remote host), clone the axolotl repo and change your current directory to the root:
+
+```bash
+git clone https://github.com/OpenAccess-AI-Collective/axolotl
+cd axolotl
+```
+
+>[!Tip]
+> If you already have axolotl cloned on your host, make sure you have the latest changes and change into the root of the project.
+
+Next, run the desired docker image and mount the current directory. Below is a docker command you can run to do this:[^2]
+
+```bash
+docker run --privileged --gpus '"all"' --shm-size 10g --rm -it --name axolotl --ipc=host --ulimit memlock=-1 --ulimit stack=67108864 --mount type=bind,src="${PWD}",target=/workspace/axolotl -v ${HOME}/.cache/huggingface:/root/.cache/huggingface winglian/axolotl:main-py3.10-cu118-2.0.1
+```
+
+>[!Tip]
+> To understand which containers are available, see the [Docker section of the README](../README.md#docker) and the [DockerHub repo](https://hub.docker.com/r/winglian/axolotl/tags). For details of how the Docker containers are built, see axolotl's [Docker CI builds](../.github/workflows/main.yml).
+
+You will now be in the container. Next, perform an editable install of Axolotl:
+
+```bash
+pip3 install packaging
+pip3 install -e '.[flash-attn,deepspeed]'
+```
+
+### Attach To Container
+
+Next, if you are using a remote host, [Remote into this host with VSCode](https://code.visualstudio.com/docs/remote/ssh). If you are using a local host, you can skip this step.
+
+Next, select `Dev Containers: Attach to Running Container...` using the command palette (`CMD + SHIFT + P`) in VSCode. You will be prompted to select a container to attach to. Select the container you just created. You will now be in the container with a working directory that is at the root of the project. Any changes you make to the code will be reflected both in the container and on the host.
+
+Now you are ready to debug as described above (see [Debugging with VSCode](#debugging-with-vscode)).
+
+### Video - Attaching To Docker On Remote Host
+
+Here is a short video that demonstrates how to attach to a Docker container on a remote host:
+
+
+
+
+[^1]: The config actually mimics the command `CUDA_VISIBLE_DEVICES=0 python -m accelerate.commands.launch -m axolotl.cli.train devtools/sharegpt.yml`, but this is the same thing.
+
+[^2]: Many of the below flags are recommended best practices by Nvidia when using nvidia-container-toolkit. You can read more about these flags [here](https://docs.nvidia.com/deeplearning/frameworks/user-guide/index.html).
diff --git a/docs/faq.qmd b/docs/faq.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..91413d24e9488385454e78b2562c76fc9fe80f3c
--- /dev/null
+++ b/docs/faq.qmd
@@ -0,0 +1,21 @@
+---
+title: FAQ
+description: Frequently asked questions
+---
+
+
+**Q: The trainer stopped and hasn't progressed in several minutes.**
+
+> A: Usually an issue with the GPUs communicating with each other. See the [NCCL doc](nccl.qmd)
+
+**Q: Exitcode -9**
+
+> A: This usually happens when you run out of system RAM.
+
+**Q: Exitcode -7 while using deepspeed**
+
+> A: Try upgrading deepspeed w: `pip install -U deepspeed`
+
+**Q: AttributeError: 'DummyOptim' object has no attribute 'step'**
+
+> A: You may be using deepspeed with single gpu. Please don't set `deepspeed:` in yaml or cli.
diff --git a/docs/fsdp_qlora.qmd b/docs/fsdp_qlora.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..7f12d44935305b2e817ce80788bf57f5e2b13b1e
--- /dev/null
+++ b/docs/fsdp_qlora.qmd
@@ -0,0 +1,43 @@
+---
+title: "FDSP + QLoRA"
+description: Use FSDP with QLoRA to fine-tune large LLMs on consumer GPUs.
+format:
+ html:
+ toc: true
+---
+
+## Background
+
+Using FSDP with QLoRA is essential for **fine-tuning larger (70b+ parameter) LLMs on consumer GPUs.** For example, you can use FSDP + QLoRA to train a 70b model on two 24GB GPUs[^1].
+
+Below, we describe how to use this feature in Axolotl.
+
+## Usage
+
+To enable `QLoRA` with `FSDP`, you need to perform the following steps:
+
+> ![Tip]
+> See the [example config](#example-config) file in addition to reading these instructions.
+
+1. Set `adapter: qlora` in your axolotl config file.
+2. Enable FSDP in your axolotl config, as [described here](https://github.com/OpenAccess-AI-Collective/axolotl?tab=readme-ov-file#fsdp).
+3. Use one of the supported model types: `llama`, `mistral` or `mixtral`.
+
+## Example Config
+
+[examples/llama-2/qlora-fsdp.yml](../examples/llama-2/qlora-fsdp.yml) contains an example of how to enable QLoRA + FSDP in axolotl.
+
+## References
+
+- [PR #1378](https://github.com/OpenAccess-AI-Collective/axolotl/pull/1378) enabling QLoRA in FSDP in Axolotl.
+- [Blog Post](https://www.answer.ai/posts/2024-03-06-fsdp-qlora.html) from the [Answer.AI](https://www.answer.ai/) team describing the work that enabled QLoRA in FSDP.
+- Related HuggingFace PRs Enabling FDSP + QLoRA:
+ - Accelerate [PR#2544](https://github.com/huggingface/accelerate/pull/2544 )
+ - Transformers [PR#29587](https://github.com/huggingface/transformers/pull/29587)
+ - TRL [PR#1416](https://github.com/huggingface/trl/pull/1416)
+ - PEFT [PR#1550](https://github.com/huggingface/peft/pull/1550)
+
+
+
+
+[^1]: This was enabled by [this work](https://www.answer.ai/posts/2024-03-06-fsdp-qlora.html) from the Answer.AI team.
diff --git a/docs/images/4d-mask.png b/docs/images/4d-mask.png
new file mode 100644
index 0000000000000000000000000000000000000000..36c8b05f61737da7203b6a855f5f25045bec7d78
--- /dev/null
+++ b/docs/images/4d-mask.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:70ddbbb087016350f54d2c4f2ca652ae3b8d838d70044d2bd219d4fc389b9e04
+size 244403
diff --git a/docs/input_output.qmd b/docs/input_output.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..3762901b3188f6548c36d8af4a7ecec53f1625ca
--- /dev/null
+++ b/docs/input_output.qmd
@@ -0,0 +1,265 @@
+---
+title: Template-free prompt construction
+description: "Template-free prompt construction with the `input_output` format"
+---
+
+
+
+- [Background](#background)
+ - [Masking Inputs](#masking-inputs)
+ - [You may not want prompt templates](#you-may-not-want-prompt-templates)
+ - [The `input_output` format](#the-input_output-format)
+- [Usage](#usage)
+ - [1. Prepare Data](#1-prepare-data)
+ - [2. Use `type: input_output`](#2-use-type-input_output)
+ - [3. Check the prompts](#3-check-the-prompts)
+
+
+
+
+
+## Background
+
+
+
+### Masking Inputs
+
+One of the most popular features of
+[axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) is
+setting the following configuration value:
+
+
+```yaml
+train_on_inputs: false
+```
+
+If you declare a [dataset formats](https://github.com/OpenAccess-AI-Collective/axolotl?tab=readme-ov-file#dataset)
+such as `alpaca` or `chatml`, axolotl knows what is an input
+(i.e. human) vs. an output (i.e. the assistant) and masks the input
+labels so that your model can focus on predicting the outputs only.
+
+
+
+### You may not want prompt templates
+
+However, there are many situations where you don't want to use one of
+these formats or templates. This is because they can:
+
+- Add unnecessary boilerplate to your prompts.
+- Create artifacts like special delimiters `<|im_start|>` that can
+ quickly become footguns if you don't include them correctly at
+ inference time.
+- Enforce a *chat* interface when you do not want one. Sometimes you
+ just want to fine-tune a model to a very specific task and do NOT
+ want multi-turn conversations, roles, etc.
+- Limit you to only certain roles that the template allows.
+
+
+
+### The `input_output` format
+
+You can construct your prompts without a template by using the
+`input_output` format, by setting `type: input_output` in your
+configuration file like this:
+
+**config.yml**
+
+```yaml
+train_on_inputs: false # Mask segments of your data
+datasets:
+ - path: output.jsonl
+ type: input_output # use template free prompt construction
+```
+
+Unlike `type: completion`, which is also template-free,
+`type: input_output` allows you to mask segments of your text. More
+details on how this works are described below.
+
+
+
+## Usage
+
+This is how you can use the `input_output` format:
+
+
+
+### 1. Prepare Data
+
+To use the `input_output` format, collect your data in the following
+format into a jsonl file (below is the first row from the file
+`output`.jsonl` pretty printed):
+
+```bash
+$ head -n1 output.jsonl | python -m json.tool
+```
+
+:::{.cell-output .cell-output-stdout}
+ {
+ "segments": [
+ {
+ "label": true,
+ "text": "Hello\n"
+ },
+ {
+ "label": true,
+ "text": "hi there!. "
+ },
+ {
+ "label": false,
+ "text": "goodbye "
+ },
+ {
+ "label": true,
+ "text": "farewell "
+ }
+ ]
+ }
+:::
+
+Set `label:false` when you want to mask a segment of text so that the
+model isn't trained on it. Some things to keep in mind:
+
+> [!IMPORTANT]
+> 1. **EOS, BOS, spaces, newlines etc. are entirely up to you. Axolotl
+ concatenates all the segments as-is.** The tokenizer doesn't add
+ anything additional. Notice how I added spaces, newlines, ``
+ (BOS), and ` ` (EOS) myself.
+> 2. Make sure you check the materialized output to validate that the
+ prompt is getting assembled how you like.
+
+
+
+### 2. Use `type: input_output`
+
+Let's materialize data with our `output.jsonl` file by setting
+`type: input_output` in our axolotl config:
+
+```yaml
+# training_config.yaml
+base_model: mistralai/Mistral-7B-v0.1
+data_seed: 49
+seed: 49
+
+datasets:
+ - path: output.jsonl
+ type: input_output
+val_set_size: 0.1
+
+sequence_len: 896
+sample_packing: false
+
+micro_batch_size: 2
+gradient_accumulation_steps: 3
+eval_batch_size: 2
+num_epochs: 1
+learning_rate: 0.0002
+
+train_on_inputs: false
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
+```
+
+You can use the following command to materialize your data. The
+`--debug` flag will print the tokens, along with the labels so you can
+verify that the correct items are being ignored:
+
+```bash
+$ python -m axolotl.cli.preprocess training_config.yaml --debug
+
+...
+[2024-03-05 23:36:46,969] [INFO] [axolotl.check_example_labels:35] [PID:607731] [RANK:0] (1, 1) Hello(22557, 22557)
+(13, 13) hi(12014, 12014) there(736, 736) !(28808, 28808) .(28723, 28723) (28705, 28705) good(-100, 1179) bye(-100, 17664) (-100, 28705) fare(19111, 19111) well(5458, 5458) (2, 2)
+
+```
+
+The format is `decoded_token`(`label`, `token_id`), for example,
+`(1, 1)` means that the token is ``, the label is `1` and the
+token_id is `1`. When the label is `-100` then that token is ignored for
+training.
+
+
+
+### 3. Check the prompts
+
+Here is another way to check the materialized output:
+
+```python
+from transformers import AutoTokenizer
+from datasets import load_from_disk
+import yaml
+
+directory = !ls last_run_prepared/
+with open('training_config.yaml', 'r') as f:
+ cfg = yaml.safe_load(f)
+model_id = cfg['base_model']
+tok = AutoTokenizer.from_pretrained(model_id)
+ds = load_from_disk(f'last_run_prepared/{directory[0]}/')
+```
+
+```python
+>>> row = ds[0]
+>>> print(tok.decode(row['input_ids']))
+ Hello
+ hi there!. goodbye farewell
+```
+
+We can check that the right tokens are ingored by comparing the labels
+to each token:
+
+```python
+import pandas as pd
+pd.DataFrame([{'token': tok.decode(i), 'label': l, 'id':i} for i,l in
+ zip(row['input_ids'], row['labels'])])
+```
+
+| token | label | id |
+|-------|-------|-------|
+| 0 | \ | 1 |
+| 1 | Hello | 22557 |
+| 2 | \\n | 13 |
+| 3 | hi | 12014 |
+| 4 | there | 736 |
+| 5 | ! | 28808 |
+| 6 | . | 28723 |
+| 7 | | 28705 |
+| 8 | good | -100 |
+| 9 | bye | -100 |
+| 10 | | -100 |
+| 11 | fare | 19111 |
+| 12 | well | 5458 |
+| 13 | \ | 2 |
+
+
+
+If we look at the input data, the above table seems correct! (The jsonl
+version is repeated below for reference):
+
+
+```bash
+$ head -n1 output.jsonl | python -m json.tool
+```
+
+:::{.cell-output .cell-output-stdout}
+ {
+ "segments": [
+ {
+ "label": true,
+ "text": "Hello\n"
+ },
+ {
+ "label": true,
+ "text": "hi there!. "
+ },
+ {
+ "label": false,
+ "text": "goodbye "
+ },
+ {
+ "label": true,
+ "text": "farewell "
+ }
+ ]
+ }
+:::
diff --git a/docs/mac.qmd b/docs/mac.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..2a83035381d8534b53917d0981d7269addec4f70
--- /dev/null
+++ b/docs/mac.qmd
@@ -0,0 +1,22 @@
+---
+title: Mac M-series
+description: Mac M-series support
+---
+
+Currently Axolotl on Mac is partially usable, many of the dependencies of Axolotl including Pytorch do not support MPS or have incomplete support.
+
+Current support:
+
+- [x] Support for all models
+- [x] Full training of models
+- [x] LoRA training
+- [x] Sample packing
+- [ ] FP16 and BF16 (awaiting AMP support for MPS in Pytorch)
+- [ ] Tri-dao's flash-attn (until it is supported use spd_attention as an alternative)
+- [ ] xformers
+- [ ] bitsandbytes (meaning no 4/8 bits loading and bnb optimizers)
+- [ ] qlora
+- [ ] DeepSpeed
+
+Untested:
+- FSDP
diff --git a/docs/multi-node.qmd b/docs/multi-node.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..5c6fa976b991c5868efeb3daaf3156a43482517b
--- /dev/null
+++ b/docs/multi-node.qmd
@@ -0,0 +1,48 @@
+---
+title: Multi Node
+description: How to use Axolotl on multiple machines
+---
+
+You will need to create a configuration for accelerate, either by using `accelerate config` and follow the instructions or you can use one of the preset below:
+
+~/.cache/huggingface/accelerate/default_config.yaml
+```yaml
+compute_environment: LOCAL_MACHINE
+debug: false
+distributed_type: FSDP
+downcast_bf16: 'no'
+machine_rank: 0 # Set to 0 for the main machine, increment by one for other machines
+main_process_ip: 10.0.0.4 # Set to main machine's IP
+main_process_port: 5000
+main_training_function: main
+mixed_precision: bf16
+num_machines: 2 # Change to the number of machines
+num_processes: 4 # That's the total number of GPUs, (for example: if you have 2 machines with 4 GPU, put 8)
+rdzv_backend: static
+same_network: true
+tpu_env: []
+tpu_use_cluster: false
+tpu_use_sudo: false
+use_cpu: false
+```
+
+Configure your model to use FSDP with for example:
+```yaml
+fsdp:
+ - full_shard
+ - auto_wrap
+fsdp_config:
+ fsdp_offload_params: true
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
+```
+
+## Machine configuration
+
+On each machine you need a copy of Axolotl, we suggest using the same commit to ensure compatibility.
+
+You will also need to have the same configuration file for your model on each machine.
+
+On the main machine only, make sure the port you set as `main_process_port` is open in TCP and reachable by other machines.
+
+All you have to do now is launch using accelerate as you would usually do on each machine and voila, the processes will start once you have launched accelerate on every machine.
diff --git a/docs/multipack.qmd b/docs/multipack.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..097bcd2e503c2e35f637824002a68682fd52979f
--- /dev/null
+++ b/docs/multipack.qmd
@@ -0,0 +1,76 @@
+---
+title: Multipack (Sample Packing)
+description: Multipack is a technique to pack multiple sequences into a single batch to increase training throughput.
+---
+
+## Visualization of Multipack with Flash Attention
+
+Because Flash Attention simply drops the attention mask, we do not need to
+construct a 4d attention mask. We only need to concatenate the sequences into
+a single batch and let flash attention know where each new sequence begins.
+
+
+4k context, bsz =4,
+each character represents 256 tokens
+X represents a padding token
+
+```
+ 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5
+[[ A A A A A A A A A A A ]
+ B B B B B B ]
+ C C C C C C C ]
+ D D D D ]]
+
+[[ E E E E E E E E ]
+ [ F F F F ]
+ [ G G G ]
+ [ H H H H ]]
+
+[[ I I I ]
+ [ J J J ]
+ [ K K K K K]
+ [ L L L ]]
+```
+
+after padding to longest input in each step
+```
+ 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5
+[[ A A A A A A A A A A A ]
+ B B B B B B X X X X X X ]
+ C C C C C C C X X X X ]
+ D D D D X X X X X X X ]]
+
+[[ E E E E E E E E ]
+ [ F F F F X X X X ]
+ [ G G G X X X X X ]
+ [ H H H H X X X X ]]
+
+[[ I I I X X ]
+ [ J J J X X ]
+ [ K K K K K ]
+ [ L L L X X ]]
+```
+
+w packing ( note it's the same effective number of tokens per step, but a true bsz of 1)
+```
+ 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5
+[[ A A A A A A A A A A A B B B B B
+ B C C C C C C C D D D D E E E E
+ E E E E F F F F F G G G H H H H
+ I I I J J J J K K K K K L L L X ]]
+```
+
+cu_seqlens:
+[[ 0, 11, 17, 24, 28, 36, 41 44, 48, 51, 55, 60, 64]]
+
+
+## Multipack without Flash Attention
+
+Multipack can still be achieved without Flash attention, but with lower packing
+efficiency as we are not able to join multiple batches into a single batch due to
+context length limits without flash attention. We can use either Pytorch's Scaled
+Dot Product Attention implementation or native Pytorch attention implementation
+along with [4d attention masks](https://github.com/huggingface/transformers/pull/27539)
+to pack sequences together and avoid cross attention.
+
+
diff --git a/docs/nccl.qmd b/docs/nccl.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..3b616aa66526e1d18cc3b292a7261c6470ce6540
--- /dev/null
+++ b/docs/nccl.qmd
@@ -0,0 +1,49 @@
+---
+title: NCCL
+description: Troubleshooting NCCL issues
+---
+
+NVIDIA NCCL is a library to facilitate and optimize multi-GPU communication operations, such as broadcast, all-gather, reduce, all-reduce, etc. Broadly, NCCL configuration is highly environment-specific and is configured via several [environment variables](https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/env.html). A common NCCL-related problem occurs when a long-running operation times out causing the training process to abort:
+
+```text
+Watchdog caught collective operation timeout: WorkNCCL(SeqNum=42, OpType=ALLGATHER, Timeout(ms)=1800000) ran for 1806948 milliseconds before timing out.
+```
+
+Often, this timeout will happen after 30 minutes (the default setting) and is accompanied by below-average power consumption with near 100% GPU utilization before the error is raised. Nvidia recommends [disabling PCI access control services (ACS)](https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/troubleshooting.html#pci-access-control-services-acs) as a possible solution if this is available to you.
+
+Forcing cross-GPU communication via [NVLink](https://en.wikipedia.org/wiki/NVLink) may help without increasing timeouts. To verify that your configuration is leveraging NVLink run the following command:
+
+```shell
+nvidia-smi nvlink --status
+```
+
+To force NCCL to use NVLink, simply set this in the environment:
+
+```shell
+export NCCL_P2P_LEVEL=NVL
+```
+
+If NVLink is not available in your environment there are other options for ``NCCL_P2P_LEVEL`` in the table below:
+
+| NCCL_P2P_LEVEL | Description |
+| -------------- | ----------- |
+| PIX | P2P data transfers through no more than a single PCIe bridge. Faster data transfer rates vs to paths involving multiple bridges, but slower compared to direct GPU-to-GPU communication. |
+| PXB | P2P data transfers through multiple PCIe bridges but not going through the PCIe Host Bridge; this path involves a complex routing process, potentially incurring a moderate level of latency. |
+| PHB | P2P data transfers occur over the PCIe and through a PCIe Host Bridge, typically involving the CPU, which can facilitate direct memory access but might introduce additional latency compared to more direct paths (ex PIX, NVL) |
+
+To validate that acceptable data transfer speeds exist for your training job, running [NCCL Tests](https://github.com/NVIDIA/nccl-tests/blob/master/README.md) can help pinpoint bottlenecks, for example:
+
+```shell
+./build/all_reduce_perf -b 8 -e 128M -f 2 -g 3
+```
+
+It can be useful when debugging NCCL communication timeouts to activate additional logging in both PyTorch and NCCL:
+
+```shell
+export NCCL_DEBUG=INFO
+export NCCL_DEBUG_SUBSYS=ALL
+export TORCH_DISTRIBUTED_DEBUG=INFO
+export TORCHELASTIC_ERROR_FILE=/PATH/TO/torcherror.log
+```
+
+Finally, if you believe your training job needs more time you can increase the timeout past 30 minutes by setting the ``ddp_timeout`` value in the Axolotl configuration. See [PyTorch init_process_group](https://pytorch.org/docs/stable/distributed.html#torch.distributed.init_process_group) for documentation on this value.
diff --git a/docs/rlhf.qmd b/docs/rlhf.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..b8b2bded09a6c67a9609218db8b6aa4f394cbbe0
--- /dev/null
+++ b/docs/rlhf.qmd
@@ -0,0 +1,72 @@
+---
+title: "RLHF (Beta)"
+description: "Reinforcement Learning from Human Feedback is a method whereby a language model is optimized from data using human feedback."
+---
+
+### Overview
+
+Reinforcement Learning from Human Feedback is a method whereby a language model is optimized from data using human
+feedback. Various methods include, but not limited to:
+
+- Proximal Policy Optimization (PPO) (not yet supported in axolotl)
+- Direct Preference Optimization (DPO)
+- Identity Preference Optimization (IPO)
+
+
+### RLHF using Axolotl
+
+>[!IMPORTANT]
+>This is a BETA feature and many features are not fully implemented. You are encouraged to open new PRs to improve the integration and functionality.
+
+The various RL training methods are implemented in trl and wrapped via axolotl. Below are various examples with how you can use various preference datasets to train models that use ChatML
+
+#### DPO
+```yaml
+rl: dpo
+datasets:
+ - path: Intel/orca_dpo_pairs
+ split: train
+ type: chatml.intel
+ - path: argilla/ultrafeedback-binarized-preferences
+ split: train
+ type: chatml.argilla
+```
+
+#### IPO
+```yaml
+rl: ipo
+```
+
+#### ORPO
+
+Paper: https://arxiv.org/abs/2403.07691
+
+```yaml
+rl: orpo
+orpo_alpha: 0.1
+remove_unused_columns: false
+
+chat_template: chatml
+datasets:
+ - path: argilla/ultrafeedback-binarized-preferences-cleaned
+ type: chat_template.argilla
+```
+
+#### Using local dataset files
+```yaml
+datasets:
+ - ds_type: json
+ data_files:
+ - orca_rlhf.jsonl
+ split: train
+ type: chatml.intel
+```
+
+#### Trl autounwrap for peft
+
+Trl supports autounwrapping peft models, so that a ref model does not need to be additionally loaded, leading to less VRAM needed. This is on by default. To turn it off, pass the following config.
+
+```yaml
+# load ref model when adapter training.
+rl_adapter_ref_model: true
+```
diff --git a/examples/cerebras/btlm-ft.yml b/examples/cerebras/btlm-ft.yml
new file mode 100644
index 0000000000000000000000000000000000000000..ba4e65daaef2149c3269d85a8eb3ef3ecb7a14ff
--- /dev/null
+++ b/examples/cerebras/btlm-ft.yml
@@ -0,0 +1,89 @@
+base_model: cerebras/btlm-3b-8k-base
+model_type: AutoModelForCausalLM
+tokenizer_type: GPT2Tokenizer
+trust_remote_code: true
+tokenizer_use_fast: true
+tokenizer_legacy: true
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+push_dataset_to_hub:
+hf_use_auth_token: true
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path: last_prepared_run
+val_set_size: 0.05
+
+adapter:
+lora_model_dir:
+sequence_len: 2048
+max_packed_sequence_len:
+sample_packing: false
+sample_packing_eff_est:
+sample_packing_seq_len_multiplier:
+total_num_tokens:
+
+lora_r:
+lora_alpha:
+lora_dropout:
+lora_target_modules:
+lora_target_linear:
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+output_dir: ./outputs/btlm-out
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 1
+optimizer: adamw_torch
+adam_beta2: 0.95
+adam_eps: 0.000000001
+max_grad_norm: 1.0
+
+torchdistx_path:
+lr_scheduler: cosine
+lr_quadratic_warmup: true
+learning_rate: 0.000085
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+
+gradient_checkpointing: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+
+xformers_attention:
+flash_attention: true
+sdp_attention:
+flash_optimum:
+
+gptq_groupsize:
+gptq_model_v1:
+
+warmup_steps: 32
+evals_per_epoch: 4
+saves_per_epoch: 1
+save_total_limit:
+
+debug:
+deepspeed:
+weight_decay: 0.1
+special_tokens:
+ pad_token: "<|endoftext|>"
+fsdp:
+# - full_shard
+# - auto_wrap
+fsdp_config:
+# fsdp_state_dict_type: FULL_STATE_DICT
+# fsdp_transformer_layer_cls_to_wrap: BTLMBlock
diff --git a/examples/cerebras/qlora.yml b/examples/cerebras/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..285607a4c81aff6141d3cca982c6810f48e52d35
--- /dev/null
+++ b/examples/cerebras/qlora.yml
@@ -0,0 +1,59 @@
+base_model: cerebras/Cerebras-GPT-1.3B
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+push_dataset_to_hub:
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+adapter: qlora
+lora_model_dir:
+sequence_len: 2048
+lora_r: 16
+lora_alpha: 32
+lora_dropout: 0.05
+lora_target_modules:
+ - c_fc
+ - c_attn
+ - c_proj
+lora_target_linear:
+lora_fan_in_fan_out:
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/qlora-out
+batch_size: 4
+micro_batch_size: 4
+num_epochs: 2
+optimizer: paged_adamw_8bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.0002
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention: true
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+special_tokens:
+ pad_token: "<|endoftext|>"
diff --git a/examples/code-llama/13b/lora.yml b/examples/code-llama/13b/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..0ba96cfaa7410a004d3afd043c87b42c27682a4a
--- /dev/null
+++ b/examples/code-llama/13b/lora.yml
@@ -0,0 +1,67 @@
+base_model: codellama/CodeLlama-13b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: CodeLlamaTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+s2_attention:
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/code-llama/13b/qlora.yml b/examples/code-llama/13b/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..787862d0102b944023fcf12b6aa14da9eee701a3
--- /dev/null
+++ b/examples/code-llama/13b/qlora.yml
@@ -0,0 +1,68 @@
+base_model: codellama/CodeLlama-13b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: CodeLlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/qlora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: paged_adamw_32bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/code-llama/34b/lora.yml b/examples/code-llama/34b/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..92d4c544a3fedb426f567a3d4bb634c8ef96108d
--- /dev/null
+++ b/examples/code-llama/34b/lora.yml
@@ -0,0 +1,67 @@
+base_model: codellama/CodeLlama-34b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: CodeLlamaTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+s2_attention:
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/code-llama/34b/qlora.yml b/examples/code-llama/34b/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..93a6de877778e51b9a20c986d43fe3575e17155d
--- /dev/null
+++ b/examples/code-llama/34b/qlora.yml
@@ -0,0 +1,68 @@
+base_model: codellama/CodeLlama-34b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: CodeLlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/qlora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: paged_adamw_32bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/code-llama/7b/lora.yml b/examples/code-llama/7b/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..d13f505325d30efa186c70b635a1eef5233e914f
--- /dev/null
+++ b/examples/code-llama/7b/lora.yml
@@ -0,0 +1,67 @@
+base_model: codellama/CodeLlama-7b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: CodeLlamaTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+s2_attention:
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/code-llama/7b/qlora.yml b/examples/code-llama/7b/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..a1026a982d04e37d56900db0848fcd627d202987
--- /dev/null
+++ b/examples/code-llama/7b/qlora.yml
@@ -0,0 +1,68 @@
+base_model: codellama/CodeLlama-7b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: CodeLlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/qlora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: paged_adamw_32bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/code-llama/README.md b/examples/code-llama/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..a5011e347283ab67b2f276de85e7fba4344bf09f
--- /dev/null
+++ b/examples/code-llama/README.md
@@ -0,0 +1,22 @@
+# Overview
+
+This is an example of CodeLLaMA configuration for 7b, 13b and 34b.
+
+The 7b variant fits on any 24GB VRAM GPU and will take up about 17 GB of VRAM during training if using qlora and 20 GB if using lora. On a RTX 4090 it trains 3 epochs of the default dataset in about 15 minutes.
+
+The 13b variant will fit if you change these settings to these values:
+gradient_accumulation_steps: 2
+micro_batch_size: 1
+
+The 34b variant does not fit on 24GB of VRAM - you will need something with +40 gb VRAM that also supports flash attention v2 - A6000 or A100 are good choices.
+
+```shell
+accelerate launch scripts/finetune.py examples/code-llama/[MODEL_SIZE]/qlora.yml
+
+```
+or
+
+```shell
+accelerate launch scripts/finetune.py examples/code-llama/[MODEL_SIZE]/lora.yml
+
+```
diff --git a/examples/colab-notebooks/colab-axolotl-example.ipynb b/examples/colab-notebooks/colab-axolotl-example.ipynb
new file mode 100644
index 0000000000000000000000000000000000000000..fc3b76194961ec170798018369cb149f5c8fa3cc
--- /dev/null
+++ b/examples/colab-notebooks/colab-axolotl-example.ipynb
@@ -0,0 +1,216 @@
+{
+ "cells": [
+ {
+ "cell_type": "markdown",
+ "metadata": {
+ "id": "AKjdG7tbTb-n"
+ },
+ "source": [
+ "# Example notebook for running Axolotl on google colab"
+ ]
+ },
+ {
+ "cell_type": "code",
+ "execution_count": null,
+ "metadata": {
+ "id": "RcbNpOgWRcii"
+ },
+ "outputs": [],
+ "source": [
+ "import torch\n",
+ "# Check so there is a gpu available, a T4(free tier) is enough to run this notebook\n",
+ "assert (torch.cuda.is_available()==True)"
+ ]
+ },
+ {
+ "cell_type": "markdown",
+ "metadata": {
+ "id": "h3nLav8oTRA5"
+ },
+ "source": [
+ "## Install Axolotl and dependencies"
+ ]
+ },
+ {
+ "cell_type": "code",
+ "execution_count": null,
+ "metadata": {
+ "colab": {
+ "base_uri": "https://localhost:8080/"
+ },
+ "id": "3c3yGAwnOIdi",
+ "outputId": "e3777b5a-40ef-424f-e181-62dfecd1dd01"
+ },
+ "outputs": [],
+ "source": [
+ "!pip install torch==\"2.1.2\"\n",
+ "!pip install -e git+https://github.com/OpenAccess-AI-Collective/axolotl#egg=axolotl\n",
+ "!pip install flash-attn==\"2.5.0\"\n",
+ "!pip install deepspeed==\"0.13.1\""
+ ]
+ },
+ {
+ "cell_type": "markdown",
+ "metadata": {
+ "id": "BW2MFr7HTjub"
+ },
+ "source": [
+ "## Create an yaml config file"
+ ]
+ },
+ {
+ "cell_type": "code",
+ "execution_count": null,
+ "metadata": {
+ "id": "9pkF2dSoQEUN"
+ },
+ "outputs": [],
+ "source": [
+ "import yaml\n",
+ "\n",
+ "# Your YAML string\n",
+ "yaml_string = \"\"\"\n",
+ "base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T\n",
+ "model_type: LlamaForCausalLM\n",
+ "tokenizer_type: LlamaTokenizer\n",
+ "is_llama_derived_model: true\n",
+ "\n",
+ "load_in_8bit: false\n",
+ "load_in_4bit: true\n",
+ "strict: false\n",
+ "\n",
+ "datasets:\n",
+ " - path: mhenrichsen/alpaca_2k_test\n",
+ " type: alpaca\n",
+ "dataset_prepared_path:\n",
+ "val_set_size: 0.05\n",
+ "output_dir: ./outputs/qlora-out\n",
+ "\n",
+ "adapter: qlora\n",
+ "lora_model_dir:\n",
+ "\n",
+ "sequence_len: 1096\n",
+ "sample_packing: true\n",
+ "pad_to_sequence_len: true\n",
+ "\n",
+ "lora_r: 32\n",
+ "lora_alpha: 16\n",
+ "lora_dropout: 0.05\n",
+ "lora_target_modules:\n",
+ "lora_target_linear: true\n",
+ "lora_fan_in_fan_out:\n",
+ "\n",
+ "wandb_project:\n",
+ "wandb_entity:\n",
+ "wandb_watch:\n",
+ "wandb_name:\n",
+ "wandb_log_model:\n",
+ "\n",
+ "mlflow_experiment_name: colab-example\n",
+ "\n",
+ "gradient_accumulation_steps: 1\n",
+ "micro_batch_size: 1\n",
+ "num_epochs: 4\n",
+ "max_steps: 20\n",
+ "optimizer: paged_adamw_32bit\n",
+ "lr_scheduler: cosine\n",
+ "learning_rate: 0.0002\n",
+ "\n",
+ "train_on_inputs: false\n",
+ "group_by_length: false\n",
+ "bf16: false\n",
+ "fp16: true\n",
+ "tf32: false\n",
+ "\n",
+ "gradient_checkpointing: true\n",
+ "early_stopping_patience:\n",
+ "resume_from_checkpoint:\n",
+ "local_rank:\n",
+ "logging_steps: 1\n",
+ "xformers_attention:\n",
+ "flash_attention: false\n",
+ "\n",
+ "warmup_steps: 10\n",
+ "evals_per_epoch:\n",
+ "saves_per_epoch:\n",
+ "debug:\n",
+ "deepspeed:\n",
+ "weight_decay: 0.0\n",
+ "fsdp:\n",
+ "fsdp_config:\n",
+ "special_tokens:\n",
+ "\n",
+ "\"\"\"\n",
+ "\n",
+ "# Convert the YAML string to a Python dictionary\n",
+ "yaml_dict = yaml.safe_load(yaml_string)\n",
+ "\n",
+ "# Specify your file path\n",
+ "file_path = 'test_axolotl.yaml'\n",
+ "\n",
+ "# Write the YAML file\n",
+ "with open(file_path, 'w') as file:\n",
+ " yaml.dump(yaml_dict, file)\n"
+ ]
+ },
+ {
+ "cell_type": "markdown",
+ "metadata": {
+ "id": "bidoj8YLTusD"
+ },
+ "source": [
+ "## Launch the training"
+ ]
+ },
+ {
+ "cell_type": "code",
+ "execution_count": null,
+ "metadata": {
+ "colab": {
+ "base_uri": "https://localhost:8080/"
+ },
+ "id": "ydTI2Jk2RStU",
+ "outputId": "d6d0df17-4b53-439c-c802-22c0456d301b"
+ },
+ "outputs": [],
+ "source": [
+ "# Buy using the ! the comand will be executed as a bash command\n",
+ "!accelerate launch -m axolotl.cli.train /content/test_axolotl.yaml"
+ ]
+ },
+ {
+ "cell_type": "markdown",
+ "metadata": {},
+ "source": [
+ "## Play with inference"
+ ]
+ },
+ {
+ "cell_type": "code",
+ "execution_count": null,
+ "metadata": {},
+ "outputs": [],
+ "source": [
+ "# Buy using the ! the comand will be executed as a bash command\n",
+ "!accelerate launch -m axolotl.cli.inference /content/test_axolotl.yaml \\\n",
+ " --qlora_model_dir=\"./qlora-out\" --gradio"
+ ]
+ }
+ ],
+ "metadata": {
+ "accelerator": "GPU",
+ "colab": {
+ "gpuType": "T4",
+ "provenance": []
+ },
+ "kernelspec": {
+ "display_name": "Python 3",
+ "name": "python3"
+ },
+ "language_info": {
+ "name": "python"
+ }
+ },
+ "nbformat": 4,
+ "nbformat_minor": 0
+}
diff --git a/examples/dbrx/16bit-lora.yaml b/examples/dbrx/16bit-lora.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..32b625ac6965b9accdd764e6b22143b8adbedfcb
--- /dev/null
+++ b/examples/dbrx/16bit-lora.yaml
@@ -0,0 +1,81 @@
+base_model: LnL-AI/dbrx-base-converted-v2
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.0
+output_dir: ./outputs/out
+
+sequence_len: 512
+sample_packing: false
+pad_to_sequence_len: false
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+adapter: lora
+lora_model_dir:
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+# w1, w2, & v1 will hang the trainer
+lora_target_modules:
+ - q_proj # attn
+ - k_proj # attn
+ - v_proj # attn
+ - out_proj # attn
+ - layer # router
+# - w1
+# - w2
+# - v1
+
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 1
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: false # don't use with fsdp_activation_checkpointing
+gradient_checkpointing_kwargs:
+ use_reentrant: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch:
+saves_per_epoch: 1
+debug:
+weight_decay: 0.0
+fsdp:
+ - full_shard
+ - auto_wrap
+fsdp_config:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: false
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: true
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+ fsdp_transformer_layer_cls_to_wrap: DbrxBlock
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_activation_checkpointing: true
diff --git a/examples/dbrx/8bit-lora.yaml b/examples/dbrx/8bit-lora.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..50ee0a016411dfdf2d36ac4bd828a54c4be660d5
--- /dev/null
+++ b/examples/dbrx/8bit-lora.yaml
@@ -0,0 +1,81 @@
+base_model: LnL-AI/dbrx-base-converted-v2
+trust_remote_code: true
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.0
+output_dir: ./outputs/out
+
+sequence_len: 512
+sample_packing: false
+pad_to_sequence_len: false
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+adapter: lora
+lora_model_dir:
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+# w1, w2, & v1 will hang the trainer
+lora_target_modules:
+ - q_proj # attn
+ - k_proj # attn
+ - v_proj # attn
+ - out_proj # attn
+ - layer # router
+# - w1
+# - w2
+# - v1
+
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 1
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: false # don't use with fsdp_activation_checkpointing
+gradient_checkpointing_kwargs:
+ use_reentrant: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch:
+saves_per_epoch: 1
+debug:
+weight_decay: 0.0
+fsdp:
+ - full_shard
+ - auto_wrap
+fsdp_config:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: false
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: true
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+ fsdp_transformer_layer_cls_to_wrap: DbrxBlock
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_activation_checkpointing: true
diff --git a/examples/dbrx/README.md b/examples/dbrx/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..99ff3dd0b73b594431eb18a109923ddcc3e329af
--- /dev/null
+++ b/examples/dbrx/README.md
@@ -0,0 +1,26 @@
+# DBRX MoE
+
+Currently, for LoRA, only the `q_proj`, `k_proj`, `v_proj` `out_proj` and `layer` Linear layers are trainable.
+
+We are using the "converted" base models based on [this issue](https://huggingface.co/databricks/dbrx-instruct/discussions/10)
+where the Experts are fused as an `nn.Parameter` rather than a `nn.Linear` layer. However, the implementation
+is still a bit buggy and attempting to train a LoRA adapter over those `w1`, `w2` and `v1` layers
+results in the trainer hanging.
+
+
+### FSDP
+We've tested using the [`LnL-AI/dbrx-base-converted-v2`](https://huggingface.co/LnL-AI/dbrx-base-converted-v2) model as the base model for FSDP.
+
+The high memory usage seen w/ FSDP is due to FSDP not supporting 8bit optimizers.
+
+- 16-bit LoRA w/ FSDP
+ - ✅ w/o CPU Offload - 8x80GB uses ~80GiB/gpu
+ - ❌ w/ CPU Offload - `paged_adamw_8bit` optimizer errors from being on cpu
+- ✅ 8-bit LoRA w/ FSDP
+- ❌ 4-bit QLoRA w/ FSDP - errors w/: `Error an illegal memory access was encountered at line 90 in file /src/csrc/ops.cu`
+- ✅ bf16 full finetune w/ FSDP, freezing all but first 8 layers (8x80GB uses ~78GiB/gpu)
+
+
+### Deepspeed
+
+WIP
diff --git a/examples/dbrx/fft-ds-zero3.yaml b/examples/dbrx/fft-ds-zero3.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..60dc201eee03bbe9a72131f63ff234c17afaa127
--- /dev/null
+++ b/examples/dbrx/fft-ds-zero3.yaml
@@ -0,0 +1,56 @@
+base_model: LnL-AI/dbrx-base-converted-v2
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.0
+output_dir: ./outputs/out
+
+sequence_len: 512
+sample_packing: false
+pad_to_sequence_len: false
+
+unfrozen_parameters:
+ - transformer.blocks.[0-7].
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 1
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch:
+saves_per_epoch: 1
+debug:
+weight_decay: 0.0
+deepspeed: deepspeed_configs/zero3_bf16.json
diff --git a/examples/falcon/config-7b-lora.yml b/examples/falcon/config-7b-lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..029ca40e0902377b1a6d019b3363c28be26dcb33
--- /dev/null
+++ b/examples/falcon/config-7b-lora.yml
@@ -0,0 +1,64 @@
+base_model: tiiuae/falcon-7b
+trust_remote_code: true
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+gptq: false
+strict: false
+push_dataset_to_hub:
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca:chat
+dataset_prepared_path:
+val_set_size: 0.05
+adapter: lora
+lora_model_dir:
+sequence_len: 2048
+max_packed_sequence_len:
+lora_r: 16
+lora_alpha: 32
+lora_dropout: 0.0
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/falcon-7b
+batch_size: 2
+micro_batch_size: 1
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.00003
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention: true
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 40
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ pad_token: "<|endoftext|>"
+ bos_token: "<|endoftext|>"
+ eos_token: "<|endoftext|>"
diff --git a/examples/falcon/config-7b-qlora.yml b/examples/falcon/config-7b-qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..4e34144ed6dc148d08c7f2ecc4f5fb07cbe187b1
--- /dev/null
+++ b/examples/falcon/config-7b-qlora.yml
@@ -0,0 +1,93 @@
+# 1b: tiiuae/falcon-rw-1b
+# 40b: tiiuae/falcon-40b
+base_model: tiiuae/falcon-7b
+# required by falcon custom model code: https://huggingface.co/tiiuae/falcon-7b/tree/main
+trust_remote_code: true
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: false
+# enable 4bit for QLoRA
+load_in_4bit: true
+gptq: false
+strict: false
+push_dataset_to_hub:
+datasets:
+ - path: QingyiSi/Alpaca-CoT
+ data_files:
+ - Chain-of-Thought/formatted_cot_data/gsm8k_train.json
+ type: "alpaca:chat"
+dataset_prepared_path:
+val_set_size: 0.05
+# enable QLoRA
+adapter: qlora
+lora_model_dir:
+sequence_len: 2048
+max_packed_sequence_len:
+
+# hyperparameters from QLoRA paper Appendix B.2
+# "We find hyperparameters to be largely robust across datasets"
+lora_r: 64
+lora_alpha: 16
+# 0.1 for models up to 13B
+# 0.05 for 33B and 65B models
+lora_dropout: 0.05
+# add LoRA modules on all linear layers of the base model
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/qlora-out
+
+# QLoRA paper Table 9
+# - 16 for 7b & 13b
+# - 32 for 33b, 64 for 64b
+# Max size tested on A6000
+# - 7b: 40
+# - 40b: 4
+# decrease if OOM, increase for max VRAM utilization
+micro_batch_size: 1
+gradient_accumulation_steps: 2
+num_epochs: 4
+# Optimizer for QLoRA
+optimizer: paged_adamw_32bit
+torchdistx_path:
+lr_scheduler: cosine
+# QLoRA paper Table 9
+# - 2e-4 for 7b & 13b
+# - 1e-4 for 33b & 64b
+learning_rate: 0.0002
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+gradient_checkpointing: true
+# stop training after this many evaluation losses have increased in a row
+# https://huggingface.co/transformers/v4.2.2/_modules/transformers/trainer_callback.html#EarlyStoppingCallback
+early_stopping_patience: 3
+resume_from_checkpoint:
+auto_resume_from_checkpoints: true
+local_rank:
+logging_steps: 1
+xformers_attention: true
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.000001
+fsdp:
+fsdp_config:
+special_tokens:
+ pad_token: "<|endoftext|>"
+ bos_token: "<|endoftext|>"
+ eos_token: "<|endoftext|>"
diff --git a/examples/falcon/config-7b.yml b/examples/falcon/config-7b.yml
new file mode 100644
index 0000000000000000000000000000000000000000..36264f063e3aa96f7d5263a2df5c84ad8cc6692a
--- /dev/null
+++ b/examples/falcon/config-7b.yml
@@ -0,0 +1,64 @@
+base_model: tiiuae/falcon-7b
+trust_remote_code: true
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+gptq: false
+strict: false
+push_dataset_to_hub:
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca:chat
+dataset_prepared_path:
+val_set_size: 0.05
+adapter:
+lora_model_dir:
+sequence_len: 2048
+max_packed_sequence_len:
+lora_r: 64
+lora_alpha: 32
+lora_dropout: 0.0
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/falcon-7b
+batch_size: 2
+micro_batch_size: 1
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.00003
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention: true
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 40
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ pad_token: "<|endoftext|>"
+ bos_token: "<|endoftext|>"
+ eos_token: "<|endoftext|>"
diff --git a/examples/gemma/qlora.yml b/examples/gemma/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..e08facfc5d4541efdb779a2cfb1dc6949b115505
--- /dev/null
+++ b/examples/gemma/qlora.yml
@@ -0,0 +1,66 @@
+# use google/gemma-7b if you have access
+base_model: mhenrichsen/gemma-7b
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+# huggingface repo
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+val_set_size: 0.1
+output_dir: ./outputs/out
+
+adapter: qlora
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+
+sequence_len: 4096
+sample_packing: true
+eval_sample_packing: false
+pad_to_sequence_len: true
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+
+gradient_accumulation_steps: 3
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_ratio: 0.1
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/gptj/qlora.yml b/examples/gptj/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..f801729fac9833ea69e70e45319b17565fa6d7d8
--- /dev/null
+++ b/examples/gptj/qlora.yml
@@ -0,0 +1,57 @@
+base_model: EleutherAI/gpt-j-6b
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+push_dataset_to_hub:
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+adapter: qlora
+lora_model_dir:
+sequence_len: 2048
+max_packed_sequence_len:
+lora_r: 8
+lora_alpha: 32
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/qlora-out
+gradient_accumulation_steps: 2
+micro_batch_size: 2
+num_epochs: 2
+optimizer: paged_adamw_8bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.0001
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention: true
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+special_tokens:
+ pad_token: "<|endoftext|>"
diff --git a/examples/jamba/README.md b/examples/jamba/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..54f5d1da9c2410f17f49fe65335df29dd6c9c348
--- /dev/null
+++ b/examples/jamba/README.md
@@ -0,0 +1,10 @@
+# Jamba
+
+- ✅ qlora w/ deepspeed Zero-2 needs at least 2x GPUs and
+ - 35GiB VRAM per GPU w minimal context length
+ - 56GiB VRAM per GPU (w multipack enabled)
+- ✅ qlora w/ deepspeed Zero-3 needs at least 2x GPUs and 67GiB VRAM (wtf?)
+- ✅ qlora single-gpu, ~51GiB VRAM
+- ✅ multipack
+- ❓ FSDP
+- ❓ 8-bit LoRA
diff --git a/examples/jamba/qlora.yaml b/examples/jamba/qlora.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..3d6f69e793c8367daed54fec7287ca16aec23068
--- /dev/null
+++ b/examples/jamba/qlora.yaml
@@ -0,0 +1,62 @@
+base_model: ai21labs/Jamba-v0.1
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.0
+output_dir: ./outputs/out
+
+sequence_len: 4096
+sample_packing: false
+pad_to_sequence_len: false
+eval_sample_packing: false
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+adapter: qlora
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+
+low_cpu_mem_usage: true
+gradient_accumulation_steps: 4
+micro_batch_size: 1
+num_epochs: 2
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 0.00001
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch:
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+special_tokens:
diff --git a/examples/jamba/qlora_deepspeed.yaml b/examples/jamba/qlora_deepspeed.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..43a76c00b106391ba5f6ac9e94365bfaf98c1218
--- /dev/null
+++ b/examples/jamba/qlora_deepspeed.yaml
@@ -0,0 +1,62 @@
+base_model: ai21labs/Jamba-v0.1
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.0
+output_dir: ./outputs/out
+
+sequence_len: 4096
+sample_packing: false
+pad_to_sequence_len: false
+eval_sample_packing: false
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+adapter: qlora
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+
+low_cpu_mem_usage: true
+gradient_accumulation_steps: 4
+micro_batch_size: 1
+num_epochs: 2
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 0.00001
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch:
+saves_per_epoch: 1
+debug:
+deepspeed: deepspeed_configs/zero2.json
+weight_decay: 0.0
+special_tokens:
diff --git a/examples/jeopardy-bot/config.yml b/examples/jeopardy-bot/config.yml
new file mode 100644
index 0000000000000000000000000000000000000000..088629c0840aa17641e3924ae4bf79b6204ad46e
--- /dev/null
+++ b/examples/jeopardy-bot/config.yml
@@ -0,0 +1,55 @@
+base_model: huggyllama/llama-7b
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+load_in_8bit: false
+datasets:
+ - path: openaccess-ai-collective/jeopardy
+ type: jeopardy
+dataset_prepared_path:
+val_set_size: 0.02
+adapter:
+lora_model_dir:
+sequence_len: 512
+max_packed_sequence_len:
+lora_r:
+lora_alpha:
+lora_dropout:
+lora_target_modules:
+lora_fan_in_fan_out: false
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/jeopardy-bot-7b
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.00003
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+tf32: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 5
+xformers_attention: true
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 20
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/llama-2/README.md b/examples/llama-2/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..2ddd711e29a34b72f34efb5745e241ca8e015675
--- /dev/null
+++ b/examples/llama-2/README.md
@@ -0,0 +1,24 @@
+# Overview
+
+This is an example of a llama-2 configuration for 7b and 13b. The yaml file contains configuration for the 7b variant, but you can just aswell use the same settings for 13b.
+
+The 7b variant fits on any 24GB VRAM GPU and will take up about 17 GB of VRAM during training if using qlora and 20 GB if using lora. On a RTX 4090 it trains 3 epochs of the default dataset in about 15 minutes.
+
+The 13b variant will fit if you change these settings to these values:
+gradient_accumulation_steps: 2
+micro_batch_size: 1
+
+```shell
+accelerate launch -m axolotl.cli.train examples/llama-2/qlora.yml
+```
+or
+
+```shell
+accelerate launch -m axolotl.cli.train examples/llama-2/lora.yml
+```
+
+To launch a full finetuning with 16-bit precision:
+
+```shell
+accelerate launch -m axolotl.cli.train examples/llama-2/fft_optimized.yml
+```
diff --git a/examples/llama-2/fft_optimized.yml b/examples/llama-2/fft_optimized.yml
new file mode 100644
index 0000000000000000000000000000000000000000..3d94b04b8b7029ac8f6ca8ac4cce925a8afc6913
--- /dev/null
+++ b/examples/llama-2/fft_optimized.yml
@@ -0,0 +1,68 @@
+base_model: NousResearch/Llama-2-7b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.05
+output_dir: ./outputs/out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter:
+lora_model_dir:
+lora_r:
+lora_alpha:
+lora_dropout:
+lora_target_linear:
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 1
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+
+warmup_steps: 100
+evals_per_epoch: 4
+eval_table_size:
+saves_per_epoch: 1
+debug:
+deepspeed: #deepspeed_configs/zero2.json # multi-gpu only
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/llama-2/gptq-lora.yml b/examples/llama-2/gptq-lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..2a706265bdde1be7e96047593074f01ff3116766
--- /dev/null
+++ b/examples/llama-2/gptq-lora.yml
@@ -0,0 +1,72 @@
+base_model: TheBloke/Llama-2-7B-GPTQ
+gptq: true
+gptq_disable_exllama: true
+model_type: AutoModelForCausalLM
+tokenizer_type: LlamaTokenizer
+tokenizer_use_fast: true
+tokenizer_legacy: true
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+push_dataset_to_hub:
+hf_use_auth_token: true
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+adapter: lora
+lora_model_dir:
+sequence_len: 4096
+sample_packing:
+lora_r: 8
+lora_alpha: 32
+lora_dropout: 0.05
+lora_target_modules:
+ - k_proj
+ - o_proj
+ - q_proj
+ - v_proj
+lora_target_linear:
+lora_fan_in_fan_out:
+wandb_project:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/model-out
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 4
+optimizer: adamw_torch
+adam_beta2: 0.95
+adam_eps: 0.00001
+max_grad_norm: 1.0
+torchdistx_path:
+lr_scheduler: cosine
+lr_quadratic_warmup: true
+learning_rate: 0.000017
+train_on_inputs: false
+group_by_length: false
+bf16: false
+fp16: false
+float16: true
+tf32: true
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention:
+sdp_attention:
+flash_optimum:
+warmup_steps: 100
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/llama-2/lisa.yml b/examples/llama-2/lisa.yml
new file mode 100644
index 0000000000000000000000000000000000000000..7012d1f6131215eebad688762f8811c1183553d3
--- /dev/null
+++ b/examples/llama-2/lisa.yml
@@ -0,0 +1,75 @@
+base_model: NousResearch/Llama-2-7b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.05
+output_dir: ./outputs/lisa-out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter:
+lora_model_dir:
+lora_r:
+lora_alpha:
+lora_dropout:
+lora_target_linear:
+lora_fan_in_fan_out:
+
+lisa_n_layers: 4
+lisa_step_interval: 20
+lisa_layers_attribute: model.layers
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 2
+micro_batch_size: 1
+num_epochs: 1
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 5e-5 # recommendation from lisa paper for 7b
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+
+warmup_steps: 100
+evals_per_epoch: 4
+eval_table_size:
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/llama-2/loftq.yml b/examples/llama-2/loftq.yml
new file mode 100644
index 0000000000000000000000000000000000000000..68d9ac01421e1c841b941b217889b914d95da8de
--- /dev/null
+++ b/examples/llama-2/loftq.yml
@@ -0,0 +1,69 @@
+base_model: NousResearch/Llama-2-7b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+peft:
+ loftq_config:
+ loftq_bits: 4
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+s2_attention:
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/llama-2/lora.yml b/examples/llama-2/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..95bfae6920b4f9734d8bffb433ef6435097cb724
--- /dev/null
+++ b/examples/llama-2/lora.yml
@@ -0,0 +1,66 @@
+base_model: NousResearch/Llama-2-7b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+s2_attention:
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/llama-2/qlora-fsdp.yml b/examples/llama-2/qlora-fsdp.yml
new file mode 100644
index 0000000000000000000000000000000000000000..88029f92d5bcb2c02ede8b0fe8df2ae451803a1f
--- /dev/null
+++ b/examples/llama-2/qlora-fsdp.yml
@@ -0,0 +1,78 @@
+base_model: NousResearch/Llama-2-7b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: yahma/alpaca-cleaned
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.05
+output_dir: ./outputs/qlora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 512
+sample_packing: false
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 4
+num_epochs: 4
+optimizer: adamw_torch
+lr_scheduler: cosine
+learning_rate: 0.00001
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+ - full_shard
+ - auto_wrap
+fsdp_config:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: true
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: true
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
+ fsdp_state_dict_type: FULL_STATE_DICT
+special_tokens:
diff --git a/examples/llama-2/qlora.yml b/examples/llama-2/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..dda32170bd812028cf377e15c7a36323cafe0ca1
--- /dev/null
+++ b/examples/llama-2/qlora.yml
@@ -0,0 +1,66 @@
+base_model: NousResearch/Llama-2-7b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/qlora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: paged_adamw_32bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/llama-2/relora.yml b/examples/llama-2/relora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..93247ce068abf39db2e1de2fb3d79c82f1ea3f37
--- /dev/null
+++ b/examples/llama-2/relora.yml
@@ -0,0 +1,73 @@
+base_model: NousResearch/Llama-2-7b-hf
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/relora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+relora_steps: 150
+relora_warmup_steps: 10
+relora_cpu_offload: false
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 4
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/llama-3/README.md b/examples/llama-3/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..5c01af27bb435f4f6af4f152648d7a6f9b4b216c
--- /dev/null
+++ b/examples/llama-3/README.md
@@ -0,0 +1,13 @@
+# Llama-3
+
+https://llama.meta.com/llama3/
+
+[8B Base Model](https://huggingface.co/meta-llama/Meta-Llama-3-8B)
+ - [Full Fine Tune](./fft-8b.yaml)
+ - Single GPU @ 48GB VRAM
+ - [LoRA](./lora-8b.yml)
+ - Single GPU @ 11GB VRAM
+
+[70B Base Model](https://huggingface.co/meta-llama/Meta-Llama-3-70B)
+ - [QLORA+FSDP](./qlora-fsdp-70b.yaml)
+ - Dual GPU @ 21GB VRAM
diff --git a/examples/llama-3/fft-8b.yaml b/examples/llama-3/fft-8b.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..a36fd740e417627859a26136ef4cb3cd94f8552e
--- /dev/null
+++ b/examples/llama-3/fft-8b.yaml
@@ -0,0 +1,58 @@
+base_model: meta-llama/Meta-Llama-3-8B
+model_type: LlamaForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.05
+output_dir: ./outputs/out
+
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 8
+micro_batch_size: 1
+num_epochs: 1
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 2e-5
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: false
+early_stopping_patience:
+resume_from_checkpoint:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 100
+evals_per_epoch: 2
+eval_table_size:
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ pad_token: <|end_of_text|>
diff --git a/examples/llama-3/lora-8b.yml b/examples/llama-3/lora-8b.yml
new file mode 100644
index 0000000000000000000000000000000000000000..6b0ebaed86cdcc9079f64798d648aa7a37529ac2
--- /dev/null
+++ b/examples/llama-3/lora-8b.yml
@@ -0,0 +1,67 @@
+base_model: meta-llama/Meta-Llama-3-8B
+model_type: LlamaForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+s2_attention:
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ pad_token: <|end_of_text|>
diff --git a/examples/llama-3/qlora-fsdp-70b.yaml b/examples/llama-3/qlora-fsdp-70b.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..9b74f6b4de687b9c0e859d1f05f1e169c15421ab
--- /dev/null
+++ b/examples/llama-3/qlora-fsdp-70b.yaml
@@ -0,0 +1,80 @@
+base_model: casperhansen/llama-3-70b-fp16
+model_type: LlamaForCausalLM
+tokenizer_type: AutoTokenizer # PreTrainedTokenizerFast
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.05
+output_dir: ./outputs/out/qlora-llama3-70b
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 512
+sample_packing: false
+pad_to_sequence_len: true
+
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 1
+num_epochs: 4
+optimizer: adamw_torch
+lr_scheduler: cosine
+learning_rate: 0.00001
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+ - full_shard
+ - auto_wrap
+fsdp_config:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: true
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: true
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_sharding_strategy: FULL_SHARD
+special_tokens:
+ pad_token: <|end_of_text|>
diff --git a/examples/llama-3/qlora.yml b/examples/llama-3/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..44120d9385db8da0d7d67c96ab689726e06ccd95
--- /dev/null
+++ b/examples/llama-3/qlora.yml
@@ -0,0 +1,67 @@
+base_model: meta-llama/Meta-Llama-3-8B
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: aaditya/alpaca_subset_1
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0
+output_dir: ./outputs/qlora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: paged_adamw_32bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ pad_token: "<|end_of_text|>"
diff --git a/examples/mamba/config.yml b/examples/mamba/config.yml
new file mode 100644
index 0000000000000000000000000000000000000000..f88f5138d916055fe3b4e47fd39f575d5c151771
--- /dev/null
+++ b/examples/mamba/config.yml
@@ -0,0 +1,61 @@
+base_model: state-spaces/mamba-2.8b
+model_type: MambaLMHeadModel
+tokenizer_type: AutoTokenizer
+tokenizer_config: EleutherAI/gpt-neox-20b
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.0
+output_dir: ./outputs/out
+
+sequence_len: 2048
+sample_packing: false
+pad_to_sequence_len: false
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 1
+num_epochs: 2
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 5e-5
+
+train_on_inputs: false
+group_by_length: true
+
+bf16: auto
+fp16:
+tf32: true
+
+gradient_checkpointing: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention:
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+tokens:
+save_safetensors: False
diff --git a/examples/mistral/README.md b/examples/mistral/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..462c2d3e7bb167d04f4bec5da151dbc66eb8a608
--- /dev/null
+++ b/examples/mistral/README.md
@@ -0,0 +1,12 @@
+**Mistral 7B** is a language model with a total of 7.3 billion parameters, showcasing a notable performance across a variety of benchmarks.
+
+Fine Tune:
+```shell
+accelerate launch -m axolotl.cli.train examples/mistral/config.yml
+
+```
+
+If you run into CUDA OOM, use deepspeed with config zero2.json:
+```shell
+accelerate launch -m axolotl.cli.train examples/mistral/config.yml --deepspeed deepspeed_configs/zero2.json
+```
diff --git a/examples/mistral/bigstral-ds-zero3.yaml b/examples/mistral/bigstral-ds-zero3.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..e993e44a7862be45ffd1a173f17e07a7aa2a59b7
--- /dev/null
+++ b/examples/mistral/bigstral-ds-zero3.yaml
@@ -0,0 +1,63 @@
+base_model: mistral-community/Mixtral-8x22B-v0.1
+model_type: AutoModelForCausalLM
+tokenizer_type: LlamaTokenizer
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+unfrozen_parameters:
+ - ^lm_head.weight$
+ - ^model.embed_tokens.weight$
+ - model.layers.4[4-9]+.block_sparse_moe.gate
+ - model.layers.4[4-9]+.block_sparse_moe.experts
+ - model.layers.5[0-5]+.block_sparse_moe.gate
+ - model.layers.5[0-5]+.block_sparse_moe.experts
+
+model_config:
+ output_router_logits: true
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.05
+output_dir: ./outputs/out
+
+sequence_len: 2048
+sample_packing: true
+pad_to_sequence_len: true
+
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 3
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0001
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+save_total_limit: 1
+save_steps:
+debug:
+deepspeed: deepspeed_configs/zero3_bf16_cpuoffload_params.json
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ eos_token: "<|im_end|>"
+tokens:
+ - "<|im_start|>"
diff --git a/examples/mistral/config.yml b/examples/mistral/config.yml
new file mode 100644
index 0000000000000000000000000000000000000000..a70937c4fd714b88edb5cee3883aab2f34be43b0
--- /dev/null
+++ b/examples/mistral/config.yml
@@ -0,0 +1,58 @@
+base_model: mistralai/Mistral-7B-v0.1
+model_type: MistralForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/out
+
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+eval_sample_packing: false
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.000005
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/mistral/lora-mps.yml b/examples/mistral/lora-mps.yml
new file mode 100644
index 0000000000000000000000000000000000000000..03c74bb59bbca7c6eb062c68b92ad9b704fff108
--- /dev/null
+++ b/examples/mistral/lora-mps.yml
@@ -0,0 +1,79 @@
+base_model: mistralai/Mistral-7B-v0.1
+model_type: MistralForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0
+output_dir: ./outputs/lora-out
+eval_sample_packing: false
+
+adapter: lora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+lora_target_modules:
+ - gate_proj
+ - down_proj
+ - up_proj
+ - q_proj
+ - v_proj
+ - k_proj
+ - o_proj
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 8
+micro_batch_size: 1
+num_epochs: 2
+optimizer: adamw_torch
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16: false
+tf32: true
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: false
+sdp_attention: true
+
+loss_watchdog_threshold: 5.0
+loss_watchdog_patience: 3
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_table_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/mistral/lora.yml b/examples/mistral/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..0d5dc9edd7c9c76adb5aa05d67832e0f9abb6d47
--- /dev/null
+++ b/examples/mistral/lora.yml
@@ -0,0 +1,77 @@
+base_model: mistralai/Mistral-7B-v0.1
+model_type: MistralForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.1
+output_dir: ./outputs/lora-out
+
+adapter: lora
+lora_model_dir:
+
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+lora_target_modules:
+ - gate_proj
+ - down_proj
+ - up_proj
+ - q_proj
+ - v_proj
+ - k_proj
+ - o_proj
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 1
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+loss_watchdog_threshold: 5.0
+loss_watchdog_patience: 3
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/mistral/mistral-qlora-fsdp.yml b/examples/mistral/mistral-qlora-fsdp.yml
new file mode 100644
index 0000000000000000000000000000000000000000..e6b07c594b1e4f2ec3fecf962a9e0c32407f3eee
--- /dev/null
+++ b/examples/mistral/mistral-qlora-fsdp.yml
@@ -0,0 +1,82 @@
+base_model: mistralai/Mixtral-8x7B-v0.1
+model_type: AutoModelForCausalLM
+tokenizer_type: LlamaTokenizer
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.02
+output_dir: ./outputs/qlora-out
+
+model_config:
+ output_router_logits: true
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 1024
+sample_packing: false
+pad_to_sequence_len: false
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 1
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+loss_watchdog_threshold: 5.0
+loss_watchdog_patience: 3
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+weight_decay: 0.0
+fsdp:
+ - full_shard
+ - auto_wrap
+fsdp_config:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: false
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: false
+ fsdp_transformer_layer_cls_to_wrap: MistralDecoderLayer
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+special_tokens:
diff --git a/examples/mistral/mistral-qlora-orpo.yml b/examples/mistral/mistral-qlora-orpo.yml
new file mode 100644
index 0000000000000000000000000000000000000000..2549ef018c6e065597e80030a0dc7f05d62ed51c
--- /dev/null
+++ b/examples/mistral/mistral-qlora-orpo.yml
@@ -0,0 +1,82 @@
+base_model: mistralai/Mistral-7B-v0.1
+model_type: MistralForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+rl: orpo
+orpo_alpha: 0.1
+remove_unused_columns: false
+
+chat_template: chatml
+datasets:
+ - path: argilla/ultrafeedback-binarized-preferences-cleaned
+ type: chat_template.argilla
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.1
+output_dir: ./outputs/mistral-qlora-orpo-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: false
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+lora_target_modules:
+ - gate_proj
+ - down_proj
+ - up_proj
+ - q_proj
+ - v_proj
+ - k_proj
+ - o_proj
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 1
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+loss_watchdog_threshold: 5.0
+loss_watchdog_patience: 3
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/mistral/mixtral-8x22b-qlora-fsdp.yml b/examples/mistral/mixtral-8x22b-qlora-fsdp.yml
new file mode 100644
index 0000000000000000000000000000000000000000..fe68b28172901ce8615174da984df916bdab2cee
--- /dev/null
+++ b/examples/mistral/mixtral-8x22b-qlora-fsdp.yml
@@ -0,0 +1,81 @@
+base_model: mistral-community/Mixtral-8x22B-v0.1
+model_type: AutoModelForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.02
+output_dir: ./outputs/qlora-out
+
+model_config:
+ output_router_logits: true
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 1024
+sample_packing: false
+pad_to_sequence_len: false
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 1
+optimizer: adamw_torch
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+loss_watchdog_threshold: 5.0
+loss_watchdog_patience: 3
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+weight_decay: 0.0
+fsdp:
+ - full_shard
+ - auto_wrap
+fsdp_config:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: true
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: true
+ fsdp_transformer_layer_cls_to_wrap: MixtralSparseMoeBlock
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+special_tokens:
diff --git a/examples/mistral/mixtral-qlora-fsdp.yml b/examples/mistral/mixtral-qlora-fsdp.yml
new file mode 100644
index 0000000000000000000000000000000000000000..c0959704027713d5dde9901e376538df17223c2e
--- /dev/null
+++ b/examples/mistral/mixtral-qlora-fsdp.yml
@@ -0,0 +1,85 @@
+base_model: mistralai/Mixtral-8x7B-v0.1
+model_type: AutoModelForCausalLM
+tokenizer_type: LlamaTokenizer
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.02
+output_dir: ./outputs/qlora-out
+
+model_config:
+ output_router_logits: true
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 1024
+sample_packing: false
+pad_to_sequence_len: false
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 1
+optimizer: adamw_torch
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+loss_watchdog_threshold: 5.0
+loss_watchdog_patience: 3
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+weight_decay: 0.0
+fsdp:
+ - full_shard
+ - auto_wrap
+fsdp_config:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: true
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: true
+ fsdp_transformer_layer_cls_to_wrap: MixtralSparseMoeBlock
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+ fsdp_sharding_strategy: FULL_SHARD
+ fsdp_forward_prefetch: false
+ fsdp_backward_prefetch: BACKWARD_PRE
+special_tokens:
diff --git a/examples/mistral/mixtral.yml b/examples/mistral/mixtral.yml
new file mode 100644
index 0000000000000000000000000000000000000000..13fbe92ab8d7fd4d96444835728c0928d95cfcbe
--- /dev/null
+++ b/examples/mistral/mixtral.yml
@@ -0,0 +1,91 @@
+base_model: mistralai/Mixtral-8x7B-v0.1
+model_type: AutoModelForCausalLM
+tokenizer_type: LlamaTokenizer
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: tatsu-lab/alpaca
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.0
+output_dir: ./outputs/qlora-out
+
+## You can optionally freeze the entire model and unfreeze a subset of parameters
+unfrozen_parameters:
+# - ^lm_head.weight$
+# - ^model.embed_tokens.weight$[:32000]
+# - model.layers.2[0-9]+.block_sparse_moe.gate
+# - model.layers.2[0-9]+.block_sparse_moe.experts
+# - model.layers.3[0-9]+.block_sparse_moe.gate
+# - model.layers.3[0-9]+.block_sparse_moe.experts
+
+model_config:
+ output_router_logits: true
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+#lora_target_modules:
+# - gate
+# - q_proj
+# - k_proj
+# - v_proj
+# - o_proj
+# - w1
+# - w2
+# - w3
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 2
+micro_batch_size: 1
+num_epochs: 1
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+loss_watchdog_threshold: 5.0
+loss_watchdog_patience: 3
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed: deepspeed_configs/zero2.json
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/mistral/mixtral_22.yml b/examples/mistral/mixtral_22.yml
new file mode 100644
index 0000000000000000000000000000000000000000..9a1e86386c28963902c1bfb2eb405d4530ae3244
--- /dev/null
+++ b/examples/mistral/mixtral_22.yml
@@ -0,0 +1,61 @@
+base_model: mistral-community/Mixtral-8x22B-v0.1
+model_type: AutoModelForCausalLM
+tokenizer_type: LlamaTokenizer
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+unfrozen_parameters:
+ - ^lm_head.weight$
+ - ^model.embed_tokens.weight$
+ - model.layers.4[4-9]+.block_sparse_moe.gate
+ - model.layers.4[4-9]+.block_sparse_moe.experts
+ - model.layers.5[0-5]+.block_sparse_moe.gate
+ - model.layers.5[0-5]+.block_sparse_moe.experts
+
+model_config:
+ output_router_logits: true
+
+datasets:
+ - path: yahma/alpaca-cleaned
+ type: alpaca
+output_dir: ./outputs/out
+
+sequence_len: 8000
+sample_packing: true
+pad_to_sequence_len: true
+
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 3
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0001
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+save_total_limit: 1
+save_steps:
+debug:
+deepspeed: deepspeed_configs/zero3_bf16_cpuoffload_all.json
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+ eos_token: "<|im_end|>"
+tokens:
+ - "<|im_start|>"
diff --git a/examples/mistral/qlora.yml b/examples/mistral/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..c7bdb155c0ec69afc0b7147e49d4ca30c141853e
--- /dev/null
+++ b/examples/mistral/qlora.yml
@@ -0,0 +1,77 @@
+base_model: mistralai/Mistral-7B-v0.1
+model_type: MistralForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.1
+output_dir: ./outputs/qlora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+lora_target_modules:
+ - gate_proj
+ - down_proj
+ - up_proj
+ - q_proj
+ - v_proj
+ - k_proj
+ - o_proj
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 1
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+loss_watchdog_threshold: 5.0
+loss_watchdog_patience: 3
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/mpt-7b/README.md b/examples/mpt-7b/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..12c55688d2aefc559e8acde1a6fc354ce863fe19
--- /dev/null
+++ b/examples/mpt-7b/README.md
@@ -0,0 +1,6 @@
+# MPT-7B
+
+```shell
+accelerate launch scripts/finetune.py examples/mpt-7b/config.yml
+
+```
diff --git a/examples/mpt-7b/config.yml b/examples/mpt-7b/config.yml
new file mode 100644
index 0000000000000000000000000000000000000000..530415de1703ab68ff64451a0c6caf109e4d46f1
--- /dev/null
+++ b/examples/mpt-7b/config.yml
@@ -0,0 +1,58 @@
+base_model: mosaicml/mpt-7b
+tokenizer_type: AutoTokenizer
+trust_remote_code: true # required for mpt as their model class is not merged into transformers yet
+load_in_8bit: false
+datasets:
+ - path: vicgalle/alpaca-gpt4
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.02
+adapter:
+lora_model_dir:
+sequence_len: 2048
+max_packed_sequence_len:
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+ - q_proj
+ - v_proj
+lora_fan_in_fan_out: false
+wandb_project: mpt-alpaca-7b
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/mpt-alpaca-7b
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.0000002
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+tf32: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 5
+xformers_attention:
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 20
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0001
+fsdp:
+fsdp_config:
+tokens:
+ pad_token: "<|padding|>"
+ bos_token: "<|endoftext|>"
+ eos_token: "<|endoftext|>"
+ unk_token: "<|endoftext|>"
diff --git a/examples/openllama-3b/README.md b/examples/openllama-3b/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..3e9501a54f61577ab8f7301b6caebc3d4802000c
--- /dev/null
+++ b/examples/openllama-3b/README.md
@@ -0,0 +1,16 @@
+# openllama-3b
+
+Basic full tune
+```shell
+accelerate launch scripts/finetune.py examples/openllama-3b/config.yml
+```
+
+LoRA
+```shell
+accelerate launch scripts/finetune.py examples/openllama-3b/lora.yml
+```
+
+QLoRA
+```shell
+accelerate launch scripts/finetune.py examples/openllama-3b/qlora.yml
+```
diff --git a/examples/openllama-3b/config.yml b/examples/openllama-3b/config.yml
new file mode 100644
index 0000000000000000000000000000000000000000..a0473213c033efe5e858fe278177b7caa3310cda
--- /dev/null
+++ b/examples/openllama-3b/config.yml
@@ -0,0 +1,62 @@
+base_model: openlm-research/open_llama_3b_v2
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+push_dataset_to_hub:
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.02
+adapter:
+lora_model_dir:
+sequence_len: 1024
+sample_packing: true
+lora_r:
+lora_alpha:
+lora_dropout:
+lora_target_modules:
+lora_target_linear:
+lora_fan_in_fan_out:
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/openllama-out
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.000003
+train_on_inputs: false
+group_by_length: false
+float16: true
+bf16: false
+fp16: false
+tf32: false
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 20
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/openllama-3b/lora.yml b/examples/openllama-3b/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..2b67849159904e990c881285256837e89bb4fa7d
--- /dev/null
+++ b/examples/openllama-3b/lora.yml
@@ -0,0 +1,68 @@
+base_model: openlm-research/open_llama_3b_v2
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+push_dataset_to_hub:
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.02
+adapter: lora
+lora_model_dir:
+sequence_len: 1024
+sample_packing: true
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.0
+lora_target_modules:
+ - gate_proj
+ - down_proj
+ - up_proj
+ - q_proj
+ - v_proj
+ - k_proj
+ - o_proj
+lora_fan_in_fan_out:
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/lora-out
+gradient_accumulation_steps: 1
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.0002
+train_on_inputs: false
+group_by_length: false
+bf16: false
+fp16: true
+tf32: false
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+gptq_groupsize:
+s2_attention:
+gptq_model_v1:
+warmup_steps: 20
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/openllama-3b/qlora.yml b/examples/openllama-3b/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..8d4dc05ca7594e6f561cf192f03f092662610b76
--- /dev/null
+++ b/examples/openllama-3b/qlora.yml
@@ -0,0 +1,61 @@
+base_model: openlm-research/open_llama_3b_v2
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+push_dataset_to_hub:
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+adapter: qlora
+lora_model_dir:
+sequence_len: 1024
+sample_packing: true
+lora_r: 8
+lora_alpha: 32
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/qlora-out
+gradient_accumulation_steps: 1
+micro_batch_size: 2
+num_epochs: 4
+optimizer: paged_adamw_32bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.0002
+train_on_inputs: false
+group_by_length: false
+bf16: false
+fp16: true
+tf32: false
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 20
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+special_tokens:
+ bos_token: ""
+ eos_token: " "
+ unk_token: ""
diff --git a/examples/phi/README.md b/examples/phi/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..1b9e8022e33fb9392bd1ce549d8c03892e5f22e7
--- /dev/null
+++ b/examples/phi/README.md
@@ -0,0 +1,11 @@
+# Phi
+
+Due to some nuances with the phi code, please use deepspeed when training phi for full finetune.
+
+```shell
+accelerate launch -m axolotl.cli.train examples/phi/phi-ft.yml --deepspeed deepspeed_configs/zero1.json
+
+# OR
+
+python -m axolotl.cli.train examples/phi/phi-qlora.yml
+```
diff --git a/examples/phi/phi-ft.yml b/examples/phi/phi-ft.yml
new file mode 100644
index 0000000000000000000000000000000000000000..0dabadc7a4e68788d060ad6ba1e174eef22a7f5f
--- /dev/null
+++ b/examples/phi/phi-ft.yml
@@ -0,0 +1,71 @@
+base_model: microsoft/phi-1_5
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: garage-bAInd/Open-Platypus
+ type: alpaca
+
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/phi-sft-out
+
+sequence_len: 2048
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter:
+lora_model_dir:
+lora_r:
+lora_alpha:
+lora_dropout:
+lora_target_linear:
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 1
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_torch
+adam_beta2: 0.95
+adam_epsilon: 0.00001
+max_grad_norm: 1.0
+lr_scheduler: cosine
+learning_rate: 0.000003
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: True
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 100
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+resize_token_embeddings_to_32x: true
+special_tokens:
+ pad_token: "<|endoftext|>"
diff --git a/examples/phi/phi-qlora.yml b/examples/phi/phi-qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..7c181a3c15968cad2d0df2c42132f3a55b8d1b17
--- /dev/null
+++ b/examples/phi/phi-qlora.yml
@@ -0,0 +1,71 @@
+base_model: microsoft/phi-1_5
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: garage-bAInd/Open-Platypus
+ type: alpaca
+
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/phi-sft-out
+
+sequence_len: 2048
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter: qlora
+lora_model_dir:
+lora_r: 64
+lora_alpha: 32
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 1
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_torch
+adam_beta2: 0.95
+adam_epsilon: 0.00001
+max_grad_norm: 1.0
+lr_scheduler: cosine
+learning_rate: 0.000003
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: True
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 100
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+resize_token_embeddings_to_32x: true
+special_tokens:
+ pad_token: "<|endoftext|>"
diff --git a/examples/phi/phi2-ft.yml b/examples/phi/phi2-ft.yml
new file mode 100644
index 0000000000000000000000000000000000000000..27815550b4c8a4d37a07c47c1ac30102693a2d09
--- /dev/null
+++ b/examples/phi/phi2-ft.yml
@@ -0,0 +1,71 @@
+base_model: microsoft/phi-2
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: garage-bAInd/Open-Platypus
+ type: alpaca
+
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/phi-sft-out
+
+sequence_len: 2048
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter:
+lora_model_dir:
+lora_r:
+lora_alpha:
+lora_dropout:
+lora_target_linear:
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 1
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_torch
+adam_beta2: 0.95
+adam_epsilon: 0.00001
+max_grad_norm: 1.0
+lr_scheduler: cosine
+learning_rate: 0.000003
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: True
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 100
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+resize_token_embeddings_to_32x: true
+special_tokens:
+ pad_token: "<|endoftext|>"
diff --git a/examples/pythia-12b/README.md b/examples/pythia-12b/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..123ffa7109a439845c88039cdcb3968da0b26a16
--- /dev/null
+++ b/examples/pythia-12b/README.md
@@ -0,0 +1,9 @@
+# Pythia 12B
+
+- Single-GPU A100 only (?)
+
+```shell
+python scripts/finetune.py examples/pythia-12b/config.yml
+```
+
+⚠️ Multiple-GPU A100 - Doesn't seem to work with multi-gpu without causing OOM! ⚠️
diff --git a/examples/pythia-12b/config.yml b/examples/pythia-12b/config.yml
new file mode 100644
index 0000000000000000000000000000000000000000..18e6beaafd2ec0405de2ae68847430063e816396
--- /dev/null
+++ b/examples/pythia-12b/config.yml
@@ -0,0 +1,48 @@
+base_model: EleutherAI/pythia-12b-deduped
+base_model_ignore_patterns: pytorch* # prefer safetensors
+model_type: GPTNeoXForCausalLM
+tokenizer_type: AutoTokenizer
+load_in_8bit: false
+load_in_4bit: false
+gptq: false
+device_map: auto
+datasets:
+ - path: vicgalle/alpaca-gpt4
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+adapter:
+lora_model_dir:
+sequence_len: 2048
+max_packed_sequence_len: 2048
+lora_r: 64
+lora_alpha: 32
+lora_dropout: 0.0
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out: true # pythia/GPTNeoX lora specific
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/pythia-12b
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 5
+learning_rate: 0.00003
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+train_on_inputs: false
+group_by_length: false
+bf16: false
+fp16: false
+float16: true
+tf32: true
+flash_optimum: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+gradient_checkpointing: true
+fsdp:
+fsdp_config:
diff --git a/examples/pythia/lora.yml b/examples/pythia/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..0aa650f67e807e1fe06f7fcc9d40cc409c073d5f
--- /dev/null
+++ b/examples/pythia/lora.yml
@@ -0,0 +1,37 @@
+base_model: EleutherAI/pythia-1.4b-deduped
+load_in_8bit: true
+datasets:
+ - path: teknium/GPT4-LLM-Cleaned
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+adapter: lora
+lora_model_dir:
+sequence_len: 512
+lora_r: 16
+lora_alpha: 32
+lora_dropout: 0.05
+lora_target_modules:
+ - query_key_value
+lora_target_linear:
+lora_fan_in_fan_out: true # pythia/GPTNeoX lora specific
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/lora-alpaca-pythia
+gradient_accumulation_steps: 1
+micro_batch_size: 4
+num_epochs: 4
+learning_rate: 0.00001
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+tf32: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+weight_decay: 0.1
+evals_per_epoch: 4
+logging_steps: 1
diff --git a/examples/qwen/README.md b/examples/qwen/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..b62fb4101be0edfe6c7f46a4081770c234c5b362
--- /dev/null
+++ b/examples/qwen/README.md
@@ -0,0 +1,10 @@
+# Qwen
+
+TODO
+
+# Qwen2 MoE
+
+✅ multipack
+✅ qwen2_moe 4-bit QLoRA
+✅ qwen2_moe 16-bit LoRA
+❓ qwen2_moe 8-bit LoRA
diff --git a/examples/qwen/lora.yml b/examples/qwen/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..dd8dc1e4f4e9452874b1ae738247e7a45e2f9085
--- /dev/null
+++ b/examples/qwen/lora.yml
@@ -0,0 +1,67 @@
+base_model: Qwen/Qwen-7B
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+trust_remote_code: true
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 2048 # supports up to 8192
+sample_packing: false
+pad_to_sequence_len:
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention:
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/qwen/qlora.yml b/examples/qwen/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..01c0c0ab864ac424630a811a87d5d62665e5c205
--- /dev/null
+++ b/examples/qwen/qlora.yml
@@ -0,0 +1,67 @@
+base_model: Qwen/Qwen-7B
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 2048 # supports up to 8192
+sample_packing: false
+pad_to_sequence_len:
+
+adapter: qlora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention:
+
+warmup_steps: 10
+evals_per_epoch: 4
+eval_table_size:
+eval_max_new_tokens: 128
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/qwen/qwen2-moe-lora.yaml b/examples/qwen/qwen2-moe-lora.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..452335e38f1bc81fc2e1e1de0a0ee19480f8be9d
--- /dev/null
+++ b/examples/qwen/qwen2-moe-lora.yaml
@@ -0,0 +1,64 @@
+base_model: Qwen/Qwen1.5-MoE-A2.7B
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/out
+
+sequence_len: 1024 # supports up to 32k
+sample_packing: false
+pad_to_sequence_len: false
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 1
+num_epochs: 4
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/qwen/qwen2-moe-qlora.yaml b/examples/qwen/qwen2-moe-qlora.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..bc11007c78777bf11fc28bfe733752470e4bf046
--- /dev/null
+++ b/examples/qwen/qwen2-moe-qlora.yaml
@@ -0,0 +1,64 @@
+base_model: Qwen/Qwen1.5-MoE-A2.7B
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/out
+
+sequence_len: 1024 # supports up to 32k
+sample_packing: false
+pad_to_sequence_len: false
+
+adapter: qlora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 1
+num_epochs: 4
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: true
+
+gradient_checkpointing: true
+gradient_checkpointing_kwargs:
+ use_reentrant: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/redpajama/README.md b/examples/redpajama/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..bbd9a93ecc1a548de071250edd14209c8bfedc8f
--- /dev/null
+++ b/examples/redpajama/README.md
@@ -0,0 +1,6 @@
+# RedPajama 3B preview release
+
+```shell
+accelerate launch scripts/finetune.py examples/redpajama/config-3b.yml
+
+```
diff --git a/examples/redpajama/config-3b.yml b/examples/redpajama/config-3b.yml
new file mode 100644
index 0000000000000000000000000000000000000000..ff395a863df28927fbbf4c86394cd9c8cc30ec3f
--- /dev/null
+++ b/examples/redpajama/config-3b.yml
@@ -0,0 +1,59 @@
+base_model: togethercomputer/RedPajama-INCITE-Chat-3B-v1
+model_type: GPTNeoXForCausalLM
+tokenizer_type: AutoTokenizer
+trust_remote_code:
+load_in_8bit: false
+datasets:
+ - path: vicgalle/alpaca-gpt4
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.02
+adapter:
+lora_model_dir:
+sequence_len: 2048
+max_packed_sequence_len:
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+ - q_proj
+ - v_proj
+lora_fan_in_fan_out: false
+wandb_project: redpajama-alpaca-3b
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/redpajama-alpaca-3b
+batch_size: 4
+micro_batch_size: 1
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+torchdistx_path:
+lr_scheduler: cosine
+learning_rate: 0.0000002
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+tf32: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 5
+xformers_attention:
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 20
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0001
+fsdp:
+fsdp_config:
+tokens:
+ pad_token: "<|padding|>"
+ bos_token: "<|endoftext|>"
+ eos_token: "<|endoftext|>"
+ unk_token: "<|endoftext|>"
diff --git a/examples/replit-3b/config-lora.yml b/examples/replit-3b/config-lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..9fee099d474f09acf553ca85e0d9de54461ad996
--- /dev/null
+++ b/examples/replit-3b/config-lora.yml
@@ -0,0 +1,55 @@
+base_model: replit/replit-code-v1-3b
+trust_remote_code: true
+load_in_8bit: false
+datasets:
+ - path: vicgalle/alpaca-gpt4
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+adapter: lora
+lora_model_dir:
+sequence_len: 2048
+max_packed_sequence_len:
+lora_r: 8
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+ - Wqkv
+ - mlp_up
+ - mlp_down
+lora_fan_in_fan_out:
+wandb_project: lora-replit
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/lora-replit
+batch_size: 8
+micro_batch_size: 1
+num_epochs: 4
+optimizer:
+torchdistx_path:
+lr_scheduler:
+learning_rate: 0.00001
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+tf32: true
+gradient_checkpointing:
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 20
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0
+fsdp:
+fsdp_config:
+#special_tokens:
diff --git a/examples/stablelm-2/1.6b/fft.yml b/examples/stablelm-2/1.6b/fft.yml
new file mode 100644
index 0000000000000000000000000000000000000000..777262a7ee43de7cb990d276ec032d8b47960620
--- /dev/null
+++ b/examples/stablelm-2/1.6b/fft.yml
@@ -0,0 +1,69 @@
+base_model: stabilityai/stablelm-2-1_6b
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+trust_remote_code: true
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path: last_run_prepared
+val_set_size: 0.05
+output_dir: ./outputs/out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter:
+lora_model_dir:
+lora_r:
+lora_alpha:
+lora_dropout:
+lora_target_linear:
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 1
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+
+warmup_steps: 100
+evals_per_epoch: 4
+eval_table_size:
+saves_per_epoch: 1
+debug:
+deepspeed: #deepspeed_configs/zero2.json # multi-gpu only
+weight_decay: 0.1
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/stablelm-2/1.6b/lora.yml b/examples/stablelm-2/1.6b/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..c65b9e4cd0617afb898bb69533ddc7c7778c4b0b
--- /dev/null
+++ b/examples/stablelm-2/1.6b/lora.yml
@@ -0,0 +1,66 @@
+base_model: stabilityai/stablelm-2-1_6b
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+trust_remote_code: true
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 1
+micro_batch_size: 1
+num_epochs: 1
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/stablelm-2/README.md b/examples/stablelm-2/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..0cdc2ffb0e532dd573051066e37fad9af192b084
--- /dev/null
+++ b/examples/stablelm-2/README.md
@@ -0,0 +1,36 @@
+# StableLM 2
+
+This repository contains examples for training and processing using StableLM-2. It also includes a section to help you estimate the GPU requirements for your specific use case.
+
+## Estimating GPU Requirements
+
+| type | deepspeed | batch size | context length | vRAM GPU (GBs) |
+|---------------|-----------|------------|----------------|----------------|
+| full finetune | N/A | 1 | 4096 | ~21.5GBs |
+| full finetune | zero2 | 1 | 4096 | ~20GBs |
+| lora | N/A | 1 | 4096 | ~16.6GBs |
+
+The above are estimates and might differ slight depending on the setup for example whether you pack your sequence lengths or not (the above assumes you do to length 4096).
+
+This blog post from Hamel Husain was a great resource for estimating these numbers: https://hamel.dev/notes/llm/03_estimating_vram.html
+
+## Training
+We have example scripts here for both full finetuning and lora using the popular alpaca dataset:
+
+```shell
+# preprocess the dataset
+CUDA_VISIBLE_DEVICES="" python -m axolotl.cli.preprocess examples/stablelm-2/1.6b/lora.yml
+```
+
+Single GPU Training:
+```shell
+python -m axolotl.cli.train examples/stablelm-2/fft.yml --deepspeed deepspeed_configs/zero2.json
+# OR
+python -m axolotl.cli.train examples/stablelm-2/1.6b/lora.yml
+```
+
+Multinode GPU Training with `accelerate`:
+```shell
+# make sure you've configured accelerate properly
+accelerate launch -m axolotl.cli.train examples/stablelm-2/1.6b/fft.yml --deepspeed deepspeed_configs/zero2.json
+```
diff --git a/examples/starcoder2/qlora.yml b/examples/starcoder2/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..83fc0d89f76ba3c842f5011eee1d3d2c13d800ed
--- /dev/null
+++ b/examples/starcoder2/qlora.yml
@@ -0,0 +1,69 @@
+base_model: bigcode/starcoder2-3b
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+
+
+dataset_prepared_path:
+val_set_size: 0.2
+output_dir: ./outputs/qlora
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_run_id:
+wandb_log_model:
+
+gradient_accumulation_steps: 8
+micro_batch_size: 2
+num_epochs: 3
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 2e-5
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16: false
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 20
+evals_per_epoch: 4
+eval_steps:
+eval_table_size:
+saves_per_epoch: 4
+save_steps:
+save_total_limit: 2
+debug:
+deepspeed:
+weight_decay:
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/tiny-llama/README.md b/examples/tiny-llama/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..467c06ec87a43569560264c549332dbc00b57b7a
--- /dev/null
+++ b/examples/tiny-llama/README.md
@@ -0,0 +1,17 @@
+# Overview
+
+This is a simple example of how to finetune TinyLlama1.1B using either lora or qlora:
+
+LoRa:
+
+```
+accelerate launch -m axolotl.cli.train examples/tiny-llama/lora.yml
+```
+
+qLoRa:
+
+```
+accelerate launch -m axolotl.cli.train examples/tiny-llama/qlora.yml
+```
+
+Both take about 10 minutes to complete on a 4090.
diff --git a/examples/tiny-llama/lora-mps.yml b/examples/tiny-llama/lora-mps.yml
new file mode 100644
index 0000000000000000000000000000000000000000..c08be82d3b95e328b3dd03c5ad84011dc9547944
--- /dev/null
+++ b/examples/tiny-llama/lora-mps.yml
@@ -0,0 +1,64 @@
+base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0
+output_dir: ./outputs/lora-out
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+eval_sample_packing: false
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_torch
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16: false
+tf32: true
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: false
+
+warmup_steps: 10
+evals_per_epoch: 0
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/tiny-llama/lora.yml b/examples/tiny-llama/lora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..c5ff0437e8cc6dd892c3fa06a99f21a0f79abcf8
--- /dev/null
+++ b/examples/tiny-llama/lora.yml
@@ -0,0 +1,64 @@
+base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: true
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/lora-out
+
+sequence_len: 4096
+sample_packing: true
+eval_sample_packing: false
+pad_to_sequence_len: true
+
+adapter: lora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/tiny-llama/pretrain.yml b/examples/tiny-llama/pretrain.yml
new file mode 100644
index 0000000000000000000000000000000000000000..e501dcb8e590f7b156bf022b5ad2862ff33d3b5c
--- /dev/null
+++ b/examples/tiny-llama/pretrain.yml
@@ -0,0 +1,58 @@
+base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
+
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+max_steps: 200
+pretraining_dataset:
+ path: c4
+ name: en
+ type: pretrain
+dataset_prepared_path:
+val_set_size: 0.0
+output_dir: ./outputs/model-out
+
+sequence_len: 2048
+sample_packing: true
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: adamw_bnb_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch:
+eval_table_size:
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/tiny-llama/qlora.yml b/examples/tiny-llama/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..0d21aca9d5fe824891b1097d7dccdb8d63551c23
--- /dev/null
+++ b/examples/tiny-llama/qlora.yml
@@ -0,0 +1,65 @@
+base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./outputs/qlora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: true
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: paged_adamw_32bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
diff --git a/examples/xgen-7b/xgen-7b-8k-qlora.yml b/examples/xgen-7b/xgen-7b-8k-qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..7e3f83cbd7b2318c15d1c5ec904190ff5be09933
--- /dev/null
+++ b/examples/xgen-7b/xgen-7b-8k-qlora.yml
@@ -0,0 +1,90 @@
+# An example finetuning Saleforce's XGen-7b model with 8k context using qlora
+# on Tim Dettmer's Guanaco dataset.
+base_model: Salesforce/xgen-7b-8k-base
+trust_remote_code: true
+model_type: AutoModelForCausalLM
+tokenizer_type: AutoTokenizer
+load_in_8bit: false
+# enable 4bit for QLoRA
+load_in_4bit: true
+gptq: false
+strict: false
+push_dataset_to_hub:
+datasets:
+ - path: timdettmers/openassistant-guanaco
+ data_files:
+ - openassistant_best_replies_train.jsonl
+ type: "completion"
+dataset_prepared_path:
+val_set_size: 0.05
+# enable QLoRA
+adapter: qlora
+lora_model_dir:
+sequence_len: 8192
+max_packed_sequence_len:
+
+# hyperparameters from QLoRA paper Appendix B.2
+# "We find hyperparameters to be largely robust across datasets"
+lora_r: 64
+lora_alpha: 16
+# 0.1 for models up to 13B
+# 0.05 for 33B and 65B models
+lora_dropout: 0.05
+# add LoRA modules on all linear layers of the base model
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+output_dir: ./outputs/qlora-out
+
+# QLoRA paper Table 9
+# - 16 for 7b & 13b
+# - 32 for 33b, 64 for 64b
+# Max size tested on A6000
+# - 7b: 40
+# - 40b: 4
+# decrease if OOM, increase for max VRAM utilization
+micro_batch_size: 1
+gradient_accumulation_steps: 1
+num_epochs: 4
+# Optimizer for QLoRA
+optimizer: paged_adamw_32bit
+torchdistx_path:
+lr_scheduler: cosine
+# QLoRA paper Table 9
+# - 2e-4 for 7b & 13b
+# - 1e-4 for 33b & 64b
+learning_rate: 0.00002
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+gradient_checkpointing: true
+# stop training after this many evaluation losses have increased in a row
+# https://huggingface.co/transformers/v4.2.2/_modules/transformers/trainer_callback.html#EarlyStoppingCallback
+early_stopping_patience: 3
+resume_from_checkpoint:
+auto_resume_from_checkpoints: true
+local_rank:
+logging_steps: 1
+xformers_attention: true
+flash_attention:
+gptq_groupsize:
+gptq_model_v1:
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+special_tokens:
+ eos_token: "<|endoftext|>"
+ bos_token: "<|endoftext|>"
+ unk_token: "<|endoftext|>"
+ pad_token: "<|endoftext|>"
diff --git a/examples/yi-34B-chat/README.md b/examples/yi-34B-chat/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..07078850fbb9454b270253765e631bfcd78cf988
--- /dev/null
+++ b/examples/yi-34B-chat/README.md
@@ -0,0 +1,5 @@
+# Overview
+
+This is an example of a Yi-34B-Chat configuration. It demonstrates that it is possible to finetune a 34B model on a GPU with 24GB of VRAM.
+
+Tested on an RTX 4090 with `python -m axolotl.cli.train examples/mistral/qlora.yml`, a single epoch of finetuning on the alpaca dataset using qlora runs in 47 mins, using 97% of available memory.
diff --git a/examples/yi-34B-chat/qlora.yml b/examples/yi-34B-chat/qlora.yml
new file mode 100644
index 0000000000000000000000000000000000000000..7fe322d63dc9840c251f46136963902d556df0e3
--- /dev/null
+++ b/examples/yi-34B-chat/qlora.yml
@@ -0,0 +1,75 @@
+base_model: 01-ai/Yi-34B-Chat
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+sequence_len: 1024
+bf16: auto
+fp16:
+tf32: false
+flash_attention: true
+special_tokens:
+ bos_token: "<|startoftext|>"
+ eos_token: "<|endoftext|>"
+ unk_token: ""
+
+# Data
+datasets:
+ - path: mhenrichsen/alpaca_2k_test
+ type: alpaca
+warmup_steps: 10
+
+# Iterations
+num_epochs: 1
+
+# Evaluation
+val_set_size: 0.1
+evals_per_epoch: 5
+eval_table_size:
+eval_max_new_tokens: 128
+eval_sample_packing: false
+eval_batch_size: 1
+
+# LoRA
+output_dir: ./outputs/qlora-out
+adapter: qlora
+lora_model_dir:
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_linear: true
+lora_fan_in_fan_out:
+lora_target_modules:
+
+# Sampling
+sample_packing: false
+pad_to_sequence_len: false
+
+# Batching
+gradient_accumulation_steps: 4
+micro_batch_size: 1
+gradient_checkpointing: true
+
+# wandb
+wandb_project:
+
+# Optimizer
+optimizer: paged_adamw_8bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+# Misc
+train_on_inputs: false
+group_by_length: false
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+debug:
+deepspeed:
+weight_decay: 0
+fsdp:
+fsdp_config:
diff --git a/favicon.jpg b/favicon.jpg
new file mode 100644
index 0000000000000000000000000000000000000000..444ed30f508408136aba3a20a4acda4e353bb6f8
--- /dev/null
+++ b/favicon.jpg
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d46d77a60557862559b1ea8fc943a7314ffed4f8a11230d2f1cff65618bcc92a
+size 4638
diff --git a/image/axolotl-badge-web.png b/image/axolotl-badge-web.png
new file mode 100644
index 0000000000000000000000000000000000000000..df542796738bdc6a66757045cb368a81f10ccd50
--- /dev/null
+++ b/image/axolotl-badge-web.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e5b5112484a73ef67562237c9742e630507b50483a6041860e24af4ab7954f8e
+size 11656
diff --git a/image/axolotl.png b/image/axolotl.png
new file mode 100644
index 0000000000000000000000000000000000000000..124d8cc6b8ad3e6d3d97b1a184827aeea3b5a3d5
--- /dev/null
+++ b/image/axolotl.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9cb7956d3e2118b22e09817760c80a2d0998f545b692159d9b0b1276e21894af
+size 957136
diff --git a/image/sticker_fixed.png b/image/sticker_fixed.png
new file mode 100644
index 0000000000000000000000000000000000000000..bd4feb407612e4899f0ea74d6e0da6d9872f7ab6
--- /dev/null
+++ b/image/sticker_fixed.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6e6f0c1c795e9fa8b5c70a2ba7ab9b6ed33c7295dfa2c7552ace8546cd727245
+size 379330
diff --git a/index.qmd b/index.qmd
new file mode 100644
index 0000000000000000000000000000000000000000..ed23d235ad81a718d93c60c571790b8701888c33
--- /dev/null
+++ b/index.qmd
@@ -0,0 +1,23 @@
+---
+toc-location: right-body
+toc-title: Table Of Contents
+toc-expand: 2
+---
+
+```{python}
+#|output: asis
+#|echo: false
+
+# This cell steals the README as the home page for now, but excludes the table of contents (quarto adds its own)
+import re
+pattern = re.compile(
+ r"\s*\s*\s*## Table of Contents.*? \s* \s*
",
+ re.DOTALL | re.IGNORECASE
+)
+
+with open('README.md', 'r') as f:
+ txt = f.read()
+
+cleaned = pattern.sub("", txt)
+print(cleaned)
+```
diff --git a/requirements-dev.txt b/requirements-dev.txt
new file mode 100644
index 0000000000000000000000000000000000000000..4b5df167b61d26ce682651410411614ca84a7d92
--- /dev/null
+++ b/requirements-dev.txt
@@ -0,0 +1,4 @@
+pre-commit
+black
+mypy
+types-requests
diff --git a/requirements-tests.txt b/requirements-tests.txt
new file mode 100644
index 0000000000000000000000000000000000000000..e079f8a6038dd2dc8512967540f96ee0de172067
--- /dev/null
+++ b/requirements-tests.txt
@@ -0,0 +1 @@
+pytest
diff --git a/requirements.txt b/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..6723b179828824843a58b8caabac0a951d9b312c
--- /dev/null
+++ b/requirements.txt
@@ -0,0 +1,44 @@
+--extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
+packaging==23.2
+peft==0.10.0
+transformers==4.40.2
+tokenizers==0.19.1
+bitsandbytes==0.43.1
+accelerate==0.30.1
+deepspeed==0.14.2
+pydantic==2.6.3
+addict
+fire
+PyYAML>=6.0
+requests
+datasets==2.19.1
+flash-attn==2.5.8
+sentencepiece
+wandb
+einops
+xformers==0.0.23.post1
+optimum==1.16.2
+hf_transfer
+colorama
+numba
+numpy>=1.24.4
+# qlora things
+evaluate==0.4.1
+scipy
+scikit-learn==1.2.2
+pynvml
+art
+fschat @ git+https://github.com/lm-sys/FastChat.git@27a05b04a35510afb1d767ae7e5990cbd278f8fe
+gradio==3.50.2
+tensorboard
+
+mamba-ssm==1.2.0.post1
+
+# remote filesystems
+s3fs
+gcsfs
+# adlfs
+
+trl==0.8.5
+zstandard==0.22.0
+fastcore
diff --git a/scripts/cloud-entrypoint.sh b/scripts/cloud-entrypoint.sh
new file mode 100644
index 0000000000000000000000000000000000000000..5b0337f2b24e8d2f6ef7ef0817b8c77197ffe1e0
--- /dev/null
+++ b/scripts/cloud-entrypoint.sh
@@ -0,0 +1,80 @@
+#!/bin/bash
+
+# Export specific ENV variables to /etc/rp_environment
+echo "Exporting environment variables..."
+printenv | grep -E '^RUNPOD_|^PATH=|^_=' | sed 's/^\(.*\)=\(.*\)$/export \1="\2"/' >> /etc/rp_environment
+echo 'source /etc/rp_environment' >> ~/.bashrc
+
+add_keys_to_authorized() {
+ local key_value=$1
+
+ # Create the ~/.ssh directory and set permissions
+ mkdir -p ~/.ssh
+ chmod 700 ~/.ssh
+
+ # Create the authorized_keys file if it doesn't exist
+ touch ~/.ssh/authorized_keys
+
+ # Initialize an empty key variable
+ local key=""
+
+ # Read the key variable word by word
+ for word in $key_value; do
+ # Check if the word looks like the start of a key
+ if [[ $word == ssh-* ]]; then
+ # If there's a key being built, add it to the authorized_keys file
+ if [[ -n $key ]]; then
+ echo $key >> ~/.ssh/authorized_keys
+ fi
+ # Start a new key
+ key=$word
+ else
+ # Append the word to the current key
+ key="$key $word"
+ fi
+ done
+
+ # Add the last key to the authorized_keys file
+ if [[ -n $key ]]; then
+ echo $key >> ~/.ssh/authorized_keys
+ fi
+
+ # Set the correct permissions
+ chmod 600 ~/.ssh/authorized_keys
+ chmod 700 -R ~/.ssh
+}
+
+if [[ $PUBLIC_KEY ]]; then
+ # runpod
+ add_keys_to_authorized "$PUBLIC_KEY"
+ # Start the SSH service in the background
+ service ssh start
+elif [[ $SSH_KEY ]]; then
+ # latitude.sh
+ add_keys_to_authorized "$SSH_KEY"
+ # Start the SSH service in the background
+ service ssh start
+else
+ echo "No PUBLIC_KEY or SSH_KEY environment variable provided, not starting openSSH daemon"
+fi
+
+# Check if JUPYTER_PASSWORD is set and not empty
+if [ -n "$JUPYTER_PASSWORD" ]; then
+ # Set JUPYTER_TOKEN to the value of JUPYTER_PASSWORD
+ export JUPYTER_TOKEN="$JUPYTER_PASSWORD"
+fi
+
+if [ "$JUPYTER_DISABLE" != "1" ]; then
+ # Run Jupyter Lab in the background
+ jupyter lab --port=8888 --ip=* --allow-root --ServerApp.allow_origin=* &
+fi
+
+if [ ! -d "/workspace/data/axolotl-artifacts" ]; then
+ mkdir -p /workspace/data/axolotl-artifacts
+fi
+if [ ! -L "/workspace/axolotl/outputs" ]; then
+ ln -sf /workspace/data/axolotl-artifacts /workspace/axolotl/outputs
+fi
+
+# Execute the passed arguments (CMD)
+exec "$@"
diff --git a/scripts/finetune.py b/scripts/finetune.py
new file mode 100644
index 0000000000000000000000000000000000000000..d5bbcaf8f019b3b86733a8d70cdd6cff24b5d3a7
--- /dev/null
+++ b/scripts/finetune.py
@@ -0,0 +1,52 @@
+"""Prepare and train a model on a dataset. Can also infer from a model or merge lora"""
+import logging
+from pathlib import Path
+
+import fire
+import transformers
+
+from axolotl.cli import (
+ check_accelerate_default_config,
+ check_user_token,
+ do_inference,
+ do_merge_lora,
+ load_cfg,
+ load_datasets,
+ print_axolotl_text_art,
+)
+from axolotl.cli.shard import shard
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+
+LOG = logging.getLogger("axolotl.scripts.finetune")
+
+
+def do_cli(config: Path = Path("examples/"), **kwargs):
+ print_axolotl_text_art()
+ LOG.warning(
+ str(
+ PendingDeprecationWarning(
+ "scripts/finetune.py will be replaced with calling axolotl.cli.train"
+ )
+ )
+ )
+ parsed_cfg = load_cfg(config, **kwargs)
+ check_accelerate_default_config()
+ check_user_token()
+ parser = transformers.HfArgumentParser((TrainerCliArgs))
+ parsed_cli_args, _ = parser.parse_args_into_dataclasses(
+ return_remaining_strings=True
+ )
+ if parsed_cli_args.inference:
+ do_inference(cfg=parsed_cfg, cli_args=parsed_cli_args)
+ elif parsed_cli_args.merge_lora:
+ do_merge_lora(cfg=parsed_cfg, cli_args=parsed_cli_args)
+ elif parsed_cli_args.shard:
+ shard(cfg=parsed_cfg, cli_args=parsed_cli_args)
+ else:
+ dataset_meta = load_datasets(cfg=parsed_cfg, cli_args=parsed_cli_args)
+ train(cfg=parsed_cfg, cli_args=parsed_cli_args, dataset_meta=dataset_meta)
+
+
+if __name__ == "__main__":
+ fire.Fire(do_cli)
diff --git a/scripts/motd b/scripts/motd
new file mode 100644
index 0000000000000000000000000000000000000000..060a5c5c6c7e6452095cd0dd52c25631bc23d083
--- /dev/null
+++ b/scripts/motd
@@ -0,0 +1,17 @@
+
+ dP dP dP
+ 88 88 88
+ .d8888b. dP. .dP .d8888b. 88 .d8888b. d8888P 88
+ 88' `88 `8bd8' 88' `88 88 88' `88 88 88
+ 88. .88 .d88b. 88. .88 88 88. .88 88 88
+ `88888P8 dP' `dP `88888P' dP `88888P' dP dP
+
+Welcome to the axolotl cloud image! If the you've mounted a disk to /workspace and the axolotl directory ie empty, run the following commands:
+
+```
+cd /workspace
+rm -rf /workspace/axolotl
+git clone https://github.com/OpenAccess-AI-Collective/axolotl.git
+cd axolotl
+pip install --no-deps -e .
+```
diff --git a/setup.py b/setup.py
new file mode 100644
index 0000000000000000000000000000000000000000..31a6d671641c713a770bcdb608f20444a1627c9c
--- /dev/null
+++ b/setup.py
@@ -0,0 +1,99 @@
+"""setup.py for axolotl"""
+
+import platform
+import re
+from importlib.metadata import PackageNotFoundError, version
+
+from setuptools import find_packages, setup
+
+
+def parse_requirements():
+ _install_requires = []
+ _dependency_links = []
+ with open("./requirements.txt", encoding="utf-8") as requirements_file:
+ lines = [r.strip() for r in requirements_file.readlines()]
+ for line in lines:
+ is_extras = (
+ "flash-attn" in line
+ or "flash-attention" in line
+ or "deepspeed" in line
+ or "mamba-ssm" in line
+ or "lion-pytorch" in line
+ )
+ if line.startswith("--extra-index-url"):
+ # Handle custom index URLs
+ _, url = line.split()
+ _dependency_links.append(url)
+ elif not is_extras and line and line[0] != "#":
+ # Handle standard packages
+ _install_requires.append(line)
+
+ try:
+ if "Darwin" in platform.system():
+ _install_requires.pop(_install_requires.index("xformers==0.0.23.post1"))
+ else:
+ torch_version = version("torch")
+ _install_requires.append(f"torch=={torch_version}")
+
+ version_match = re.match(r"^(\d+)\.(\d+)(?:\.(\d+))?", torch_version)
+ if version_match:
+ major, minor, patch = version_match.groups()
+ major, minor = int(major), int(minor)
+ patch = (
+ int(patch) if patch is not None else 0
+ ) # Default patch to 0 if not present
+ else:
+ raise ValueError("Invalid version format")
+
+ if (major, minor) >= (2, 3):
+ _install_requires.pop(_install_requires.index("xformers==0.0.23.post1"))
+ _install_requires.append("xformers>=0.0.26.post1")
+ elif (major, minor) >= (2, 2):
+ _install_requires.pop(_install_requires.index("xformers==0.0.23.post1"))
+ _install_requires.append("xformers>=0.0.25.post1")
+ except PackageNotFoundError:
+ pass
+
+ return _install_requires, _dependency_links
+
+
+install_requires, dependency_links = parse_requirements()
+
+
+setup(
+ name="axolotl",
+ version="0.4.0",
+ description="LLM Trainer",
+ long_description="Axolotl is a tool designed to streamline the fine-tuning of various AI models, offering support for multiple configurations and architectures.",
+ package_dir={"": "src"},
+ packages=find_packages(),
+ install_requires=install_requires,
+ dependency_links=dependency_links,
+ extras_require={
+ "flash-attn": [
+ "flash-attn==2.5.8",
+ ],
+ "fused-dense-lib": [
+ "fused-dense-lib @ git+https://github.com/Dao-AILab/flash-attention@v2.5.8#subdirectory=csrc/fused_dense_lib",
+ ],
+ "deepspeed": [
+ "deepspeed==0.14.2",
+ "deepspeed-kernels",
+ ],
+ "mamba-ssm": [
+ "mamba-ssm==1.2.0.post1",
+ ],
+ "auto-gptq": [
+ "auto-gptq==0.5.1",
+ ],
+ "mlflow": [
+ "mlflow",
+ ],
+ "lion-pytorch": [
+ "lion-pytorch==0.1.2",
+ ],
+ "galore": [
+ "galore_torch",
+ ],
+ },
+)
diff --git a/src/axolotl.egg-info/PKG-INFO b/src/axolotl.egg-info/PKG-INFO
new file mode 100644
index 0000000000000000000000000000000000000000..45f79a239d7d2c617669440a967e2321f8bec9a3
--- /dev/null
+++ b/src/axolotl.egg-info/PKG-INFO
@@ -0,0 +1,59 @@
+Metadata-Version: 2.1
+Name: axolotl
+Version: 0.4.0
+Summary: LLM Trainer
+License-File: LICENSE
+Requires-Dist: packaging==23.2
+Requires-Dist: peft==0.10.0
+Requires-Dist: transformers==4.40.2
+Requires-Dist: tokenizers==0.19.1
+Requires-Dist: bitsandbytes==0.43.1
+Requires-Dist: accelerate==0.30.1
+Requires-Dist: pydantic==2.6.3
+Requires-Dist: addict
+Requires-Dist: fire
+Requires-Dist: PyYAML>=6.0
+Requires-Dist: requests
+Requires-Dist: datasets==2.19.1
+Requires-Dist: sentencepiece
+Requires-Dist: wandb
+Requires-Dist: einops
+Requires-Dist: xformers==0.0.23.post1
+Requires-Dist: optimum==1.16.2
+Requires-Dist: hf_transfer
+Requires-Dist: colorama
+Requires-Dist: numba
+Requires-Dist: numpy>=1.24.4
+Requires-Dist: evaluate==0.4.1
+Requires-Dist: scipy
+Requires-Dist: scikit-learn==1.2.2
+Requires-Dist: pynvml
+Requires-Dist: art
+Requires-Dist: fschat@ git+https://github.com/lm-sys/FastChat.git@27a05b04a35510afb1d767ae7e5990cbd278f8fe
+Requires-Dist: gradio==3.50.2
+Requires-Dist: tensorboard
+Requires-Dist: s3fs
+Requires-Dist: gcsfs
+Requires-Dist: trl==0.8.5
+Requires-Dist: zstandard==0.22.0
+Requires-Dist: fastcore
+Requires-Dist: torch==2.1.2+cu118
+Provides-Extra: flash-attn
+Requires-Dist: flash-attn==2.5.8; extra == "flash-attn"
+Provides-Extra: fused-dense-lib
+Requires-Dist: fused-dense-lib@ git+https://github.com/Dao-AILab/flash-attention@v2.5.8#subdirectory=csrc/fused_dense_lib ; extra == "fused-dense-lib"
+Provides-Extra: deepspeed
+Requires-Dist: deepspeed==0.14.2; extra == "deepspeed"
+Requires-Dist: deepspeed-kernels; extra == "deepspeed"
+Provides-Extra: mamba-ssm
+Requires-Dist: mamba-ssm==1.2.0.post1; extra == "mamba-ssm"
+Provides-Extra: auto-gptq
+Requires-Dist: auto-gptq==0.5.1; extra == "auto-gptq"
+Provides-Extra: mlflow
+Requires-Dist: mlflow; extra == "mlflow"
+Provides-Extra: lion-pytorch
+Requires-Dist: lion-pytorch==0.1.2; extra == "lion-pytorch"
+Provides-Extra: galore
+Requires-Dist: galore_torch; extra == "galore"
+
+Axolotl is a tool designed to streamline the fine-tuning of various AI models, offering support for multiple configurations and architectures.
diff --git a/src/axolotl.egg-info/SOURCES.txt b/src/axolotl.egg-info/SOURCES.txt
new file mode 100644
index 0000000000000000000000000000000000000000..d93a2948554d0e2ffbcaa3604cd4184a4e79169e
--- /dev/null
+++ b/src/axolotl.egg-info/SOURCES.txt
@@ -0,0 +1,22 @@
+LICENSE
+README.md
+setup.py
+src/axolotl.egg-info/PKG-INFO
+src/axolotl.egg-info/SOURCES.txt
+src/axolotl.egg-info/dependency_links.txt
+src/axolotl.egg-info/requires.txt
+src/axolotl.egg-info/top_level.txt
+tests/test_data.py
+tests/test_datasets.py
+tests/test_dict.py
+tests/test_expand_mask.py
+tests/test_freeze.py
+tests/test_normalize_config.py
+tests/test_packed_batch_sampler.py
+tests/test_packed_dataset.py
+tests/test_packed_pretraining.py
+tests/test_prompt_tokenizers.py
+tests/test_prompters.py
+tests/test_schedulers.py
+tests/test_tokenizers.py
+tests/test_validation.py
\ No newline at end of file
diff --git a/src/axolotl.egg-info/dependency_links.txt b/src/axolotl.egg-info/dependency_links.txt
new file mode 100644
index 0000000000000000000000000000000000000000..513af3175af1d5c7d2d6ecc19b20e8730983a6ae
--- /dev/null
+++ b/src/axolotl.egg-info/dependency_links.txt
@@ -0,0 +1 @@
+https://huggingface.github.io/autogptq-index/whl/cu118/
diff --git a/src/axolotl.egg-info/requires.txt b/src/axolotl.egg-info/requires.txt
new file mode 100644
index 0000000000000000000000000000000000000000..621ee6c2f68617e4e89b19e3a44d4d31da232ac5
--- /dev/null
+++ b/src/axolotl.egg-info/requires.txt
@@ -0,0 +1,60 @@
+packaging==23.2
+peft==0.10.0
+transformers==4.40.2
+tokenizers==0.19.1
+bitsandbytes==0.43.1
+accelerate==0.30.1
+pydantic==2.6.3
+addict
+fire
+PyYAML>=6.0
+requests
+datasets==2.19.1
+sentencepiece
+wandb
+einops
+xformers==0.0.23.post1
+optimum==1.16.2
+hf_transfer
+colorama
+numba
+numpy>=1.24.4
+evaluate==0.4.1
+scipy
+scikit-learn==1.2.2
+pynvml
+art
+fschat@ git+https://github.com/lm-sys/FastChat.git@27a05b04a35510afb1d767ae7e5990cbd278f8fe
+gradio==3.50.2
+tensorboard
+s3fs
+gcsfs
+trl==0.8.5
+zstandard==0.22.0
+fastcore
+torch==2.1.2+cu118
+
+[auto-gptq]
+auto-gptq==0.5.1
+
+[deepspeed]
+deepspeed==0.14.2
+deepspeed-kernels
+
+[flash-attn]
+flash-attn==2.5.8
+
+[fused-dense-lib]
+fused-dense-lib@ git+https://github.com/Dao-AILab/flash-attention@v2.5.8#subdirectory=csrc/fused_dense_lib
+
+[galore]
+galore_torch
+
+[lion-pytorch]
+lion-pytorch==0.1.2
+
+[mamba-ssm]
+mamba-ssm==1.2.0.post1
+
+[mlflow]
+mlflow
diff --git a/src/axolotl.egg-info/top_level.txt b/src/axolotl.egg-info/top_level.txt
new file mode 100644
index 0000000000000000000000000000000000000000..8b137891791fe96927ad78e64b0aad7bded08bdc
--- /dev/null
+++ b/src/axolotl.egg-info/top_level.txt
@@ -0,0 +1 @@
+
diff --git a/src/axolotl/__init__.py b/src/axolotl/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/axolotl/__pycache__/__init__.cpython-310.pyc b/src/axolotl/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..e55c70958a19e59b2c5046d84867dfa7e3291e61
Binary files /dev/null and b/src/axolotl/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/__pycache__/datasets.cpython-310.pyc b/src/axolotl/__pycache__/datasets.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..22f4a1acde1c777a90eb11b022022272d293dc8f
Binary files /dev/null and b/src/axolotl/__pycache__/datasets.cpython-310.pyc differ
diff --git a/src/axolotl/__pycache__/logging_config.cpython-310.pyc b/src/axolotl/__pycache__/logging_config.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..8bba57bcd6e75c8be0c9dba2b115227844a145bc
Binary files /dev/null and b/src/axolotl/__pycache__/logging_config.cpython-310.pyc differ
diff --git a/src/axolotl/__pycache__/loraplus.cpython-310.pyc b/src/axolotl/__pycache__/loraplus.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..1536e29a77647800d6cb16e47eced31f1cda862c
Binary files /dev/null and b/src/axolotl/__pycache__/loraplus.cpython-310.pyc differ
diff --git a/src/axolotl/__pycache__/prompt_tokenizers.cpython-310.pyc b/src/axolotl/__pycache__/prompt_tokenizers.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..b6f833bcbb52e6df1788fd97089ec1bb137da9ea
Binary files /dev/null and b/src/axolotl/__pycache__/prompt_tokenizers.cpython-310.pyc differ
diff --git a/src/axolotl/__pycache__/prompters.cpython-310.pyc b/src/axolotl/__pycache__/prompters.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..dafec63d07e3050dde478a280d8d71705fa91b4b
Binary files /dev/null and b/src/axolotl/__pycache__/prompters.cpython-310.pyc differ
diff --git a/src/axolotl/__pycache__/train.cpython-310.pyc b/src/axolotl/__pycache__/train.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..3eea5ddd11ef6bce6ac8d6638249ffa21b90d9d8
Binary files /dev/null and b/src/axolotl/__pycache__/train.cpython-310.pyc differ
diff --git a/src/axolotl/cli/__init__.py b/src/axolotl/cli/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..7ec3f524ab5b0f6171be71c5a4d8426eb754d3c8
--- /dev/null
+++ b/src/axolotl/cli/__init__.py
@@ -0,0 +1,490 @@
+"""Prepare and train a model on a dataset. Can also infer from a model or merge lora"""
+
+import importlib
+import json
+import logging
+import math
+import os
+import random
+import sys
+import tempfile
+from pathlib import Path
+from threading import Thread
+from typing import Any, Dict, List, Optional, Union
+from urllib.parse import urlparse
+
+import requests
+import torch
+import yaml
+
+# add src to the pythonpath so we don't need to pip install this
+from accelerate.commands.config import config_args
+from art import text2art
+from huggingface_hub import HfApi
+from huggingface_hub.utils import LocalTokenNotFoundError
+from transformers import GenerationConfig, TextIteratorStreamer, TextStreamer
+from transformers.utils import is_torch_bf16_gpu_available
+from transformers.utils.import_utils import _is_package_available
+
+from axolotl.common.cli import TrainerCliArgs, load_model_and_tokenizer
+from axolotl.logging_config import configure_logging
+from axolotl.train import TrainDatasetMeta
+from axolotl.utils.config import (
+ normalize_cfg_datasets,
+ normalize_config,
+ validate_config,
+)
+from axolotl.utils.data import load_prepare_dpo_datasets, prepare_dataset
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.distributed import is_main_process
+from axolotl.utils.mlflow_ import setup_mlflow_env_vars
+from axolotl.utils.models import load_tokenizer
+from axolotl.utils.tokenization import check_dataset_labels
+from axolotl.utils.trainer import prepare_optim_env
+from axolotl.utils.wandb_ import setup_wandb_env_vars
+
+project_root = os.path.abspath(os.path.join(os.path.dirname(__file__), ".."))
+src_dir = os.path.join(project_root, "src")
+sys.path.insert(0, src_dir)
+
+configure_logging()
+LOG = logging.getLogger("axolotl.scripts")
+
+os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
+
+
+def print_axolotl_text_art(suffix=None):
+ font = "nancyj"
+ ascii_text = " axolotl"
+ if suffix:
+ ascii_text += f" x {suffix}"
+ ascii_art = text2art(ascii_text, font=font)
+
+ if is_main_process():
+ print(ascii_art)
+
+ print_dep_versions()
+
+
+def print_dep_versions():
+ packages = ["accelerate", "peft", "transformers", "trl", "torch", "bitsandbytes"]
+ max_len = max(len(pkg) for pkg in packages)
+ if is_main_process():
+ print("*" * 40)
+ print("**** Axolotl Dependency Versions *****")
+ for pkg in packages:
+ version = _is_package_available(pkg, return_version=True)
+ print(f"{pkg: >{max_len}}: {version[1]: <15}")
+ print("*" * 40)
+
+
+def check_remote_config(config: Union[str, Path]):
+ # Check if the config is a valid HTTPS URL to a .yml or .yaml file
+ if not (isinstance(config, str) and config.startswith("https://")):
+ return config # Return the original value if it's not a valid URL
+
+ filename = os.path.basename(urlparse(config).path)
+ temp_dir = tempfile.mkdtemp()
+
+ try:
+ response = requests.get(config, timeout=30)
+ response.raise_for_status() # Check for HTTP errors
+
+ content = response.content
+ try:
+ # Try parsing as JSON first to catch cases where JSON content is mistakenly considered YAML
+ json.loads(content)
+ # Log a warning but do not raise an error; JSON is technically valid YAML - this can happen when you forget to point to a raw github link
+ LOG.warning(
+ f"Warning: The content of the file at {config} is JSON, which is technically valid YAML but might not be intended."
+ )
+ except json.JSONDecodeError:
+ # If it's not valid JSON, verify it's valid YAML
+ try:
+ yaml.safe_load(content)
+ except yaml.YAMLError as err:
+ raise ValueError(
+ f"Failed to parse the content at {config} as YAML: {err}"
+ ) from err
+
+ # Write the content to a file if it's valid YAML (or JSON treated as YAML)
+ output_path = Path(temp_dir) / filename
+ with open(output_path, "wb") as file:
+ file.write(content)
+ LOG.info(
+ f"Using the following config obtained from {config}:\n\n{content.decode('utf-8')}\n"
+ )
+ return output_path
+
+ except requests.RequestException as err:
+ # This catches all requests-related exceptions including HTTPError
+ raise RuntimeError(f"Failed to download {config}: {err}") from err
+ except Exception as err:
+ # Catch-all for any other exceptions
+ raise err
+
+
+def get_multi_line_input() -> Optional[str]:
+ print("Give me an instruction (Ctrl + D to submit): ")
+ instruction = ""
+ for line in sys.stdin:
+ instruction += line # pylint: disable=consider-using-join
+ # instruction = pathlib.Path("/proc/self/fd/0").read_text()
+ return instruction
+
+
+def do_merge_lora(
+ *,
+ cfg: DictDefault,
+ cli_args: TrainerCliArgs,
+):
+ model, tokenizer = load_model_and_tokenizer(cfg=cfg, cli_args=cli_args)
+ safe_serialization = cfg.save_safetensors is True
+
+ LOG.info("running merge of LoRA with base model")
+ model = model.merge_and_unload(progressbar=True)
+ try:
+ model.to(dtype=cfg.torch_dtype)
+ except RuntimeError:
+ pass
+ model.generation_config.do_sample = True
+
+ if cfg.local_rank == 0:
+ LOG.info(f"saving merged model to: {str(Path(cfg.output_dir) / 'merged')}")
+ model.save_pretrained(
+ str(Path(cfg.output_dir) / "merged"),
+ safe_serialization=safe_serialization,
+ progressbar=True,
+ )
+ tokenizer.save_pretrained(str(Path(cfg.output_dir) / "merged"))
+
+
+def do_inference(
+ *,
+ cfg: DictDefault,
+ cli_args: TrainerCliArgs,
+):
+ model, tokenizer = load_model_and_tokenizer(cfg=cfg, cli_args=cli_args)
+ prompter = cli_args.prompter
+ default_tokens = {"unk_token": "", "bos_token": "", "eos_token": " "}
+
+ for token, symbol in default_tokens.items():
+ # If the token isn't already specified in the config, add it
+ if not (cfg.special_tokens and token in cfg.special_tokens):
+ tokenizer.add_special_tokens({token: symbol})
+
+ prompter_module = None
+ if prompter:
+ prompter_module = getattr(
+ importlib.import_module("axolotl.prompters"), prompter
+ )
+
+ model = model.to(cfg.device, dtype=cfg.torch_dtype)
+
+ while True:
+ print("=" * 80)
+ # support for multiline inputs
+ instruction = get_multi_line_input()
+ if not instruction:
+ return
+ if prompter_module:
+ prompt: str = next(
+ prompter_module().build_prompt(instruction=instruction.strip("\n"))
+ )
+ else:
+ prompt = instruction.strip()
+ batch = tokenizer(prompt, return_tensors="pt", add_special_tokens=True)
+
+ print("=" * 40)
+ model.eval()
+ with torch.no_grad():
+ generation_config = GenerationConfig(
+ repetition_penalty=1.1,
+ max_new_tokens=1024,
+ temperature=0.9,
+ top_p=0.95,
+ top_k=40,
+ bos_token_id=tokenizer.bos_token_id,
+ eos_token_id=tokenizer.eos_token_id,
+ pad_token_id=tokenizer.pad_token_id,
+ do_sample=True,
+ use_cache=True,
+ return_dict_in_generate=True,
+ output_attentions=False,
+ output_hidden_states=False,
+ output_scores=False,
+ )
+ streamer = TextStreamer(tokenizer)
+ generated = model.generate(
+ inputs=batch["input_ids"].to(cfg.device),
+ generation_config=generation_config,
+ streamer=streamer,
+ )
+ print("=" * 40)
+ print(tokenizer.decode(generated["sequences"].cpu().tolist()[0]))
+
+
+def do_inference_gradio(
+ *,
+ cfg: DictDefault,
+ cli_args: TrainerCliArgs,
+):
+ import gradio as gr
+
+ model, tokenizer = load_model_and_tokenizer(cfg=cfg, cli_args=cli_args)
+ prompter = cli_args.prompter
+ default_tokens = {"unk_token": "", "bos_token": "", "eos_token": " "}
+
+ for token, symbol in default_tokens.items():
+ # If the token isn't already specified in the config, add it
+ if not (cfg.special_tokens and token in cfg.special_tokens):
+ tokenizer.add_special_tokens({token: symbol})
+
+ prompter_module = None
+ if prompter:
+ prompter_module = getattr(
+ importlib.import_module("axolotl.prompters"), prompter
+ )
+
+ model = model.to(cfg.device, dtype=cfg.torch_dtype)
+
+ def generate(instruction):
+ if not instruction:
+ return
+ if prompter_module:
+ # pylint: disable=stop-iteration-return
+ prompt: str = next(
+ prompter_module().build_prompt(instruction=instruction.strip("\n"))
+ )
+ else:
+ prompt = instruction.strip()
+ batch = tokenizer(prompt, return_tensors="pt", add_special_tokens=True)
+
+ model.eval()
+ with torch.no_grad():
+ generation_config = GenerationConfig(
+ repetition_penalty=1.1,
+ max_new_tokens=cfg.get("gradio_max_new_tokens", 1024),
+ temperature=cfg.get("gradio_temperature", 0.9),
+ top_p=0.95,
+ top_k=40,
+ bos_token_id=tokenizer.bos_token_id,
+ eos_token_id=tokenizer.eos_token_id,
+ pad_token_id=tokenizer.pad_token_id,
+ do_sample=True,
+ use_cache=True,
+ return_dict_in_generate=True,
+ output_attentions=False,
+ output_hidden_states=False,
+ output_scores=False,
+ )
+ streamer = TextIteratorStreamer(tokenizer)
+ generation_kwargs = {
+ "inputs": batch["input_ids"].to(cfg.device),
+ "generation_config": generation_config,
+ "streamer": streamer,
+ }
+
+ thread = Thread(target=model.generate, kwargs=generation_kwargs)
+ thread.start()
+
+ all_text = ""
+
+ for new_text in streamer:
+ all_text += new_text
+ yield all_text
+
+ demo = gr.Interface(
+ fn=generate,
+ inputs="textbox",
+ outputs="text",
+ title=cfg.get("gradio_title", "Axolotl Gradio Interface"),
+ )
+
+ demo.queue().launch(
+ show_api=False,
+ share=cfg.get("gradio_share", True),
+ server_name=cfg.get("gradio_server_name", "127.0.0.1"),
+ server_port=cfg.get("gradio_server_port", None),
+ )
+
+
+def choose_config(path: Path):
+ yaml_files = list(path.glob("*.yml"))
+
+ if not yaml_files:
+ raise ValueError(
+ "No YAML config files found in the specified directory. Are you using a .yml extension?"
+ )
+
+ if len(yaml_files) == 1:
+ print(f"Using default YAML file '{yaml_files[0]}'")
+ return yaml_files[0]
+
+ print("Choose a YAML file:")
+ for idx, file in enumerate(yaml_files):
+ print(f"{idx + 1}. {file}")
+
+ chosen_file = None
+ while chosen_file is None:
+ try:
+ choice = int(input("Enter the number of your choice: "))
+ if 1 <= choice <= len(yaml_files):
+ chosen_file = yaml_files[choice - 1]
+ else:
+ print("Invalid choice. Please choose a number from the list.")
+ except ValueError:
+ print("Invalid input. Please enter a number.")
+
+ return chosen_file
+
+
+def check_not_in(list1: List[str], list2: Union[Dict[str, Any], List[str]]) -> bool:
+ return not any(el in list2 for el in list1)
+
+
+def load_cfg(config: Union[str, Path] = Path("examples/"), **kwargs):
+ config = check_remote_config(config)
+ if Path(config).is_dir():
+ config = choose_config(Path(config))
+
+ # load the config from the yaml file
+ with open(config, encoding="utf-8") as file:
+ cfg: DictDefault = DictDefault(yaml.safe_load(file))
+ # if there are any options passed in the cli, if it is something that seems valid from the yaml,
+ # then overwrite the value
+ cfg_keys = cfg.keys()
+ for k, _ in kwargs.items():
+ # if not strict, allow writing to cfg even if it's not in the yml already
+ if k in cfg_keys or not cfg.strict:
+ # handle booleans
+ if isinstance(cfg[k], bool):
+ cfg[k] = bool(kwargs[k])
+ else:
+ cfg[k] = kwargs[k]
+
+ cfg.axolotl_config_path = config
+
+ try:
+ device_props = torch.cuda.get_device_properties("cuda")
+ gpu_version = "sm_" + str(device_props.major) + str(device_props.minor)
+ except: # pylint: disable=bare-except # noqa: E722
+ gpu_version = None
+
+ cfg = validate_config(
+ cfg,
+ capabilities={
+ "bf16": is_torch_bf16_gpu_available(),
+ "n_gpu": os.environ.get("WORLD_SIZE", 1),
+ "compute_capability": gpu_version,
+ },
+ )
+
+ prepare_optim_env(cfg)
+
+ normalize_config(cfg)
+
+ normalize_cfg_datasets(cfg)
+
+ setup_wandb_env_vars(cfg)
+
+ setup_mlflow_env_vars(cfg)
+
+ return cfg
+
+
+def load_datasets(
+ *,
+ cfg: DictDefault,
+ cli_args: TrainerCliArgs,
+) -> TrainDatasetMeta:
+ tokenizer = load_tokenizer(cfg)
+
+ train_dataset, eval_dataset, total_num_steps, prompters = prepare_dataset(
+ cfg, tokenizer
+ )
+
+ if cli_args.debug or cfg.debug:
+ LOG.info("check_dataset_labels...")
+ check_dataset_labels(
+ train_dataset.select(
+ [
+ random.randrange(0, len(train_dataset) - 1) # nosec
+ for _ in range(cli_args.debug_num_examples)
+ ]
+ ),
+ tokenizer,
+ num_examples=cli_args.debug_num_examples,
+ text_only=cli_args.debug_text_only,
+ )
+
+ LOG.info("printing prompters...")
+ for prompter in prompters:
+ LOG.info(prompter)
+
+ return TrainDatasetMeta(
+ train_dataset=train_dataset,
+ eval_dataset=eval_dataset,
+ total_num_steps=total_num_steps,
+ )
+
+
+def load_rl_datasets(
+ *,
+ cfg: DictDefault,
+ cli_args: TrainerCliArgs, # pylint: disable=unused-argument
+) -> TrainDatasetMeta:
+ train_dataset, eval_dataset = load_prepare_dpo_datasets(cfg)
+ total_num_steps = int(
+ math.ceil(len(train_dataset) * cfg.num_epochs / cfg.batch_size)
+ )
+
+ if cli_args.debug or cfg.debug:
+ LOG.info("check_dataset_labels...")
+
+ tokenizer = load_tokenizer(cfg)
+ check_dataset_labels(
+ train_dataset.select(
+ [
+ random.randrange(0, len(train_dataset) - 1) # nosec
+ for _ in range(cli_args.debug_num_examples)
+ ]
+ ),
+ tokenizer,
+ num_examples=cli_args.debug_num_examples,
+ text_only=cli_args.debug_text_only,
+ rl_mode=True,
+ )
+
+ return TrainDatasetMeta(
+ train_dataset=train_dataset,
+ eval_dataset=eval_dataset,
+ total_num_steps=total_num_steps,
+ )
+
+
+def check_accelerate_default_config():
+ if Path(config_args.default_yaml_config_file).exists():
+ LOG.warning(
+ f"accelerate config file found at {config_args.default_yaml_config_file}. This can lead to unexpected errors"
+ )
+
+
+def check_user_token():
+ # Skip check if HF_HUB_OFFLINE is set to True
+ if os.getenv("HF_HUB_OFFLINE") == "1":
+ LOG.info(
+ "Skipping HuggingFace token verification because HF_HUB_OFFLINE is set to True. Only local files will be used."
+ )
+ return True
+
+ # Verify if token is valid
+ api = HfApi()
+ try:
+ user_info = api.whoami()
+ return bool(user_info)
+ except LocalTokenNotFoundError:
+ LOG.warning(
+ "Error verifying HuggingFace token. Remember to log in using `huggingface-cli login` and get your access token from https://huggingface.co/settings/tokens if you want to use gated models or datasets."
+ )
+ return False
diff --git a/src/axolotl/cli/__pycache__/__init__.cpython-310.pyc b/src/axolotl/cli/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..bc09fb5abb475249d80df3404178ff57572a6f77
Binary files /dev/null and b/src/axolotl/cli/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/cli/__pycache__/preprocess.cpython-310.pyc b/src/axolotl/cli/__pycache__/preprocess.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..3c1aea63e37601aaffcdaade11c4bb085843f03e
Binary files /dev/null and b/src/axolotl/cli/__pycache__/preprocess.cpython-310.pyc differ
diff --git a/src/axolotl/cli/__pycache__/train.cpython-310.pyc b/src/axolotl/cli/__pycache__/train.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..8be2fb10b46776c37b8ed14658b91e84aeb771d3
Binary files /dev/null and b/src/axolotl/cli/__pycache__/train.cpython-310.pyc differ
diff --git a/src/axolotl/cli/inference.py b/src/axolotl/cli/inference.py
new file mode 100644
index 0000000000000000000000000000000000000000..86ad8409ff386ef857214207f460c22843af6be3
--- /dev/null
+++ b/src/axolotl/cli/inference.py
@@ -0,0 +1,36 @@
+"""
+CLI to run inference on a trained model
+"""
+from pathlib import Path
+
+import fire
+import transformers
+
+from axolotl.cli import (
+ do_inference,
+ do_inference_gradio,
+ load_cfg,
+ print_axolotl_text_art,
+)
+from axolotl.common.cli import TrainerCliArgs
+
+
+def do_cli(config: Path = Path("examples/"), gradio=False, **kwargs):
+ # pylint: disable=duplicate-code
+ print_axolotl_text_art()
+ parsed_cfg = load_cfg(config, **kwargs)
+ parsed_cfg.sample_packing = False
+ parser = transformers.HfArgumentParser((TrainerCliArgs))
+ parsed_cli_args, _ = parser.parse_args_into_dataclasses(
+ return_remaining_strings=True
+ )
+ parsed_cli_args.inference = True
+
+ if gradio:
+ do_inference_gradio(cfg=parsed_cfg, cli_args=parsed_cli_args)
+ else:
+ do_inference(cfg=parsed_cfg, cli_args=parsed_cli_args)
+
+
+if __name__ == "__main__":
+ fire.Fire(do_cli)
diff --git a/src/axolotl/cli/merge_lora.py b/src/axolotl/cli/merge_lora.py
new file mode 100644
index 0000000000000000000000000000000000000000..8db3fa989722c2f6414b5ac2e353401bd78b268c
--- /dev/null
+++ b/src/axolotl/cli/merge_lora.py
@@ -0,0 +1,51 @@
+"""
+CLI to run merge a trained LoRA into a base model
+"""
+from pathlib import Path
+
+import fire
+import transformers
+
+from axolotl.cli import do_merge_lora, load_cfg, print_axolotl_text_art
+from axolotl.common.cli import TrainerCliArgs
+
+
+def do_cli(config: Path = Path("examples/"), **kwargs):
+ # pylint: disable=duplicate-code
+ print_axolotl_text_art()
+ parser = transformers.HfArgumentParser((TrainerCliArgs))
+ parsed_cli_args, _ = parser.parse_args_into_dataclasses(
+ return_remaining_strings=True
+ )
+ parsed_cli_args.merge_lora = True
+
+ parsed_cfg = load_cfg(
+ config,
+ merge_lora=True,
+ load_in_8bit=False,
+ load_in_4bit=False,
+ flash_attention=False,
+ deepspeed=None,
+ fsdp=None,
+ **kwargs,
+ )
+
+ if not parsed_cfg.lora_model_dir and parsed_cfg.output_dir:
+ parsed_cfg.lora_model_dir = parsed_cfg.output_dir
+ if not Path(parsed_cfg.lora_model_dir).exists():
+ raise ValueError(
+ f"Target directory for merge: `{parsed_cfg.lora_model_dir}` does not exist."
+ )
+
+ parsed_cfg.load_in_4bit = False
+ parsed_cfg.load_in_8bit = False
+ parsed_cfg.flash_attention = False
+ parsed_cfg.deepspeed = None
+ parsed_cfg.fsdp = None
+ parsed_cfg.fsdp_config = None
+
+ do_merge_lora(cfg=parsed_cfg, cli_args=parsed_cli_args)
+
+
+if __name__ == "__main__":
+ fire.Fire(do_cli)
diff --git a/src/axolotl/cli/preprocess.py b/src/axolotl/cli/preprocess.py
new file mode 100644
index 0000000000000000000000000000000000000000..e7b3596a4f3396e48dc00af1c3bf60e85298beb3
--- /dev/null
+++ b/src/axolotl/cli/preprocess.py
@@ -0,0 +1,82 @@
+"""
+CLI to run training on a model
+"""
+import logging
+from pathlib import Path
+from typing import Union
+
+import fire
+import transformers
+from colorama import Fore
+
+from axolotl.cli import (
+ check_accelerate_default_config,
+ check_user_token,
+ load_cfg,
+ load_datasets,
+ load_rl_datasets,
+ print_axolotl_text_art,
+)
+from axolotl.common.cli import PreprocessCliArgs
+from axolotl.common.const import DEFAULT_DATASET_PREPARED_PATH
+from axolotl.prompt_strategies.sharegpt import (
+ register_chatml_template,
+ register_llama3_template,
+)
+
+LOG = logging.getLogger("axolotl.cli.preprocess")
+
+
+def do_cli(config: Union[Path, str] = Path("examples/"), **kwargs):
+ # pylint: disable=duplicate-code
+ print_axolotl_text_art()
+ parsed_cfg = load_cfg(config, **kwargs)
+ parsed_cfg.is_preprocess = True
+ check_accelerate_default_config()
+ check_user_token()
+ parser = transformers.HfArgumentParser((PreprocessCliArgs))
+ parsed_cli_args, _ = parser.parse_args_into_dataclasses(
+ return_remaining_strings=True
+ )
+
+ if parsed_cfg.chat_template == "chatml":
+ if parsed_cfg.default_system_message:
+ LOG.info(
+ f"ChatML set. Adding default system message: {parsed_cfg.default_system_message}"
+ )
+ register_chatml_template(parsed_cfg.default_system_message)
+ else:
+ register_chatml_template()
+ elif parsed_cfg.chat_template == "llama3":
+ if parsed_cfg.default_system_message:
+ LOG.info(
+ f"LLaMA-3 set. Adding default system message: {parsed_cfg.default_system_message}"
+ )
+ register_llama3_template(parsed_cfg.default_system_message)
+ else:
+ register_llama3_template()
+
+ if not parsed_cfg.dataset_prepared_path:
+ msg = (
+ Fore.RED
+ + "preprocess CLI called without dataset_prepared_path set, "
+ + f"using default path: {DEFAULT_DATASET_PREPARED_PATH}"
+ + Fore.RESET
+ )
+ LOG.warning(msg)
+ parsed_cfg.dataset_prepared_path = DEFAULT_DATASET_PREPARED_PATH
+
+ if parsed_cfg.rl: # and parsed_cfg.rl != "orpo":
+ load_rl_datasets(cfg=parsed_cfg, cli_args=parsed_cli_args)
+ else:
+ load_datasets(cfg=parsed_cfg, cli_args=parsed_cli_args)
+
+ LOG.info(
+ Fore.GREEN
+ + f"Success! Preprocessed data path: `dataset_prepared_path: {parsed_cfg.dataset_prepared_path}`"
+ + Fore.RESET
+ )
+
+
+if __name__ == "__main__":
+ fire.Fire(do_cli)
diff --git a/src/axolotl/cli/shard.py b/src/axolotl/cli/shard.py
new file mode 100644
index 0000000000000000000000000000000000000000..48f22790ac50d99fd0630b9c78d832430b198985
--- /dev/null
+++ b/src/axolotl/cli/shard.py
@@ -0,0 +1,43 @@
+"""
+CLI to shard a trained model into 10GiB chunks
+"""
+import logging
+from pathlib import Path
+from typing import Union
+
+import fire
+import transformers
+
+from axolotl.cli import load_cfg, print_axolotl_text_art
+from axolotl.common.cli import TrainerCliArgs, load_model_and_tokenizer
+from axolotl.utils.dict import DictDefault
+
+LOG = logging.getLogger("axolotl.scripts")
+
+
+def shard(
+ *,
+ cfg: DictDefault,
+ cli_args: TrainerCliArgs,
+):
+ model, _ = load_model_and_tokenizer(cfg=cfg, cli_args=cli_args)
+ safe_serialization = cfg.save_safetensors is True
+ LOG.debug("Re-saving model w/ sharding")
+ model.save_pretrained(cfg.output_dir, safe_serialization=safe_serialization)
+
+
+def do_cli(config: Union[Path, str] = Path("examples/"), **kwargs):
+ # pylint: disable=duplicate-code
+ print_axolotl_text_art()
+ parsed_cfg = load_cfg(config, **kwargs)
+ parser = transformers.HfArgumentParser((TrainerCliArgs))
+ parsed_cli_args, _ = parser.parse_args_into_dataclasses(
+ return_remaining_strings=True
+ )
+ parsed_cli_args.shard = True
+
+ shard(cfg=parsed_cfg, cli_args=parsed_cli_args)
+
+
+if __name__ == "__main__":
+ fire.Fire(do_cli)
diff --git a/src/axolotl/cli/train.py b/src/axolotl/cli/train.py
new file mode 100644
index 0000000000000000000000000000000000000000..7bb4a51844586fbb691767893b8e5f49682e12c3
--- /dev/null
+++ b/src/axolotl/cli/train.py
@@ -0,0 +1,70 @@
+"""
+CLI to run training on a model
+"""
+import logging
+from pathlib import Path
+from typing import Tuple, Union
+
+import fire
+from transformers.hf_argparser import HfArgumentParser
+from transformers.modeling_utils import PreTrainedModel
+from transformers.tokenization_utils import PreTrainedTokenizer
+
+from axolotl.cli import (
+ check_accelerate_default_config,
+ check_user_token,
+ load_cfg,
+ load_datasets,
+ load_rl_datasets,
+ print_axolotl_text_art,
+)
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.prompt_strategies.sharegpt import (
+ register_chatml_template,
+ register_llama3_template,
+)
+from axolotl.train import train
+
+LOG = logging.getLogger("axolotl.cli.train")
+
+
+def do_cli(config: Union[Path, str] = Path("examples/"), **kwargs):
+ # pylint: disable=duplicate-code
+ parsed_cfg = load_cfg(config, **kwargs)
+ parser = HfArgumentParser((TrainerCliArgs))
+ parsed_cli_args, _ = parser.parse_args_into_dataclasses(
+ return_remaining_strings=True
+ )
+ return do_train(parsed_cfg, parsed_cli_args)
+
+
+def do_train(cfg, cli_args) -> Tuple[PreTrainedModel, PreTrainedTokenizer]:
+ print_axolotl_text_art()
+ check_accelerate_default_config()
+ check_user_token()
+ if cfg.chat_template == "chatml" and cfg.default_system_message:
+ LOG.info(
+ f"ChatML set. Adding default system message: {cfg.default_system_message}"
+ )
+ register_chatml_template(cfg.default_system_message)
+ else:
+ register_chatml_template()
+
+ if cfg.chat_template == "llama3" and cfg.default_system_message:
+ LOG.info(
+ f"LLaMA-3 set. Adding default system message: {cfg.default_system_message}"
+ )
+ register_llama3_template(cfg.default_system_message)
+ else:
+ register_llama3_template()
+
+ if cfg.rl: # and cfg.rl != "orpo":
+ dataset_meta = load_rl_datasets(cfg=cfg, cli_args=cli_args)
+ else:
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ return train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+
+
+if __name__ == "__main__":
+ fire.Fire(do_cli)
diff --git a/src/axolotl/common/__init__.py b/src/axolotl/common/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/axolotl/common/__pycache__/__init__.cpython-310.pyc b/src/axolotl/common/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..e9f20a2fed89a49310f992c69abf5954d7697d72
Binary files /dev/null and b/src/axolotl/common/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/common/__pycache__/cli.cpython-310.pyc b/src/axolotl/common/__pycache__/cli.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..0f7129cf9708a232608f3c36e8ce604620f0c42e
Binary files /dev/null and b/src/axolotl/common/__pycache__/cli.cpython-310.pyc differ
diff --git a/src/axolotl/common/__pycache__/const.cpython-310.pyc b/src/axolotl/common/__pycache__/const.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..4bf245a865556c194f8b77bea9b2ecd5df1cbe3e
Binary files /dev/null and b/src/axolotl/common/__pycache__/const.cpython-310.pyc differ
diff --git a/src/axolotl/common/cli.py b/src/axolotl/common/cli.py
new file mode 100644
index 0000000000000000000000000000000000000000..636a23ba522f79f739927d0cef91f4843dacd526
--- /dev/null
+++ b/src/axolotl/common/cli.py
@@ -0,0 +1,55 @@
+"""
+shared module for cli specific things
+"""
+
+import logging
+from dataclasses import dataclass, field
+from typing import Optional
+
+import axolotl.monkeypatch.data.batch_dataset_fetcher # pylint: disable=unused-import # noqa: F401
+from axolotl.logging_config import configure_logging
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.models import load_model, load_tokenizer
+
+configure_logging()
+LOG = logging.getLogger("axolotl.common.cli")
+
+
+@dataclass
+class TrainerCliArgs:
+ """
+ dataclass representing the various non-training arguments
+ """
+
+ debug: bool = field(default=False)
+ debug_text_only: bool = field(default=False)
+ debug_num_examples: int = field(default=5)
+ inference: bool = field(default=False)
+ merge_lora: bool = field(default=False)
+ prompter: Optional[str] = field(default=None)
+ shard: bool = field(default=False)
+
+
+@dataclass
+class PreprocessCliArgs:
+ """
+ dataclass representing arguments for preprocessing only
+ """
+
+ debug: bool = field(default=False)
+ debug_text_only: bool = field(default=False)
+ debug_num_examples: int = field(default=1)
+ prompter: Optional[str] = field(default=None)
+
+
+def load_model_and_tokenizer(
+ *,
+ cfg: DictDefault,
+ cli_args: TrainerCliArgs,
+):
+ LOG.info(f"loading tokenizer... {cfg.tokenizer_config or cfg.base_model_config}")
+ tokenizer = load_tokenizer(cfg)
+ LOG.info("loading model and (optionally) peft_config...")
+ model, _ = load_model(cfg, tokenizer, inference=cli_args.inference)
+
+ return model, tokenizer
diff --git a/src/axolotl/common/const.py b/src/axolotl/common/const.py
new file mode 100644
index 0000000000000000000000000000000000000000..fd34ad4694f082fd8a1a0f135fba162e977b868c
--- /dev/null
+++ b/src/axolotl/common/const.py
@@ -0,0 +1,5 @@
+"""
+Various shared constants
+"""
+
+DEFAULT_DATASET_PREPARED_PATH = "last_run_prepared"
diff --git a/src/axolotl/convert.py b/src/axolotl/convert.py
new file mode 100644
index 0000000000000000000000000000000000000000..357e0ec50e1a055f28f1a26e39ca3f2c961a6954
--- /dev/null
+++ b/src/axolotl/convert.py
@@ -0,0 +1,78 @@
+"""Module containing File Reader, File Writer, Json Parser, and Jsonl Serializer classes"""
+
+
+import json
+import sys
+
+
+class FileReader:
+ """
+ Reads a file and returns its contents as a string
+ """
+
+ def read(self, file_path):
+ with open(file_path, encoding="utf-8") as file:
+ return file.read()
+
+
+class FileWriter:
+ """
+ Writes a string to a file
+ """
+
+ def __init__(self, file_path):
+ self.file_path = file_path
+
+ def write(self, content):
+ with open(self.file_path, "w", encoding="utf-8") as file:
+ file.write(content)
+
+
+class StdoutWriter:
+ """
+ Writes a string to stdout
+ """
+
+ def write(self, content):
+ sys.stdout.write(content)
+ sys.stdout.write("\n")
+
+
+class JsonParser:
+ """
+ Parses a string as JSON and returns the result
+ """
+
+ def parse(self, content):
+ return json.loads(content)
+
+
+class JsonlSerializer:
+ """
+ Serializes a list of JSON objects into a JSONL string
+ """
+
+ def serialize(self, data):
+ lines = [json.dumps(item) for item in data]
+ return "\n".join(lines)
+
+
+class JsonToJsonlConverter:
+ """
+ Converts a JSON file to JSONL
+ """
+
+ def __init__(self, file_reader, file_writer, json_parser, jsonl_serializer):
+ self.file_reader = file_reader
+ self.file_writer = file_writer
+ self.json_parser = json_parser
+ self.jsonl_serializer = jsonl_serializer
+
+ def convert(
+ self, input_file_path, output_file_path
+ ): # pylint: disable=unused-argument
+ content = self.file_reader.read(input_file_path)
+ data = self.json_parser.parse(content)
+ # data = [r for r in data if r["conversations"]] # vicuna cleaned has rows with empty conversations
+ jsonl_content = self.jsonl_serializer.serialize(data)
+ self.file_writer.write(jsonl_content)
diff --git a/src/axolotl/core/__init__.py b/src/axolotl/core/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/axolotl/core/__pycache__/__init__.cpython-310.pyc b/src/axolotl/core/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..83f6d9d3cf0278e7c0c3c7e26156349fe79c2a47
Binary files /dev/null and b/src/axolotl/core/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/core/__pycache__/trainer_builder.cpython-310.pyc b/src/axolotl/core/__pycache__/trainer_builder.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..e013d1d3cdc7eee0aead1f29c6e99c7ea4b6b9f7
Binary files /dev/null and b/src/axolotl/core/__pycache__/trainer_builder.cpython-310.pyc differ
diff --git a/src/axolotl/core/trainer_builder.py b/src/axolotl/core/trainer_builder.py
new file mode 100644
index 0000000000000000000000000000000000000000..2f38b12dc1f56dd12d6627718137f9da441b05c2
--- /dev/null
+++ b/src/axolotl/core/trainer_builder.py
@@ -0,0 +1,1616 @@
+# pylint: disable=too-many-lines
+"""
+Builder for the training args and trainer
+"""
+
+import abc
+import importlib
+import importlib.util
+import logging
+import math
+import sys
+from abc import abstractmethod
+from collections import defaultdict
+from dataclasses import dataclass, field
+from functools import wraps
+from pathlib import Path
+from typing import Dict, List, Literal, Optional, Type, Union
+
+import torch
+import transformers
+from datasets import Dataset
+from torch.optim.lr_scheduler import OneCycleLR
+from torch.utils.data import BatchSampler, DataLoader, RandomSampler, SequentialSampler
+from transformers import (
+ EarlyStoppingCallback,
+ PreTrainedModel,
+ Trainer,
+ TrainerCallback,
+ TrainingArguments,
+)
+from transformers.trainer_utils import seed_worker
+from transformers.utils import is_sagemaker_mp_enabled
+from trl import DPOTrainer, ORPOConfig, ORPOTrainer
+from trl.trainer.utils import pad_to_length
+
+from axolotl.loraplus import create_loraplus_optimizer
+from axolotl.monkeypatch.multipack import SUPPORTED_MULTIPACK_MODEL_TYPES
+from axolotl.monkeypatch.relora import ReLoRACallback, ReLoRAScheduler
+from axolotl.utils import is_mlflow_available
+from axolotl.utils.callbacks import (
+ EvalFirstStepCallback,
+ GPUStatsCallback,
+ LossWatchDogCallback,
+ SaveAxolotlConfigtoWandBCallback,
+ SaveBetterTransformerModelCallback,
+ SaveModelOnTrainEndCallback,
+ bench_eval_callback_factory,
+ causal_lm_bench_eval_callback_factory,
+ log_prediction_callback_factory,
+)
+from axolotl.utils.callbacks.lisa import lisa_callback_factory
+from axolotl.utils.collators import (
+ BatchSamplerDataCollatorForSeq2Seq,
+ DataCollatorForSeq2Seq,
+ MambaDataCollator,
+ V2BatchSamplerDataCollatorForSeq2Seq,
+)
+from axolotl.utils.models import ensure_dtype
+from axolotl.utils.samplers import MultipackBatchSampler, get_dataset_lengths
+from axolotl.utils.schedulers import (
+ get_cosine_schedule_with_min_lr,
+ get_cosine_schedule_with_quadratic_warmup,
+ get_cosine_schedule_with_warmup_decay_constant,
+)
+
+if is_sagemaker_mp_enabled():
+ import smdistributed.modelparallel.torch as smp
+
+try:
+ import torch._dynamo # pylint: disable=ungrouped-imports
+except ImportError:
+ pass
+
+LOG = logging.getLogger("axolotl.core.trainer_builder")
+
+
+def _sanitize_kwargs_for_tagging(tag_names, kwargs=None):
+ if isinstance(tag_names, str):
+ tag_names = [tag_names]
+
+ if kwargs is not None:
+ if "tags" not in kwargs:
+ kwargs["tags"] = tag_names
+ elif "tags" in kwargs and isinstance(kwargs["tags"], list):
+ kwargs["tags"].extend(tag_names)
+ elif "tags" in kwargs and isinstance(kwargs["tags"], str):
+ tag_names.append(kwargs["tags"])
+ kwargs["tags"] = tag_names
+
+ return kwargs
+
+
+@dataclass
+class AxolotlTrainingArguments(TrainingArguments):
+ """
+ Extend the base TrainingArguments for axolotl helpers
+ """
+
+ model_type: Optional[str] = field(
+ default=None, metadata={"help": "HF model configuration model_type."}
+ )
+ lr_quadratic_warmup: bool = field(
+ default=False,
+ metadata={"help": "Use quadratic warmup for cosine scheduling."},
+ )
+ pretraining: bool = field(
+ default=False,
+ metadata={
+ "help": "Indicates to trainer whether we are doing continued pretraining."
+ },
+ )
+ sample_packing: bool = field(
+ default=False,
+ metadata={"help": "Use sample packing for efficient training."},
+ )
+ multipack_real_batches: bool = field(
+ default=False,
+ metadata={"help": "Use real batches for efficient training."},
+ )
+ eval_sample_packing: Optional[bool] = field(
+ default=None,
+ metadata={"help": "Use sample packing for efficient evals."},
+ )
+ sample_packing_efficiency: float = field(
+ default=1.0,
+ metadata={"help": "Sample packing efficiency for calculating batch length."},
+ )
+ max_seq_length: int = field(
+ default=2048,
+ metadata={"help": "The maximum sequence length the model can handle"},
+ )
+ sample_packing_seq_len_multiplier: int = field(
+ default=1,
+ metadata={"help": "the multiplier for the max len for packed sequences"},
+ )
+ relora_steps: Optional[int] = field(
+ default=None,
+ metadata={"help": "how often to reset for ReLoRA"},
+ )
+ relora_warmup_steps: Optional[int] = field(
+ default=None,
+ metadata={"help": "how many warmup steps to take after reset for ReLoRA"},
+ )
+ relora_anneal_steps: Optional[int] = field(
+ default=None,
+ metadata={"help": "how many warmup steps to take after reset for ReLoRA"},
+ )
+ relora_prune_ratio: Optional[float] = field(
+ default=0.9,
+ metadata={"help": "prune ratio for magnitude pruning of the optimizer"},
+ )
+ bench_split: Optional[str] = field(
+ default="eval", metadata={"help": "The benchmark split to run on"}
+ )
+ bench_dataset: Optional[str] = field(
+ default="pharaouk/dharma-1/dharma_1_mini.json",
+ metadata={
+ "help": "Benchmark dataset to use: options are `mmlu-zs`, `mmlu-fs`, or the full path to the dataset file"
+ },
+ )
+ do_bench_eval: Optional[bool] = field(
+ default=False, metadata={"help": "Whether to run the Benchmark evaluation."}
+ )
+ do_causal_lm_eval: Optional[bool] = field(
+ default=False, metadata={"help": "Whether to run the Causal LM evaluation."}
+ )
+ max_bench_samples: Optional[int] = field(
+ default=None,
+ metadata={
+ "help": "If set, only evaluates on `max_bench_samples` of the benchmark dataset."
+ },
+ )
+ bench_source_max_len: int = field(
+ default=2048, metadata={"help": "Maximum source sequence length for bench."}
+ )
+ dataloader_prefetch_factor: Optional[int] = field(
+ default=None,
+ metadata={"help": "prefetch_factor argument to the dataloader"},
+ )
+ cosine_min_lr_ratio: Optional[float] = field(
+ default=None,
+ metadata={"help": "Minimum learning rate is min_lr_ratio * learning_rate"},
+ )
+ cosine_constant_lr_ratio: Optional[float] = field(
+ default=None,
+ metadata={
+ "help": "Starting constant learning rate step is cosine_constant_lr_ratio * max_steps"
+ },
+ )
+ loraplus_lr_ratio: Optional[float] = field(
+ default=None, metadata={"help": "loraplus learning rate ratio lr_B / lr_A."}
+ )
+ loraplus_lr_embedding: Optional[float] = field(
+ default=1e-6,
+ metadata={"help": "loraplus learning rate for lora embedding layers."},
+ )
+ qlora: bool = field(
+ default=False,
+ metadata={"help": "whether this is a qlora training"},
+ )
+ orpo_alpha: Optional[float] = field(
+ default=None,
+ )
+ lisa_n_layers: Optional[int] = field(
+ default=None,
+ metadata={"help": "the number of activate layers in LISA"},
+ )
+ lisa_step_interval: Optional[int] = field(
+ default=None,
+ metadata={"help": "how often to switch layers in LISA"},
+ )
+ lisa_layers_attribute: Optional[str] = field(
+ default=None,
+ metadata={"help": "path under the model to access the layers"},
+ )
+ curriculum_sampling: Optional[bool] = field(
+ default=None,
+ metadata={"help": "whether to use sequential sampling for curriculum learning"},
+ )
+
+
+class AxolotlTrainer(Trainer):
+ """
+ Extend the base Trainer for axolotl helpers
+ """
+
+ args = None # type: AxolotlTrainingArguments
+ tag_names = ["axolotl"]
+
+ def __init__(
+ self,
+ *_args,
+ num_epochs=1,
+ bench_data_collator=None,
+ eval_data_collator=None,
+ **kwargs,
+ ):
+ self.num_epochs = num_epochs
+ self.bench_data_collator = bench_data_collator
+ self.eval_data_collator = eval_data_collator
+ super().__init__(*_args, **kwargs)
+ self.train_data_collator = self.data_collator
+ self._stored_metrics = defaultdict(lambda: defaultdict(list))
+ if self.args.orpo_alpha:
+ self.loss_fct = torch.nn.CrossEntropyLoss(reduction="none")
+
+ def create_optimizer(self):
+ if self.args.loraplus_lr_ratio is None:
+ return super().create_optimizer()
+
+ opt_model = self.model_wrapped if is_sagemaker_mp_enabled() else self.model
+ if self.optimizer is None: # pylint: disable=access-member-before-definition
+ optimizer_cls, optimizer_kwargs = Trainer.get_optimizer_cls_and_kwargs(
+ self.args,
+ opt_model,
+ )
+
+ loraplus_lr_ratio = getattr(self.args, "loraplus_lr_ratio", None)
+ loraplus_lr_embedding = getattr(self.args, "loraplus_lr_embedding", None)
+ self.optimizer = create_loraplus_optimizer( # pylint: disable=attribute-defined-outside-init
+ opt_model,
+ optimizer_cls,
+ optimizer_kwargs,
+ loraplus_lr_ratio,
+ loraplus_lr_embedding,
+ )
+
+ if is_sagemaker_mp_enabled():
+ self.optimizer = smp.DistributedOptimizer( # pylint: disable=attribute-defined-outside-init
+ self.optimizer
+ )
+
+ return self.optimizer
+
+ def create_scheduler(
+ self, num_training_steps: int, optimizer: torch.optim.Optimizer = None
+ ):
+ """
+ Setup the scheduler. The optimizer of the trainer must have been set up either before this method is called or
+ passed as an argument.
+
+ Args:
+ num_training_steps (int): The number of training steps to do.
+ optimizer (torch.optim.Optimizer): The training optimizer
+ """
+ use_cosine_quadratic = (
+ self.args.lr_scheduler_type == "cosine"
+ and self.args.lr_quadratic_warmup is True
+ )
+
+ use_cosine_min_lr = (
+ self.args.lr_scheduler_type == "cosine"
+ and self.args.cosine_min_lr_ratio is not None
+ )
+
+ # fmt: off
+ if self.lr_scheduler is None: # type: ignore # pylint: disable=access-member-before-definition
+ # fmt: on
+ if use_cosine_quadratic:
+ if use_cosine_min_lr:
+ LOG.warning("Both cosine quadratic warmup and min lr detected. Using quadratic warmup.")
+
+ self.lr_scheduler = get_cosine_schedule_with_quadratic_warmup( # pylint: disable=attribute-defined-outside-init
+ optimizer,
+ num_warmup_steps=self.args.get_warmup_steps(num_training_steps),
+ num_training_steps=num_training_steps,
+ )
+ elif self.args.cosine_min_lr_ratio and self.args.cosine_constant_lr_ratio and use_cosine_min_lr:
+ assert 0 <= self.args.cosine_min_lr_ratio <= 1.0, "cosine_min_lr_ratio must be between 0.0 and 1.0"
+ assert 0 <= self.args.cosine_constant_lr_ratio <= 1.0, "cosine_constant_lr_ratio must be between 0.0 and 1.0"
+ self.lr_scheduler = get_cosine_schedule_with_warmup_decay_constant( # pylint: disable=attribute-defined-outside-init
+ optimizer,
+ num_warmup_steps=self.args.get_warmup_steps(num_training_steps),
+ num_training_steps=num_training_steps,
+ min_lr_ratio=self.args.cosine_min_lr_ratio,
+ constant_lr_ratio=self.args.cosine_constant_lr_ratio,
+ )
+ elif self.args.cosine_min_lr_ratio and use_cosine_min_lr:
+ assert 0 <= self.args.cosine_min_lr_ratio <= 1.0, "cosine_min_lr_ratio must be between 0.0 and 1.0"
+ self.lr_scheduler = get_cosine_schedule_with_min_lr( # pylint: disable=attribute-defined-outside-init
+ optimizer,
+ num_warmup_steps=self.args.get_warmup_steps(num_training_steps),
+ num_training_steps=num_training_steps,
+ min_lr_ratio=self.args.cosine_min_lr_ratio,
+ )
+ else:
+ return super().create_scheduler(num_training_steps, optimizer)
+ else:
+ if use_cosine_quadratic:
+ LOG.warning("axolotl's cosine scheduler with quadratic warmup not used (e.g., because of deepspeed).")
+
+ if use_cosine_min_lr:
+ LOG.warning("axolotl's cosine scheduler with min lr not used (e.g., because of deepspeed).")
+
+ return self.lr_scheduler
+
+ def _get_train_sampler(self) -> Optional[torch.utils.data.Sampler]:
+ if self.args.sample_packing and not self.args.pretraining:
+ if self.args.multipack_real_batches:
+ batch_size = self.args.per_device_train_batch_size
+ batch_max_len = self.args.max_seq_length
+ else:
+ batch_size = 1
+ batch_max_len = (
+ self.args.per_device_train_batch_size * self.args.max_seq_length
+ )
+ return MultipackBatchSampler(
+ RandomSampler(self.train_dataset),
+ batch_size=batch_size,
+ drop_last=True,
+ batch_max_len=batch_max_len,
+ lengths=get_dataset_lengths(self.train_dataset),
+ packing_efficiency_estimate=self.args.sample_packing_efficiency,
+ )
+ if self.args.curriculum_sampling:
+ return SequentialSampler(self.train_dataset)
+ return super()._get_train_sampler()
+
+ def _get_eval_sampler(
+ self, eval_dataset: Dataset
+ ) -> Optional[torch.utils.data.Sampler]:
+ if self.args.sample_packing and self.args.eval_sample_packing is not False:
+ if self.args.multipack_real_batches:
+ batch_size = self.args.per_device_eval_batch_size
+ batch_max_len = self.args.max_seq_length
+ else:
+ batch_size = 1
+ batch_max_len = (
+ self.args.per_device_eval_batch_size * self.args.max_seq_length
+ )
+ return MultipackBatchSampler(
+ SequentialSampler(eval_dataset),
+ batch_size=batch_size,
+ drop_last=True,
+ batch_max_len=batch_max_len,
+ lengths=get_dataset_lengths(eval_dataset),
+ packing_efficiency_estimate=self.args.sample_packing_efficiency,
+ )
+ return super()._get_eval_sampler(eval_dataset)
+
+ def get_train_dataloader(self) -> DataLoader:
+ if self.args.sample_packing and not self.args.pretraining:
+ train_dataset = self.train_dataset
+ if "length" in train_dataset.features.keys():
+ train_dataset = train_dataset.remove_columns(["length"])
+ data_collator = self.data_collator
+ dataloader_params = {
+ "batch_size": self._train_batch_size,
+ "collate_fn": data_collator,
+ "num_workers": self.args.dataloader_num_workers,
+ "pin_memory": self.args.dataloader_pin_memory,
+ }
+ if self.args.dataloader_prefetch_factor:
+ dataloader_params[
+ "prefetch_factor"
+ ] = self.args.dataloader_prefetch_factor
+
+ sampler = self._get_train_sampler()
+ if isinstance(sampler, BatchSampler):
+ dataloader_params["batch_sampler"] = sampler
+ del dataloader_params["batch_size"]
+ else:
+ dataloader_params["sampler"] = sampler
+ dataloader_params["drop_last"] = self.args.dataloader_drop_last
+ dataloader_params["worker_init_fn"] = seed_worker
+
+ self.accelerator.even_batches = False
+ return self.accelerator.prepare_data_loader(
+ DataLoader(train_dataset, **dataloader_params)
+ )
+ return super().get_train_dataloader()
+
+ def get_eval_dataloader(self, eval_dataset: Optional[Dataset] = None) -> DataLoader:
+ if self.args.sample_packing and self.args.eval_sample_packing is False:
+ self.data_collator = ( # pylint: disable=attribute-defined-outside-init
+ self.eval_data_collator
+ )
+ dataloader = super().get_eval_dataloader(eval_dataset)
+ self.data_collator = ( # pylint: disable=attribute-defined-outside-init
+ self.train_data_collator
+ )
+ return dataloader
+
+ if self.args.sample_packing and self.args.eval_sample_packing is not False:
+ eval_dataset = (
+ eval_dataset if eval_dataset is not None else self.eval_dataset
+ )
+
+ eval_sampler = self._get_eval_sampler(eval_dataset)
+ eval_dataset = eval_dataset.remove_columns(["length"])
+ data_collator = self.data_collator
+ dataloader_params = {
+ "batch_size": self.args.eval_batch_size,
+ "collate_fn": data_collator,
+ "num_workers": self.args.dataloader_num_workers,
+ "pin_memory": self.args.dataloader_pin_memory,
+ }
+ if self.args.dataloader_prefetch_factor:
+ dataloader_params[
+ "prefetch_factor"
+ ] = self.args.dataloader_prefetch_factor
+
+ if isinstance(eval_sampler, BatchSampler):
+ dataloader_params["batch_sampler"] = eval_sampler
+ del dataloader_params["batch_size"]
+ else:
+ dataloader_params["sampler"] = eval_sampler
+ dataloader_params["drop_last"] = self.args.dataloader_drop_last
+
+ self.accelerator.even_batches = False
+ return self.accelerator.prepare_data_loader(
+ DataLoader(eval_dataset, **dataloader_params)
+ )
+
+ return super().get_eval_dataloader(eval_dataset)
+
+ def _get_bench_sampler(
+ self, bench_dataset: Dataset
+ ) -> Optional[torch.utils.data.Sampler]:
+ if self.args.world_size <= 1:
+ return SequentialSampler(bench_dataset)
+ return None
+
+ def get_bench_dataloader(
+ self,
+ bench_dataset: Dataset,
+ ) -> DataLoader:
+ dataloader_params = {
+ "batch_size": self.args.eval_batch_size,
+ "collate_fn": self.bench_data_collator,
+ "num_workers": self.args.dataloader_num_workers,
+ "pin_memory": self.args.dataloader_pin_memory,
+ }
+ if self.args.dataloader_prefetch_factor:
+ dataloader_params["prefetch_factor"] = self.args.dataloader_prefetch_factor
+
+ if not isinstance(bench_dataset, torch.utils.data.IterableDataset):
+ dataloader_params["sampler"] = self._get_bench_sampler(bench_dataset)
+ dataloader_params["drop_last"] = self.args.dataloader_drop_last
+
+ return DataLoader(bench_dataset, **dataloader_params)
+ # return self.accelerator.prepare(DataLoader(bench_dataset, **dataloader_params))
+
+ def compute_loss(self, model, inputs, return_outputs=False):
+ # use one's weighted cross entropy loss calc
+ # if self.args.sample_packing:
+ # labels = inputs.pop("labels")
+ # outputs = model(**inputs)
+ # loss = trainer_weighted_loss(outputs, labels, shift_labels=True)
+ # return (loss, outputs) if return_outputs else loss
+ if self.args.orpo_alpha:
+ return self.orpo_compute_loss(model, inputs, return_outputs=return_outputs)
+ return super().compute_loss(model, inputs, return_outputs=return_outputs)
+
+ @staticmethod
+ def orpo_concatenate_inputs(inputs, label_pad_token=-100, pad_token=0, device=None):
+ concatenated_batch = {}
+
+ max_length = max(
+ inputs["input_ids"].shape[1], inputs["rejected_input_ids"].shape[1]
+ )
+ # Concatenate positive and negative inputs
+ concatenated_batch["input_ids"] = pad_to_length(
+ inputs["input_ids"], max_length, pad_token
+ )
+ concatenated_batch["rejected_input_ids"] = pad_to_length(
+ inputs["rejected_input_ids"], max_length, pad_token
+ )
+ concatenated_batch["labels"] = pad_to_length(
+ inputs["labels"], max_length, label_pad_token
+ )
+ concatenated_batch["rejected_labels"] = pad_to_length(
+ inputs["rejected_labels"], max_length, label_pad_token
+ )
+ concatenated_batch["attention_mask"] = pad_to_length(
+ inputs["attention_mask"], max_length, 0
+ )
+ concatenated_batch["rejected_attention_mask"] = pad_to_length(
+ inputs["rejected_attention_mask"], max_length, 0
+ )
+ concatenated_batch["prompt_attention_mask"] = pad_to_length(
+ inputs["prompt_attention_mask"], max_length, 0
+ ).to(device=device)
+
+ input_ids = torch.cat(
+ [concatenated_batch["input_ids"], concatenated_batch["rejected_input_ids"]],
+ dim=0,
+ ).to(device=device)
+ attention_mask = torch.cat(
+ [
+ concatenated_batch["attention_mask"],
+ concatenated_batch["rejected_attention_mask"],
+ ],
+ dim=0,
+ ).to(device=device)
+ labels = torch.cat(
+ [concatenated_batch["labels"], concatenated_batch["rejected_labels"]], dim=0
+ ).to(device=device)
+
+ return {
+ "input_ids": input_ids,
+ "labels": labels,
+ "attention_mask": attention_mask,
+ "prompt_attention_mask": concatenated_batch["prompt_attention_mask"],
+ }
+
+ def orpo_compute_custom_loss(self, logits, labels):
+ logits = logits.contiguous()
+ loss = 0.0
+
+ if labels is not None:
+ # move labels to correct device to enable model parallelism
+ labels = labels.to(logits.device)
+ # Shift so that tokens < n predict n
+ shift_logits = logits[..., :-1, :].contiguous()
+ shift_labels = labels[..., 1:].contiguous()
+
+ # Flatten the tokens
+ loss = self.loss_fct(shift_logits.transpose(2, 1), shift_labels).mean(
+ dim=-1
+ )
+
+ return loss
+
+ def orpo_compute_logps(
+ self, prompt_attention_mask, chosen_inputs, chosen_attention_mask, logits
+ ):
+ # Get the shape of chosen_attention_mask[:, :-1]
+ chosen_shape = chosen_attention_mask[:, :-1].shape
+
+ # Calculate the padding size
+ pad_length = chosen_shape[1] - (prompt_attention_mask.shape[1] - 1)
+
+ # Pad prompt_attention_mask with zeros to match the desired shape
+ prompt_attention_mask_padded = torch.nn.functional.pad(
+ prompt_attention_mask[:, 1:], (0, pad_length), mode="constant", value=0
+ )
+
+ # Perform the subtraction operation
+ mask = chosen_attention_mask[:, :-1] > prompt_attention_mask_padded
+
+ per_token_logps = torch.gather(
+ logits[:, :-1, :].log_softmax(-1),
+ dim=2,
+ index=(mask * chosen_inputs[:, 1:]).unsqueeze(2),
+ ).squeeze(2)
+ return torch.mul(per_token_logps, mask).sum(dim=1) / mask.sum(dim=1)
+
+ def orpo_compute_loss(self, model, inputs, return_outputs=False):
+ concat_inputs = AxolotlTrainer.orpo_concatenate_inputs(
+ inputs,
+ label_pad_token=-100,
+ pad_token=self.tokenizer.pad_token_id,
+ device=self.accelerator.device,
+ )
+
+ # Perform a single forward pass
+ outputs = model(
+ **{
+ "input_ids": concat_inputs["input_ids"],
+ "attention_mask": concat_inputs["attention_mask"],
+ "labels": concat_inputs["labels"],
+ },
+ output_hidden_states=True,
+ )
+
+ # Split the outputs for positive and negative examples
+ outputs_pos, outputs_neg = outputs.logits.chunk(2)
+
+ # Calculate NLL loss
+ pos_loss = self.orpo_compute_custom_loss(
+ logits=outputs_pos, labels=concat_inputs["input_ids"].chunk(2)[0]
+ )
+
+ # Calculate Log Probability
+ pos_prob = self.orpo_compute_logps(
+ prompt_attention_mask=concat_inputs["prompt_attention_mask"],
+ chosen_inputs=concat_inputs["input_ids"].chunk(2)[0],
+ chosen_attention_mask=concat_inputs["attention_mask"].chunk(2)[0],
+ logits=outputs_pos,
+ )
+ neg_prob = self.orpo_compute_logps(
+ prompt_attention_mask=concat_inputs["prompt_attention_mask"],
+ chosen_inputs=concat_inputs["input_ids"].chunk(2)[1],
+ chosen_attention_mask=concat_inputs["attention_mask"].chunk(2)[1],
+ logits=outputs_neg,
+ )
+
+ # Calculate log odds
+ log_odds = (pos_prob - neg_prob) - (
+ torch.log(1 - torch.exp(pos_prob)) - torch.log(1 - torch.exp(neg_prob))
+ )
+ sig_ratio = torch.nn.functional.sigmoid(log_odds)
+ ratio = torch.log(sig_ratio)
+
+ # Calculate the Final Loss
+ loss = torch.mean(pos_loss - self.args.orpo_alpha * ratio).to(
+ dtype=torch.bfloat16
+ )
+
+ metrics = {}
+ metrics["chosen_geometric_mean"] = torch.mean(pos_prob).cpu().item()
+ metrics["rejected_geometric_mean"] = torch.mean(neg_prob).cpu().item()
+ metrics["log_odds_ratio"] = torch.mean(ratio).cpu().item()
+ metrics["log_odds"] = torch.mean(log_odds).cpu().item()
+ self.store_metrics(metrics, train_eval="train")
+
+ return (loss, outputs_pos) if return_outputs else loss
+
+ @wraps(Trainer.push_to_hub)
+ def push_to_hub(self, *args, **kwargs) -> str:
+ """
+ Overwrite the `push_to_hub` method in order to force-add the tags when pushing the
+ model on the Hub. Please refer to `~transformers.Trainer.push_to_hub` for more details.
+ """
+ kwargs = _sanitize_kwargs_for_tagging(tag_names=self.tag_names, kwargs=kwargs)
+
+ return super().push_to_hub(*args, **kwargs)
+
+ @wraps(Trainer.create_accelerator_and_postprocess)
+ def create_accelerator_and_postprocess(self):
+ res = super().create_accelerator_and_postprocess()
+
+ if self.is_fsdp_enabled:
+ if (
+ "limit_all_gathers" in self.args.fsdp_config
+ and self.args.fsdp_config["limit_all_gathers"]
+ ):
+ self.accelerator.state.fsdp_plugin.limit_all_gathers = True
+
+ return res
+
+ def log(self, logs: Dict[str, float]) -> None:
+ """
+ Log `logs` on the various objects watching training, including stored metrics.
+
+ Args:
+ logs (`Dict[str, float]`):
+ The values to log.
+ """
+ # logs either has 'loss' or 'eval_loss'
+ train_eval = "train" if "loss" in logs else "eval"
+ # Add averaged stored metrics to logs
+ for key, metrics in self._stored_metrics[train_eval].items():
+ logs[key] = torch.tensor(metrics).mean().item()
+ del self._stored_metrics[train_eval]
+ return super().log(logs)
+
+ def store_metrics(
+ self, metrics: Dict[str, float], train_eval: Literal["train", "eval"] = "train"
+ ) -> None:
+ for key, value in metrics.items():
+ self._stored_metrics[train_eval][key].append(value)
+
+
+class AxolotlMambaTrainer(AxolotlTrainer):
+ """
+ Mamba specific trainer to handle loss calculation
+ """
+
+ tag_names = ["axolotl", "mamba"]
+
+ def compute_loss(
+ self,
+ model,
+ inputs,
+ return_outputs=False, # pylint: disable=unused-argument
+ ):
+ input_ids = inputs.pop("input_ids")
+ lm_logits = model(input_ids).logits
+
+ labels = input_ids.to(lm_logits.device)
+ shift_logits = lm_logits[:, :-1, :].contiguous()
+ labels = labels[:, 1:].contiguous()
+
+ loss_fct = torch.nn.CrossEntropyLoss()
+ lm_loss = loss_fct(
+ shift_logits.view(-1, shift_logits.size(-1)), labels.view(-1)
+ )
+
+ return lm_loss
+
+
+class OneCycleLRSchedulerTrainer(AxolotlTrainer):
+ """
+ Trainer subclass that uses the OneCycleLR scheduler
+ """
+
+ tag_names = ["axolotl", "onecycle"]
+
+ def __init__(self, *args, **kwargs):
+ super().__init__(*args, **kwargs)
+ self.lr_scheduler = None
+
+ def create_scheduler(
+ self,
+ num_training_steps: int,
+ optimizer: Optional[torch.optim.Optimizer] = None,
+ ):
+ optimizer = self.optimizer if optimizer is None else optimizer
+ num_warmup_steps = self.args.get_warmup_steps(num_training_steps)
+ pct_start = num_warmup_steps / num_training_steps
+
+ self.lr_scheduler = OneCycleLR(
+ optimizer,
+ max_lr=self.args.learning_rate,
+ total_steps=num_training_steps,
+ pct_start=pct_start,
+ div_factor=6,
+ )
+
+ return self.lr_scheduler
+
+
+class ReLoRATrainer(AxolotlTrainer):
+ """
+ Trainer subclass that uses the OneCycleLR scheduler
+ """
+
+ tag_names = ["axolotl", "relora"]
+
+ def __init__(self, *args, **kwargs):
+ super().__init__(*args, **kwargs)
+ self.lr_scheduler = None
+
+ def create_scheduler(
+ self,
+ num_training_steps: int,
+ optimizer: Optional[torch.optim.Optimizer] = None,
+ ):
+ optimizer = self.optimizer if optimizer is None else optimizer
+ lr_scheduler = super().create_scheduler(num_training_steps, optimizer)
+
+ if self.args.relora_steps:
+ warmup_steps = (
+ self.args.relora_warmup_steps if self.args.relora_warmup_steps else 10
+ )
+ anneal_steps = (
+ self.args.relora_anneal_steps if self.args.relora_anneal_steps else 1
+ )
+ self.lr_scheduler = ReLoRAScheduler(
+ optimizer,
+ lr_scheduler,
+ self.args.relora_steps,
+ anneal_steps,
+ warmup_steps,
+ )
+ else:
+ self.lr_scheduler = lr_scheduler
+
+ return self.lr_scheduler
+
+
+class AxolotlDPOTrainer(DPOTrainer):
+ """
+ Extend the base DPOTrainer for axolotl helpers
+ """
+
+ tag_names = ["axolotl", "dpo"]
+
+ @wraps(DPOTrainer.push_to_hub)
+ def push_to_hub(self, *args, **kwargs) -> str:
+ """
+ Overwrite the `push_to_hub` method in order to force-add the tags when pushing the
+ model on the Hub. Please refer to `~transformers.Trainer.push_to_hub` for more details.
+ """
+ kwargs = _sanitize_kwargs_for_tagging(tag_names=self.tag_names, kwargs=kwargs)
+
+ return super().push_to_hub(*args, **kwargs)
+
+ def tokenize_row(
+ self, feature, model: Optional[Union[PreTrainedModel, torch.nn.Module]] = None
+ ) -> Dict:
+ res = super().tokenize_row(feature, model=model)
+ if self.tokenizer.bos_token_id is None and res["prompt_input_ids"][0] is None:
+ for key in res.keys():
+ res[key] = res[key][1:]
+ return res
+
+
+class AxolotlORPOTrainer(ORPOTrainer):
+ """
+ Extend the base ORPOTrainer for axolotl helpers
+ """
+
+ tag_names = ["axolotl", "orpo"]
+
+
+class TrainerBuilderBase(abc.ABC):
+ """
+ Base class for trainer builder
+ """
+
+ _train_dataset = None
+ _eval_dataset = None
+ _model_ref = None
+ _peft_config = None
+
+ def __init__(self, cfg, model, tokenizer):
+ self.cfg = cfg
+ self.model = model
+ self.tokenizer = tokenizer
+
+ # in case the model supports tagging, add the axolotl tag.
+ # This makes sure the tag is correctly pushed even if a user calls
+ # model.push_to_hub instad of trainer.push_to_hub.
+ if hasattr(model, "add_model_tags"):
+ model.add_model_tags(["axolotl"])
+
+ @property
+ def model_ref(self):
+ return self._model_ref
+
+ @model_ref.setter
+ def model_ref(self, model):
+ self._model_ref = model
+
+ @property
+ def train_dataset(self):
+ return self._train_dataset
+
+ @train_dataset.setter
+ def train_dataset(self, dataset):
+ self._train_dataset = dataset
+
+ @property
+ def eval_dataset(self):
+ return self._eval_dataset
+
+ @eval_dataset.setter
+ def eval_dataset(self, dataset):
+ self._eval_dataset = dataset
+
+ @property
+ def peft_config(self):
+ return self._peft_config
+
+ @peft_config.setter
+ def peft_config(self, peft_config):
+ self._peft_config = peft_config
+
+ @abstractmethod
+ def build(self, total_num_steps):
+ pass
+
+ def get_callbacks(self) -> List[TrainerCallback]:
+ callbacks = []
+ if self.cfg.use_wandb:
+ callbacks.append(
+ SaveAxolotlConfigtoWandBCallback(self.cfg.axolotl_config_path)
+ )
+ if self.cfg.use_mlflow and is_mlflow_available():
+ from axolotl.utils.callbacks.mlflow_ import (
+ SaveAxolotlConfigtoMlflowCallback,
+ )
+
+ callbacks.append(
+ SaveAxolotlConfigtoMlflowCallback(self.cfg.axolotl_config_path)
+ )
+
+ return callbacks
+
+ @abstractmethod
+ def get_post_trainer_create_callbacks(self, trainer):
+ """
+ Callbacks added after the trainer is created, usually b/c these need access to the trainer
+ """
+
+ def hook_pre_create_training_args(self, training_arguments_kwargs):
+ # TODO
+ return training_arguments_kwargs
+
+ def hook_post_create_training_args(self, training_arguments):
+ # TODO
+ return training_arguments
+
+ def hook_pre_create_trainer(self, trainer_kwargs, trainer_cls):
+ # TODO
+ return trainer_kwargs, trainer_cls
+
+ def hook_post_create_trainer(self, trainer):
+ # TODO
+ return trainer
+
+
+class HFCausalTrainerBuilder(TrainerBuilderBase):
+ """
+ Build the HuggingFace training args/trainer for Causal models
+ """
+
+ def get_callbacks(self):
+ callbacks = super().get_callbacks()
+ callbacks.append(GPUStatsCallback(self.cfg))
+ callbacks.append(EvalFirstStepCallback())
+
+ if self.cfg.relora_steps:
+ callbacks.append(ReLoRACallback(self.cfg))
+
+ if (
+ hasattr(self.model, "use_bettertransformer")
+ and self.model.use_bettertransformer is True
+ ):
+ callbacks.append(SaveBetterTransformerModelCallback())
+
+ if self.cfg.loss_watchdog_threshold is not None:
+ callbacks.append(LossWatchDogCallback(self.cfg))
+
+ callbacks.append(SaveModelOnTrainEndCallback())
+
+ return callbacks
+
+ def get_post_trainer_create_callbacks(self, trainer):
+ callbacks = []
+ if self.cfg.use_wandb and self.cfg.eval_table_size > 0:
+ LogPredictionCallback = log_prediction_callback_factory(
+ trainer, self.tokenizer, "wandb"
+ )
+ callbacks.append(LogPredictionCallback(self.cfg))
+ if (
+ self.cfg.use_mlflow
+ and is_mlflow_available()
+ and self.cfg.eval_table_size > 0
+ ):
+ LogPredictionCallback = log_prediction_callback_factory(
+ trainer, self.tokenizer, "mlflow"
+ )
+ callbacks.append(LogPredictionCallback(self.cfg))
+
+ if self.cfg.do_bench_eval:
+ callbacks.append(bench_eval_callback_factory(trainer, self.tokenizer))
+ if self.cfg.do_causal_lm_eval:
+ CausalLMBenchEvalCallback = causal_lm_bench_eval_callback_factory(
+ trainer, self.tokenizer
+ )
+ callbacks.append(CausalLMBenchEvalCallback(self.cfg))
+
+ if self.cfg.early_stopping_patience:
+ early_stop_cb = EarlyStoppingCallback(
+ self.cfg.early_stopping_patience,
+ )
+ callbacks.append(early_stop_cb)
+
+ if self.cfg.lisa_step_interval and self.cfg.lisa_n_layers:
+ callbacks.append(lisa_callback_factory(trainer))
+ return callbacks
+
+ def _get_trainer_cls(self):
+ if self.cfg.lr_scheduler == "one_cycle" and (
+ self.cfg.fsdp or self.cfg.adapter == "qlora"
+ ):
+ return OneCycleLRSchedulerTrainer
+ if self.cfg.relora_steps:
+ return ReLoRATrainer
+ if self.cfg.model_config_type == "mamba":
+ return AxolotlMambaTrainer
+ return AxolotlTrainer
+
+ def build(self, total_num_steps):
+ warmup_steps = None
+ if self.cfg.warmup_steps is not None:
+ warmup_steps = self.cfg.warmup_steps
+ elif self.cfg.warmup_ratio is not None:
+ warmup_steps = max(int(self.cfg.warmup_ratio * total_num_steps), 0)
+ else:
+ warmup_steps = min(int(0.03 * total_num_steps), 100)
+
+ logging_steps = (
+ self.cfg.logging_steps
+ if self.cfg.logging_steps is not None
+ else max(min(int(0.005 * total_num_steps), 10), 1)
+ )
+
+ training_arguments_kwargs = {}
+ if self.cfg.bf16 == "full":
+ training_arguments_kwargs["bf16_full_eval"] = True
+ else:
+ training_arguments_kwargs["bf16"] = self.cfg.bf16
+ training_arguments_kwargs["fp16"] = (
+ self.cfg.fp16 and not self.cfg.bf16
+ ) or False
+ training_arguments_kwargs["tf32"] = self.cfg.tf32
+ training_arguments_kwargs["warmup_steps"] = warmup_steps
+ training_arguments_kwargs["logging_steps"] = logging_steps
+
+ if self.cfg.seed:
+ training_arguments_kwargs["seed"] = self.cfg.seed
+
+ if self.cfg.gradient_checkpointing:
+ training_arguments_kwargs[
+ "gradient_checkpointing"
+ ] = self.cfg.gradient_checkpointing
+ if self.cfg.gradient_checkpointing_kwargs is not None:
+ training_arguments_kwargs[
+ "gradient_checkpointing_kwargs"
+ ] = self.cfg.gradient_checkpointing_kwargs
+ if self.cfg.fsdp:
+ training_arguments_kwargs["fsdp"] = self.cfg.fsdp
+ if self.cfg.fsdp_config:
+ training_arguments_kwargs["fsdp_config"] = dict(self.cfg.fsdp_config)
+
+ if self.cfg.adapter == "qlora":
+ training_arguments_kwargs["qlora"] = True
+
+ # deepspeed
+ if self.cfg.deepspeed:
+ training_arguments_kwargs["deepspeed"] = self.cfg.deepspeed
+
+ if self.cfg.lr_quadratic_warmup is not None:
+ training_arguments_kwargs[
+ "lr_quadratic_warmup"
+ ] = self.cfg.lr_quadratic_warmup
+
+ if self.cfg.adam_beta1:
+ training_arguments_kwargs["adam_beta1"] = self.cfg.adam_beta1
+ if self.cfg.adam_beta2:
+ training_arguments_kwargs["adam_beta2"] = self.cfg.adam_beta2
+ if self.cfg.adam_epsilon:
+ training_arguments_kwargs["adam_epsilon"] = self.cfg.adam_epsilon
+ if self.cfg.max_grad_norm:
+ training_arguments_kwargs["max_grad_norm"] = self.cfg.max_grad_norm
+
+ if self.cfg.hub_model_id:
+ training_arguments_kwargs["hub_model_id"] = self.cfg.hub_model_id
+ training_arguments_kwargs["push_to_hub"] = True
+ training_arguments_kwargs["hub_private_repo"] = True
+ training_arguments_kwargs["hub_always_push"] = True
+
+ if self.cfg.hub_strategy:
+ training_arguments_kwargs["hub_strategy"] = self.cfg.hub_strategy
+
+ if self.cfg.save_safetensors is not None:
+ training_arguments_kwargs["save_safetensors"] = self.cfg.save_safetensors
+
+ if self.cfg.sample_packing_eff_est:
+ training_arguments_kwargs[
+ "sample_packing_efficiency"
+ ] = self.cfg.sample_packing_eff_est
+
+ if self.cfg.dataloader_pin_memory is not None:
+ training_arguments_kwargs[
+ "dataloader_pin_memory"
+ ] = self.cfg.dataloader_pin_memory
+ if self.cfg.dataloader_num_workers is not None:
+ training_arguments_kwargs[
+ "dataloader_num_workers"
+ ] = self.cfg.dataloader_num_workers
+ if self.cfg.dataloader_prefetch_factor is not None:
+ training_arguments_kwargs[
+ "dataloader_prefetch_factor"
+ ] = self.cfg.dataloader_prefetch_factor
+ if self.cfg.dataloader_drop_last is not None:
+ training_arguments_kwargs[
+ "dataloader_drop_last"
+ ] = self.cfg.dataloader_drop_last
+ elif self.cfg.sample_packing and self.cfg.eval_sample_packing is False:
+ training_arguments_kwargs["dataloader_drop_last"] = True
+
+ if self.cfg.remove_unused_columns is not None:
+ training_arguments_kwargs[
+ "remove_unused_columns"
+ ] = self.cfg.remove_unused_columns
+
+ if not self.cfg.test_datasets and self.cfg.val_set_size == 0:
+ # no eval set, so don't eval
+ training_arguments_kwargs["evaluation_strategy"] = "no"
+ elif self.cfg.eval_steps:
+ training_arguments_kwargs["evaluation_strategy"] = "steps"
+ training_arguments_kwargs["eval_steps"] = self.cfg.eval_steps
+ elif self.cfg.evaluation_strategy:
+ training_arguments_kwargs[
+ "evaluation_strategy"
+ ] = self.cfg.evaluation_strategy
+ else:
+ # we have an eval set, but no steps defined, default to use epoch
+ training_arguments_kwargs["evaluation_strategy"] = "epoch"
+
+ if self.cfg.save_steps:
+ training_arguments_kwargs["save_strategy"] = "steps"
+ training_arguments_kwargs["save_steps"] = self.cfg.save_steps
+ elif self.cfg.save_strategy:
+ training_arguments_kwargs["save_strategy"] = self.cfg.save_strategy
+ else:
+ # default to saving each epoch if not defined
+ training_arguments_kwargs["save_strategy"] = "epoch"
+
+ if self.cfg.do_bench_eval:
+ training_arguments_kwargs["do_bench_eval"] = self.cfg.do_bench_eval
+ if self.cfg.bench_dataset:
+ training_arguments_kwargs["bench_dataset"] = self.cfg.bench_dataset
+ if self.cfg.do_causal_lm_eval:
+ training_arguments_kwargs["do_causal_lm_eval"] = self.cfg.do_causal_lm_eval
+ if self.cfg.metric_for_best_model:
+ training_arguments_kwargs[
+ "metric_for_best_model"
+ ] = self.cfg.metric_for_best_model
+ if self.cfg.greater_is_better:
+ training_arguments_kwargs["greater_is_better"] = self.cfg.greater_is_better
+
+ if self.cfg.torch_compile:
+ if torch.__version__ < "2.1.0": # pylint: disable=protected-access
+ LOG.warning("torch>=2.1.0 required for torch_compile to work properly")
+ elif torch._dynamo: # pylint: disable=protected-access
+ torch._dynamo.config.suppress_errors = ( # pylint: disable=protected-access
+ True
+ )
+ training_arguments_kwargs["torch_compile"] = self.cfg.torch_compile
+ if self.cfg.torch_compile_backend:
+ training_arguments_kwargs[
+ "torch_compile_backend"
+ ] = self.cfg.torch_compile_backend
+
+ # DDP Config
+ if self.cfg.ddp_timeout:
+ training_arguments_kwargs["ddp_timeout"] = self.cfg.ddp_timeout
+ # see https://pytorch.org/docs/stable/generated/torch.nn.parallel.DistributedDataParallel.html
+ if self.cfg.ddp_bucket_cap_mb:
+ training_arguments_kwargs["ddp_bucket_cap_mb"] = self.cfg.ddp_bucket_cap_mb
+ if self.cfg.ddp_broadcast_buffers is not None:
+ training_arguments_kwargs[
+ "ddp_broadcast_buffers"
+ ] = self.cfg.ddp_broadcast_buffers
+
+ # these are all the "standard" kwargs that are def used
+ training_arguments_kwargs["max_steps"] = (
+ total_num_steps if self.cfg.max_steps else -1
+ )
+ training_arguments_kwargs["max_seq_length"] = self.cfg.sequence_len
+ training_arguments_kwargs[
+ "per_device_train_batch_size"
+ ] = self.cfg.micro_batch_size
+ if self.cfg.eval_batch_size:
+ training_arguments_kwargs[
+ "per_device_eval_batch_size"
+ ] = self.cfg.eval_batch_size
+ training_arguments_kwargs[
+ "gradient_accumulation_steps"
+ ] = self.cfg.gradient_accumulation_steps
+ training_arguments_kwargs[
+ "eval_accumulation_steps"
+ ] = self.cfg.gradient_accumulation_steps
+ training_arguments_kwargs["num_train_epochs"] = self.cfg.num_epochs
+ training_arguments_kwargs["learning_rate"] = self.cfg.learning_rate
+ training_arguments_kwargs["output_dir"] = self.cfg.output_dir
+ training_arguments_kwargs["save_total_limit"] = (
+ self.cfg.save_total_limit if self.cfg.save_total_limit else 4
+ )
+ training_arguments_kwargs["load_best_model_at_end"] = (
+ (
+ self.cfg.load_best_model_at_end is not False
+ or self.cfg.early_stopping_patience
+ )
+ and (
+ (not self.cfg.test_datasets and self.cfg.val_set_size > 0)
+ or (self.cfg.test_datasets and self.cfg.val_set_size == 0)
+ )
+ and self.cfg.save_steps
+ and self.cfg.eval_steps
+ and self.cfg.save_steps % self.cfg.eval_steps == 0
+ ) or False
+ training_arguments_kwargs["ddp_find_unused_parameters"] = (
+ False if self.cfg.ddp else None
+ )
+ training_arguments_kwargs["group_by_length"] = self.cfg.group_by_length
+ training_arguments_kwargs["curriculum_sampling"] = self.cfg.curriculum_sampling
+ report_to = None
+ if self.cfg.use_wandb:
+ report_to = "wandb"
+ if self.cfg.use_mlflow:
+ report_to = "mlflow"
+ training_arguments_kwargs["report_to"] = report_to
+ training_arguments_kwargs["run_name"] = (
+ self.cfg.wandb_name if self.cfg.use_wandb else None
+ )
+ training_arguments_kwargs["optim"] = (
+ self.cfg.optimizer if self.cfg.optimizer else "adamw_hf"
+ )
+ if self.cfg.optim_args:
+ if isinstance(self.cfg.optim_args, dict):
+ optim_args = ",".join(
+ [f"{key}={value}" for key, value in self.cfg.optim_args.items()]
+ )
+ else:
+ optim_args = self.cfg.optim_args
+ training_arguments_kwargs["optim_args"] = optim_args
+ if self.cfg.optim_target_modules:
+ training_arguments_kwargs[
+ "optim_target_modules"
+ ] = self.cfg.optim_target_modules
+ training_arguments_kwargs["loraplus_lr_ratio"] = self.cfg.loraplus_lr_ratio
+ training_arguments_kwargs[
+ "loraplus_lr_embedding"
+ ] = self.cfg.loraplus_lr_embedding
+ training_arguments_kwargs["lr_scheduler_type"] = (
+ self.cfg.lr_scheduler
+ if self.cfg.lr_scheduler
+ and self.cfg.lr_scheduler not in ("one_cycle", "log_sweep")
+ else "cosine"
+ )
+ training_arguments_kwargs["lr_scheduler_kwargs"] = (
+ self.cfg.lr_scheduler_kwargs if self.cfg.lr_scheduler_kwargs else {}
+ )
+ training_arguments_kwargs["cosine_min_lr_ratio"] = self.cfg.cosine_min_lr_ratio
+ training_arguments_kwargs[
+ "cosine_constant_lr_ratio"
+ ] = self.cfg.cosine_constant_lr_ratio
+ training_arguments_kwargs["weight_decay"] = (
+ self.cfg.weight_decay if self.cfg.weight_decay is not None else 0.0
+ )
+ training_arguments_kwargs["sample_packing"] = (
+ self.cfg.sample_packing if self.cfg.sample_packing else False
+ )
+ training_arguments_kwargs["multipack_real_batches"] = (
+ self.cfg.flash_attention is not True
+ )
+ training_arguments_kwargs["eval_sample_packing"] = (
+ self.cfg.sample_packing
+ if self.cfg.eval_sample_packing is not False
+ else False
+ )
+ training_arguments_kwargs[
+ "sample_packing_seq_len_multiplier"
+ ] = self.cfg.micro_batch_size
+ if self.cfg.relora_steps:
+ training_arguments_kwargs["relora_steps"] = self.cfg.relora_steps
+ training_arguments_kwargs[
+ "relora_warmup_steps"
+ ] = self.cfg.relora_warmup_steps
+ if self.cfg.relora_anneal_steps:
+ training_arguments_kwargs[
+ "relora_anneal_steps"
+ ] = self.cfg.relora_anneal_steps
+ if self.cfg.relora_prune_ratio:
+ training_arguments_kwargs[
+ "relora_prune_ratio"
+ ] = self.cfg.relora_prune_ratio
+
+ if self.cfg.lisa_step_interval and self.cfg.lisa_n_layers:
+ training_arguments_kwargs["lisa_n_layers"] = self.cfg.lisa_n_layers
+ training_arguments_kwargs[
+ "lisa_step_interval"
+ ] = self.cfg.lisa_step_interval
+ training_arguments_kwargs[
+ "lisa_layers_attribute"
+ ] = self.cfg.lisa_layers_attribute
+
+ training_arguments_kwargs = self.hook_pre_create_training_args(
+ training_arguments_kwargs
+ )
+ training_arguments_kwargs["model_type"] = self.cfg.model_config_type
+ training_arguments_kwargs["pretraining"] = bool(self.cfg.pretraining_dataset)
+
+ if self.cfg.rl == "orpo":
+ training_arguments_kwargs["orpo_alpha"] = self.cfg.orpo_alpha
+
+ if self.cfg.neftune_noise_alpha is not None:
+ training_arguments_kwargs[
+ "neftune_noise_alpha"
+ ] = self.cfg.neftune_noise_alpha
+
+ trainer_kwargs = {}
+
+ if self.cfg.optimizer == "lion_pytorch":
+ from lion_pytorch import Lion
+
+ lion_kwargs = {"lr": training_arguments_kwargs["learning_rate"]}
+ if "weight_decay" in training_arguments_kwargs:
+ lion_kwargs["weight_decay"] = training_arguments_kwargs["weight_decay"]
+
+ if (
+ "adam_beta1" in training_arguments_kwargs
+ and "adam_beta2" in training_arguments_kwargs
+ ):
+ lion_kwargs["betas"] = (
+ training_arguments_kwargs["adam_beta1"],
+ training_arguments_kwargs["adam_beta2"],
+ )
+
+ trainer_kwargs["optimizers"] = (
+ Lion(params=self.model.parameters(), **lion_kwargs),
+ None,
+ )
+ # Set default so transformers doesn't throw
+ training_arguments_kwargs["optim"] = "adamw_hf"
+
+ if self.cfg.optimizer == "adamw_anyprecision":
+ if Path(self.cfg.torchdistx_path).exists():
+ sys.path.append(self.cfg.torchdistx_path)
+ importlib.import_module("torchdistx")
+
+ training_args = (
+ AxolotlTrainingArguments( # pylint: disable=unexpected-keyword-arg
+ **training_arguments_kwargs,
+ )
+ )
+ training_args = self.hook_post_create_training_args(training_args)
+
+ data_collator_kwargs = {
+ "padding": True, # True/"longest" is the default
+ }
+ if self.cfg.pad_to_sequence_len:
+ data_collator_kwargs["pad_to_multiple_of"] = 64 * math.ceil(
+ self.cfg.sequence_len / 64
+ )
+ else:
+ # A100 is best at 64, while others at 8. Let's use the larger so we don't have to check
+ # https://docs.nvidia.com/deeplearning/performance/dl-performance-matrix-multiplication/index.html
+ data_collator_kwargs["pad_to_multiple_of"] = 64
+
+ trainer_cls = self._get_trainer_cls()
+ trainer_kwargs, trainer_cls = self.hook_pre_create_trainer(
+ trainer_kwargs, trainer_cls
+ )
+ trainer = trainer_cls(
+ model=self.model,
+ train_dataset=self.train_dataset,
+ eval_dataset=self.eval_dataset,
+ args=training_args,
+ tokenizer=self.tokenizer,
+ data_collator=self.build_collator(training_args, **data_collator_kwargs),
+ eval_data_collator=self.build_collator(
+ training_args, is_eval=True, **data_collator_kwargs
+ ),
+ bench_data_collator=transformers.DataCollatorForSeq2Seq(
+ self.tokenizer,
+ return_tensors="pt",
+ **data_collator_kwargs,
+ ),
+ callbacks=self.get_callbacks(),
+ num_epochs=self.cfg.num_epochs,
+ **trainer_kwargs,
+ )
+ trainer = self.hook_post_create_trainer(trainer)
+ for callback in self.get_post_trainer_create_callbacks(trainer):
+ trainer.add_callback(callback)
+
+ if self.cfg.deepspeed and self.cfg.sample_packing:
+ trainer.accelerator.state.deepspeed_plugin.deepspeed_config[
+ "train_micro_batch_size_per_gpu"
+ ] = self.cfg.micro_batch_size
+
+ return trainer
+
+ def build_collator(
+ self, training_args: AxolotlTrainingArguments, is_eval=False, **kwargs
+ ):
+ if training_args.pretraining:
+ return None
+
+ if self.cfg.model_config_type == "mamba":
+ return MambaDataCollator(tokenizer=self.tokenizer)
+
+ use_batch_sampler_collator = False
+ if is_eval is False and training_args.sample_packing:
+ use_batch_sampler_collator = True
+ if is_eval and training_args.eval_sample_packing:
+ use_batch_sampler_collator = True
+
+ collator: Type[
+ Union[
+ V2BatchSamplerDataCollatorForSeq2Seq,
+ BatchSamplerDataCollatorForSeq2Seq,
+ DataCollatorForSeq2Seq,
+ ]
+ ]
+ if use_batch_sampler_collator:
+ if self.cfg.model_config_type in SUPPORTED_MULTIPACK_MODEL_TYPES:
+ collator = V2BatchSamplerDataCollatorForSeq2Seq
+ elif (
+ self.cfg.model_config_type in ["llama"]
+ and self.cfg.flash_attention is not True
+ ):
+ collator = V2BatchSamplerDataCollatorForSeq2Seq
+ else:
+ collator = BatchSamplerDataCollatorForSeq2Seq
+ else:
+ collator = DataCollatorForSeq2Seq
+
+ return collator(
+ self.tokenizer,
+ return_tensors="pt",
+ **kwargs,
+ )
+
+
+class HFRLTrainerBuilder(TrainerBuilderBase):
+ """
+ Trainer factory class for DPO Trainer
+ """
+
+ def get_callbacks(self):
+ callbacks = super().get_callbacks()
+ callbacks.append(SaveModelOnTrainEndCallback())
+
+ return callbacks
+
+ def get_post_trainer_create_callbacks(self, trainer):
+ callbacks = []
+ return callbacks
+
+ def build_training_arguments(self, total_num_steps):
+ training_args_kwargs = {}
+ for arg in [
+ "adam_beta1",
+ "adam_beta2",
+ "adam_epsilon",
+ "dataloader_num_workers",
+ "dataloader_pin_memory",
+ ]:
+ if hasattr(self.cfg, arg) and getattr(self.cfg, arg) is not None:
+ training_args_kwargs[arg] = getattr(self.cfg, arg)
+
+ if self.cfg.hub_model_id:
+ training_args_kwargs["hub_model_id"] = self.cfg.hub_model_id
+ training_args_kwargs["push_to_hub"] = True
+ training_args_kwargs["hub_private_repo"] = True
+ training_args_kwargs["hub_always_push"] = True
+
+ if self.cfg.hub_strategy:
+ training_args_kwargs["hub_strategy"] = self.cfg.hub_strategy
+
+ if self.cfg.save_safetensors is not None:
+ training_args_kwargs["save_safetensors"] = self.cfg.save_safetensors
+
+ if self.eval_dataset:
+ training_args_kwargs["evaluation_strategy"] = "steps"
+ training_args_kwargs["eval_steps"] = self.cfg.eval_steps
+ else:
+ training_args_kwargs["evaluation_strategy"] = "no"
+
+ if self.cfg.bf16 or self.cfg.bfloat16:
+ training_args_kwargs["bf16"] = True
+
+ training_args_kwargs["lr_scheduler_type"] = (
+ self.cfg.lr_scheduler if self.cfg.lr_scheduler else "cosine"
+ )
+ training_args_kwargs["lr_scheduler_kwargs"] = (
+ self.cfg.lr_scheduler_kwargs if self.cfg.lr_scheduler_kwargs else {}
+ )
+ if self.cfg.remove_unused_columns is not None:
+ training_args_kwargs[
+ "remove_unused_columns"
+ ] = self.cfg.remove_unused_columns
+ else:
+ training_args_kwargs["remove_unused_columns"] = False
+
+ if self.cfg.dataloader_pin_memory is not None:
+ training_args_kwargs[
+ "dataloader_pin_memory"
+ ] = self.cfg.dataloader_pin_memory
+ if self.cfg.dataloader_num_workers is not None:
+ training_args_kwargs[
+ "dataloader_num_workers"
+ ] = self.cfg.dataloader_num_workers
+ if self.cfg.dataloader_prefetch_factor is not None:
+ training_args_kwargs[
+ "dataloader_prefetch_factor"
+ ] = self.cfg.dataloader_prefetch_factor
+ if self.cfg.gradient_checkpointing:
+ training_args_kwargs[
+ "gradient_checkpointing"
+ ] = self.cfg.gradient_checkpointing
+ if self.cfg.gradient_checkpointing_kwargs is not None:
+ training_args_kwargs[
+ "gradient_checkpointing_kwargs"
+ ] = self.cfg.gradient_checkpointing_kwargs
+ else:
+ training_args_kwargs["gradient_checkpointing_kwargs"] = {
+ "use_reentrant": False
+ }
+
+ # set save_strategy and save_steps
+ if self.cfg.save_steps:
+ training_args_kwargs["save_strategy"] = "steps"
+ training_args_kwargs["save_steps"] = self.cfg.save_steps
+ elif self.cfg.save_strategy:
+ training_args_kwargs["save_strategy"] = self.cfg.save_strategy
+ else:
+ # default to saving each epoch if not defined
+ training_args_kwargs["save_strategy"] = "epoch"
+
+ if self.cfg.orpo_alpha:
+ # trl does some odd mapping of alpha to beta to reuse the beta parameter ???
+ training_args_kwargs["beta"] = self.cfg.orpo_alpha
+
+ training_args_cls = TrainingArguments
+ if self.cfg.rl == "orpo":
+ training_args_cls = ORPOConfig
+ training_args_kwargs["dataset_num_proc"] = self.cfg.dataset_processes
+ training_args_kwargs["max_length"] = self.cfg.sequence_len
+ if self.cfg.max_prompt_len:
+ training_args_kwargs["max_prompt_length"] = self.cfg.max_prompt_len
+
+ training_args = training_args_cls(
+ per_device_train_batch_size=self.cfg.micro_batch_size,
+ max_steps=self.cfg.max_steps or total_num_steps,
+ gradient_accumulation_steps=self.cfg.gradient_accumulation_steps,
+ learning_rate=self.cfg.learning_rate,
+ output_dir=self.cfg.output_dir,
+ warmup_steps=self.cfg.warmup_steps,
+ logging_first_step=True,
+ logging_steps=1,
+ optim=self.cfg.optimizer,
+ save_total_limit=self.cfg.save_total_limit or 5,
+ **training_args_kwargs,
+ )
+
+ return training_args
+
+ def build(self, total_num_steps):
+ training_args = self.build_training_arguments(total_num_steps)
+ dpo_trainer_kwargs = {}
+ if self.cfg.rl == "ipo":
+ dpo_trainer_kwargs["loss_type"] = "ipo"
+ if self.cfg.dpo_label_smoothing:
+ dpo_trainer_kwargs["label_smoothing"] = self.cfg.dpo_label_smoothing
+ elif self.cfg.rl == "kto_pair":
+ dpo_trainer_kwargs["loss_type"] = "kto_pair"
+ if self.eval_dataset:
+ dpo_trainer_kwargs["eval_dataset"] = self.eval_dataset
+ if self.cfg.adapter and self.peft_config:
+ dpo_trainer_kwargs["peft_config"] = self.peft_config
+ if self.cfg.precompute_ref_log_probs is not None:
+ dpo_trainer_kwargs[
+ "precompute_ref_log_probs"
+ ] = self.cfg.precompute_ref_log_probs
+ if self.cfg.rl in ["dpo", "ipo", "kto_pair"]:
+ trainer_cls = AxolotlDPOTrainer
+ dpo_trainer_kwargs["beta"] = self.cfg.dpo_beta or 0.1
+ trainer_cls_args = [self.model, self.model_ref]
+
+ # these aren't used for the ORPO trainer
+ dpo_trainer_kwargs["max_length"] = self.cfg.sequence_len
+ dpo_trainer_kwargs["max_target_length"] = None
+ dpo_trainer_kwargs["max_prompt_length"] = self.cfg.sequence_len
+ dpo_trainer_kwargs["generate_during_eval"] = True
+ if self.cfg.rl == "dpo":
+ dpo_trainer_kwargs["dataset_num_proc"] = self.cfg.dataset_processes
+ elif self.cfg.rl == "orpo":
+ trainer_cls = AxolotlORPOTrainer
+ trainer_cls_args = [self.model]
+ else:
+ raise ValueError(f"Unsupported RL: {self.cfg.rl}")
+ dpo_trainer = trainer_cls(
+ *trainer_cls_args,
+ args=training_args,
+ train_dataset=self.train_dataset,
+ tokenizer=self.tokenizer,
+ callbacks=self.get_callbacks(),
+ **dpo_trainer_kwargs,
+ )
+ if self.cfg.fsdp:
+ ensure_dtype(dpo_trainer.model, dtype=self.cfg.torch_dtype)
+
+ dpo_trainer = self.hook_post_create_trainer(dpo_trainer)
+ for callback in self.get_post_trainer_create_callbacks(dpo_trainer):
+ dpo_trainer.add_callback(callback)
+
+ return dpo_trainer
+
+
+class HFPPOTrainerBuilder(TrainerBuilderBase):
+ """
+ HF Factory class for PPO Trainer
+ """
+
+ def get_callbacks(self):
+ callbacks = []
+ return callbacks
+
+ def get_post_trainer_create_callbacks(self, trainer):
+ callbacks = []
+ return callbacks
+
+ def build(self, total_num_steps):
+ # build PPOConfig
+ pass
diff --git a/src/axolotl/core/trainers/__init__.py b/src/axolotl/core/trainers/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/axolotl/core/trainers/trl.py b/src/axolotl/core/trainers/trl.py
new file mode 100644
index 0000000000000000000000000000000000000000..24c0b041232f59e3df49b933bd1361000aa68b72
--- /dev/null
+++ b/src/axolotl/core/trainers/trl.py
@@ -0,0 +1,66 @@
+"""
+module for TRL PPO training
+"""
+import torch
+from tqdm import tqdm
+from trl import PPOTrainer
+
+
+class TRLPPOTrainer(PPOTrainer):
+ """
+ wrapper for ppo trainer to handle customizations
+ """
+
+ def train(
+ self,
+ reward_pipe,
+ resume_from_checkpoint=None, # pylint: disable=unused-argument
+ ):
+ generation_kwargs = {
+ "min_length": -1,
+ "top_k": 0.0,
+ "top_p": 1.0,
+ "do_sample": True,
+ "pad_token_id": self.tokenizer.eos_token_id,
+ "max_new_tokens": 32,
+ }
+ sent_kwargs = {
+ "return_all_scores": True,
+ "function_to_apply": "none",
+ "batch_size": 16,
+ }
+
+ for epoch, batch in tqdm( # pylint: disable=unused-variable
+ enumerate(self.dataloader)
+ ):
+ query_tensors = batch["input_ids"]
+
+ # generate model response
+ response_tensors, ref_response_tensors = self.generate(
+ query_tensors,
+ return_prompt=False,
+ generate_ref_response=True,
+ **generation_kwargs
+ )
+ batch["response"] = self.tokenizer.batch_decode(response_tensors)
+ batch["ref_response"] = self.tokenizer.batch_decode(ref_response_tensors)
+
+ # Compute sentiment score
+ texts = [q + r for q, r in zip(batch["query"], batch["response"])]
+ pipe_outputs = reward_pipe(texts, **sent_kwargs)
+ rewards = [torch.tensor(output[1]["score"]) for output in pipe_outputs]
+ ref_texts = [q + r for q, r in zip(batch["query"], batch["ref_response"])]
+ ref_pipe_outputs = reward_pipe(ref_texts, **sent_kwargs)
+ ref_rewards = [
+ torch.tensor(output[1]["score"]) for output in ref_pipe_outputs
+ ]
+ batch["ref_rewards"] = ref_rewards
+
+ # Run PPO step
+ stats = self.step(query_tensors, response_tensors, rewards)
+ self.log_stats(
+ stats,
+ batch,
+ rewards,
+ columns_to_log=["query", "response", "ref_response", "ref_rewards"],
+ )
diff --git a/src/axolotl/datasets.py b/src/axolotl/datasets.py
new file mode 100644
index 0000000000000000000000000000000000000000..b5638a614d56d44f2749e2ceaf494a6ffda656ac
--- /dev/null
+++ b/src/axolotl/datasets.py
@@ -0,0 +1,188 @@
+"""Module containing Dataset functionality"""
+
+import logging
+import os
+from typing import List, Optional
+
+import torch
+from datasets import Dataset, IterableDataset
+
+from .prompt_tokenizers import PromptTokenizingStrategy
+
+# We want this to be a wrapper for an existing dataset that we have loaded
+# lets use the concept of middlewares to wrap each dataset, for example
+# ConstantLengthDataset(ShuffledDataset([TokenizedPromptDataset(alpaca_dataset)]))
+# let's check to ensure we don't truncate an item in the middle, we'll use
+# the collators later on to pad the datasets
+
+LOG = logging.getLogger("axolotl")
+
+
+class TokenizedPromptDataset(Dataset):
+ """
+ Dataset that returns tokenized prompts from a stream of text files.
+ Args:
+ prompt_tokenizer (PromptTokenizingStrategy): The prompt tokenizing method for processing the data.
+ dataset (dataset.Dataset): Dataset with text files.
+ process_count (int): Number of processes to use for tokenizing.
+ keep_in_memory (bool): Whether to keep the tokenized dataset in memory.
+ """
+
+ def __init__( # pylint: disable=super-init-not-called
+ self,
+ prompt_tokenizer: PromptTokenizingStrategy,
+ dataset: Dataset,
+ process_count: Optional[int] = None,
+ keep_in_memory: Optional[bool] = False,
+ **kwargs,
+ ):
+ self.prompt_tokenizer = prompt_tokenizer
+ self.process_count = process_count
+ self.keep_in_memory = keep_in_memory
+ super().__init__(
+ self.process(dataset).data,
+ **kwargs,
+ )
+
+ def process(self, dataset):
+ features = dataset.features.keys()
+ num_proc = min(64, self.process_count if self.process_count else os.cpu_count())
+
+ map_kwargs = {}
+ if self.prompt_tokenizer.supports_batched:
+ map_kwargs["batched"] = True
+ map_kwargs["batch_size"] = 100
+ return dataset.map(
+ self.prompt_tokenizer.tokenize_prompt,
+ num_proc=num_proc,
+ remove_columns=features,
+ keep_in_memory=self.keep_in_memory,
+ desc="Tokenizing Prompts",
+ **map_kwargs,
+ )
+
+
+# TODO this isn't the best since it can't interleave datasets
+class ConstantLengthDataset(IterableDataset):
+ """
+ Iterable dataset that returns constant length chunks of tokens from stream of text files.
+ Args:
+ tokenizer (Tokenizer): The processor used for processing the data.
+ dataset (dataset.Dataset): Dataset with text files.
+ seq_length (int): Length of token sequences to return.
+ """
+
+ def __init__( # pylint: disable=super-init-not-called
+ self,
+ tokenizer,
+ datasets,
+ seq_length=2048,
+ ):
+ self.tokenizer = tokenizer
+ self.concat_token_id = tokenizer.eos_token_id
+ self.datasets: List[IterableDataset] = datasets
+ self.seq_length = seq_length
+
+ vocab_size = len(tokenizer.get_vocab())
+
+ if vocab_size <= torch.iinfo(torch.int16).max:
+ self.tokens_dtype = torch.int16
+ elif vocab_size <= torch.iinfo(torch.int32).max:
+ self.tokens_dtype = torch.int32
+ else:
+ self.tokens_dtype = torch.int64
+
+ def __iter__(self):
+ buffer = {
+ "input_ids": [],
+ "attention_mask": [],
+ "labels": [],
+ "position_ids": [],
+ }
+ buffer_len = 0
+ for dataset in self.datasets:
+ idx = 0
+ iterator = iter(dataset)
+ more_examples = True
+ while more_examples:
+ try:
+ example = next(iterator)
+ idx += 1
+ except StopIteration:
+ more_examples = False
+ example = None
+
+ add_concat_token = False
+ if example:
+ example_len = len(example["input_ids"])
+ add_concat_token = example["input_ids"][-1] != self.concat_token_id
+ else:
+ example_len = 0
+
+ if not example_len or (
+ buffer_len + int(add_concat_token) + example_len > self.seq_length
+ ):
+ if buffer["input_ids"]:
+ input_ids = torch.cat(buffer["input_ids"], dim=-1)[
+ : self.seq_length
+ ]
+ attention_mask = torch.cat(buffer["attention_mask"], dim=-1)[
+ : self.seq_length
+ ]
+ position_ids = torch.cat(buffer["position_ids"], dim=-1)[
+ : self.seq_length
+ ]
+ labels = torch.cat(buffer["labels"], dim=-1)[: self.seq_length]
+ if labels.size() == input_ids.size() and (
+ attention_mask.size() == input_ids.size()
+ ):
+ yield {
+ "input_ids": input_ids,
+ "labels": labels,
+ "attention_mask": attention_mask,
+ "position_ids": position_ids,
+ }
+ else:
+ LOG.warning(
+ f"dropping batch due to tensor size mismatch input_ids: {input_ids.size()}, labels: {labels.size()}, attention_mask: {attention_mask.size()}"
+ )
+ buffer = {
+ "input_ids": [],
+ "attention_mask": [],
+ "labels": [],
+ "position_ids": [],
+ }
+ buffer_len = 0
+ idx = 1
+
+ if example:
+ # FIXME
+ # just going to drop data points that are too long
+ if len(example["input_ids"]) <= self.seq_length:
+ input_ids = example["input_ids"]
+ attention_mask = example["attention_mask"]
+ labels = example["labels"]
+
+ if add_concat_token:
+ input_ids.append(self.concat_token_id)
+ attention_mask.append(1)
+ labels.append(self.concat_token_id)
+
+ input_ids_with_concat = torch.tensor(
+ input_ids, dtype=self.tokens_dtype
+ )
+ attention_mask_with_concat = torch.tensor(
+ [idx * m for m in attention_mask], dtype=torch.int16
+ )
+ labels_with_concat = torch.tensor(
+ labels, dtype=self.tokens_dtype
+ )
+ position_ids = torch.arange(
+ len(input_ids), dtype=self.tokens_dtype
+ )
+
+ buffer["input_ids"].append(input_ids_with_concat)
+ buffer["attention_mask"].append(attention_mask_with_concat)
+ buffer["labels"].append(labels_with_concat)
+ buffer["position_ids"].append(position_ids)
+ buffer_len += len(input_ids)
diff --git a/src/axolotl/logging_config.py b/src/axolotl/logging_config.py
new file mode 100644
index 0000000000000000000000000000000000000000..2ddf89a8c4b0efe68389e7776b21a3180c9c4191
--- /dev/null
+++ b/src/axolotl/logging_config.py
@@ -0,0 +1,72 @@
+"""
+Common logging module for axolotl
+"""
+
+import os
+import sys
+from logging import Formatter
+from logging.config import dictConfig
+from typing import Any, Dict
+
+from colorama import Fore, Style, init
+
+
+class ColorfulFormatter(Formatter):
+ """
+ Formatter to add coloring to log messages by log type
+ """
+
+ COLORS = {
+ "WARNING": Fore.YELLOW,
+ "ERROR": Fore.RED,
+ "CRITICAL": Fore.RED + Style.BRIGHT,
+ }
+
+ def format(self, record):
+ record.rank = int(os.getenv("LOCAL_RANK", "0"))
+ log_message = super().format(record)
+ return self.COLORS.get(record.levelname, "") + log_message + Fore.RESET
+
+
+DEFAULT_LOGGING_CONFIG: Dict[str, Any] = {
+ "version": 1,
+ "disable_existing_loggers": False,
+ "formatters": {
+ "simple": {
+ "format": "[%(asctime)s] [%(levelname)s] [%(name)s.%(funcName)s:%(lineno)d] [PID:%(process)d] %(message)s",
+ },
+ "colorful": {
+ "()": ColorfulFormatter,
+ "format": "[%(asctime)s] [%(levelname)s] [%(name)s.%(funcName)s:%(lineno)d] [PID:%(process)d] [RANK:%(rank)d] %(message)s",
+ },
+ },
+ "filters": {},
+ "handlers": {
+ "console": {
+ "class": "logging.StreamHandler",
+ "formatter": "simple",
+ "filters": [],
+ "stream": sys.stdout,
+ },
+ "color_console": {
+ "class": "logging.StreamHandler",
+ "formatter": "colorful",
+ "filters": [],
+ "stream": sys.stdout,
+ },
+ },
+ "root": {"handlers": ["console"], "level": os.getenv("LOG_LEVEL", "INFO")},
+ "loggers": {
+ "axolotl": {
+ "handlers": ["color_console"],
+ "level": "DEBUG",
+ "propagate": False,
+ },
+ },
+}
+
+
+def configure_logging():
+ """Configure with default logging"""
+ init() # Initialize colorama
+ dictConfig(DEFAULT_LOGGING_CONFIG)
diff --git a/src/axolotl/loraplus.py b/src/axolotl/loraplus.py
new file mode 100644
index 0000000000000000000000000000000000000000..b4abec55ad4aa4669718f4a418a27fcdb99f6fd3
--- /dev/null
+++ b/src/axolotl/loraplus.py
@@ -0,0 +1,133 @@
+"""Module for LoRA+"""
+
+# MIT License
+#
+# Copyright (c) 2024 nikhil-ghosh-berkeley
+# https://github.com/nikhil-ghosh-berkeley/loraplus
+
+import logging
+from functools import reduce
+
+from peft.tuners import lora
+from torch import nn
+from transformers.pytorch_utils import ALL_LAYERNORM_LAYERS
+from transformers.trainer_pt_utils import get_parameter_names
+
+LOG = logging.getLogger("axolotl.loraplus")
+
+
+def get_module(name, opt_model):
+ """
+ Retrieve a module from a model using its parameter name.
+ Args:
+ name (str): Full name of the parameter, typically including module path.
+ opt_model (torch.nn.Module): The model from which to retrieve the module.
+
+ Returns:
+ Module corresponding to the given name.
+ """
+ parent_idx = 2 if "lora" in name else 1
+ module_names = name.split(sep=".")[:-parent_idx]
+ module = reduce(getattr, module_names, opt_model)
+ return module
+
+
+def create_loraplus_optimizer(
+ opt_model,
+ optimizer_cls,
+ optimizer_kwargs,
+ loraplus_lr_ratio,
+ loraplus_lr_embedding=None,
+):
+ """
+ Creates an optimizer for the given model, applying LoRA-specific learning rate adjustments to different parameter groups.
+
+ Args:
+ opt_model (torch.nn.Module): The model for which the optimizer is being created.
+ optimizer_cls (class): The class of the optimizer to be used (e.g., torch.optim.Adam).
+ optimizer_kwargs (dict): A dictionary of keyword arguments for the optimizer's initialization.
+ loraplus_lr_ratio (float): The learning rate ratio to be applied to LoRA parameters.
+ loraplus_lr_embedding (float, optional): A specific learning rate for embedding parameters, with a default value if not provided.
+
+ Returns:
+ An instance of the specified optimizer class configured with the model's parameters organized into groups with custom learning rates.
+ """
+
+ assert loraplus_lr_ratio is not None, "loraplus_lr_ratio must be provided."
+
+ if loraplus_lr_embedding is None:
+ loraplus_lr_embedding = 1e-6
+
+ decay_parameters = get_parameter_names(opt_model, ALL_LAYERNORM_LAYERS)
+ decay_parameters = [name for name in decay_parameters if "bias" not in name]
+ param_groups = {
+ "groupA": {},
+ "groupB": {},
+ "groupB_no_decay": {},
+ "embedding": {},
+ }
+
+ for name, param in opt_model.named_parameters():
+ if not param.requires_grad:
+ continue
+
+ module = get_module(name, opt_model)
+ if isinstance(module, lora.Embedding):
+ param_groups["embedding"][name] = param
+ elif "lora_B" in name or param.ndim == 1:
+ if name in decay_parameters:
+ param_groups["groupB"][name] = param
+ else:
+ param_groups["groupB_no_decay"][name] = param
+ else:
+ param_groups["groupA"][name] = param
+
+ assigned_param_groups = ""
+ for group, group_params in param_groups.items():
+ assigned_param_groups += f"{group}\n {list(group_params.keys())}\n\n"
+ LOG.info(assigned_param_groups)
+
+ lr = optimizer_kwargs["lr"] # pylint: disable=invalid-name
+ weight_decay = optimizer_kwargs.get("weight_decay", 0.0)
+
+ optimizer_grouped_parameters = [
+ {
+ "params": list(param_groups["groupA"].values()),
+ "weight_decay": weight_decay,
+ "lr": lr,
+ },
+ {
+ "params": list(param_groups["embedding"].values()),
+ "weight_decay": weight_decay,
+ "lr": loraplus_lr_embedding,
+ },
+ {
+ "params": list(param_groups["groupB"].values()),
+ "weight_decay": weight_decay,
+ "lr": lr * loraplus_lr_ratio,
+ },
+ {
+ "params": list(param_groups["groupB_no_decay"].values()),
+ "weight_decay": 0.0,
+ "lr": lr * loraplus_lr_ratio,
+ },
+ ]
+
+ optimizer = optimizer_cls(optimizer_grouped_parameters, **optimizer_kwargs)
+ if optimizer_cls.__name__ == "Adam8bit":
+ import bitsandbytes
+
+ manager = bitsandbytes.optim.GlobalOptimManager.get_instance()
+
+ skipped = 0
+ for module in opt_model.modules():
+ if isinstance(module, nn.Embedding):
+ skipped += sum(
+ {p.data_ptr(): p.numel() for p in module.parameters()}.values()
+ )
+ LOG.info(f"skipped {module}: {skipped/2**20}M params")
+ manager.register_module_override(module, "weight", {"optim_bits": 32})
+ LOG.debug(f"bitsandbytes: will optimize {module} in fp32")
+ LOG.info(f"skipped: {skipped/2**20}M params")
+
+ return optimizer
diff --git a/src/axolotl/models/__init__.py b/src/axolotl/models/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/axolotl/models/__pycache__/__init__.cpython-310.pyc b/src/axolotl/models/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..75113b8ca583ba347eb73f794b08f57b38a744be
Binary files /dev/null and b/src/axolotl/models/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/models/mamba/__init__.py b/src/axolotl/models/mamba/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..fee88e3a434979b72f2c47d7741dbbdf21b1996d
--- /dev/null
+++ b/src/axolotl/models/mamba/__init__.py
@@ -0,0 +1,24 @@
+"""
+Modeling module for Mamba models
+"""
+
+import importlib
+
+
+def check_mamba_ssm_installed():
+ mamba_ssm_spec = importlib.util.find_spec("mamba_ssm")
+ if mamba_ssm_spec is None:
+ raise ImportError(
+ "MambaLMHeadModel requires mamba_ssm. Please install it with `pip install -e .[mamba-ssm]`"
+ )
+
+
+def fix_mamba_attn_for_loss():
+ check_mamba_ssm_installed()
+
+ from mamba_ssm.models import mixer_seq_simple
+
+ from .modeling_mamba import MambaLMHeadModel as MambaLMHeadModelFixed
+
+ mixer_seq_simple.MambaLMHeadModel = MambaLMHeadModelFixed
+ return mixer_seq_simple.MambaLMHeadModel # pylint: disable=invalid-name
diff --git a/src/axolotl/models/mamba/__pycache__/__init__.cpython-310.pyc b/src/axolotl/models/mamba/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..80f8e3d8a354840cab1df9690c40312342b28e99
Binary files /dev/null and b/src/axolotl/models/mamba/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/models/mamba/configuration_mamba.py b/src/axolotl/models/mamba/configuration_mamba.py
new file mode 100644
index 0000000000000000000000000000000000000000..5160ee8d7e06a6d7c2dd3d0d9ae401ba3a288ef8
--- /dev/null
+++ b/src/axolotl/models/mamba/configuration_mamba.py
@@ -0,0 +1,42 @@
+"""
+HF Transformers MambaConfig
+"""
+from transformers import PretrainedConfig
+
+
+class MambaConfig(PretrainedConfig):
+ """
+ modeling configuration for state space model/mamba
+ """
+
+ model_type = "mamba"
+
+ def __init__(
+ self,
+ vocab_size=50280,
+ d_model=2560,
+ n_layer=64,
+ rms_norm=True,
+ residual_in_fp32=True,
+ fused_add_norm=True,
+ pad_vocab_size_multiple=8,
+ pad_token_id=50277,
+ bos_token_id=0,
+ eos_token_id=0,
+ tie_word_embeddings=False,
+ **kwargs,
+ ):
+ self.vocab_size = vocab_size
+ self.d_model = d_model
+ self.n_layer = n_layer
+ self.rms_norm = rms_norm
+ self.residual_in_fp32 = residual_in_fp32
+ self.fused_add_norm = fused_add_norm
+ self.pad_vocab_size_multiple = pad_vocab_size_multiple
+ super().__init__(
+ pad_token_id=pad_token_id,
+ bos_token_id=bos_token_id,
+ eos_token_id=eos_token_id,
+ tie_word_embeddings=tie_word_embeddings,
+ **kwargs,
+ )
diff --git a/src/axolotl/models/mamba/modeling_mamba.py b/src/axolotl/models/mamba/modeling_mamba.py
new file mode 100644
index 0000000000000000000000000000000000000000..70e9c88c882f595b849a63e2c274f6853ccc5ec2
--- /dev/null
+++ b/src/axolotl/models/mamba/modeling_mamba.py
@@ -0,0 +1,128 @@
+# pylint: skip-file
+import os
+from collections import namedtuple
+from functools import partial
+from typing import Optional, Union
+
+import torch
+from mamba_ssm.models.mixer_seq_simple import MixerModel, _init_weights
+from mamba_ssm.utils.generation import GenerationMixin
+from mamba_ssm.utils.hf import load_config_hf, load_state_dict_hf
+from torch import nn
+from torch.nn import CrossEntropyLoss
+
+from axolotl.models.mamba.configuration_mamba import MambaConfig
+
+
+class MambaLMHeadModel(nn.Module, GenerationMixin):
+ def __init__(
+ self,
+ d_model: int,
+ n_layer: int,
+ vocab_size: int,
+ initializer_cfg=None,
+ pad_vocab_size_multiple: int = 1,
+ device=None,
+ dtype=None,
+ **backbone_kwargs,
+ ) -> None:
+ factory_kwargs = {"device": device, "dtype": dtype}
+ super().__init__()
+ if vocab_size % pad_vocab_size_multiple != 0:
+ vocab_size += pad_vocab_size_multiple - (
+ vocab_size % pad_vocab_size_multiple
+ )
+ self.config = MambaConfig(
+ vocab_size=vocab_size,
+ d_model=d_model,
+ n_layer=n_layer,
+ pad_vocab_size_multiple=pad_vocab_size_multiple,
+ )
+ self.backbone = MixerModel(
+ d_model=d_model,
+ n_layer=n_layer,
+ vocab_size=vocab_size,
+ initializer_cfg=initializer_cfg,
+ **backbone_kwargs,
+ **factory_kwargs,
+ )
+ self.lm_head = nn.Linear(d_model, vocab_size, bias=False, **factory_kwargs)
+
+ # Initialize weights and apply final processing
+ self.apply(
+ partial(
+ _init_weights,
+ n_layer=n_layer,
+ **(initializer_cfg if initializer_cfg is not None else {}),
+ )
+ )
+ self.tie_weights()
+
+ def tie_weights(self):
+ self.lm_head.weight = self.backbone.embedding.weight
+
+ def allocate_inference_cache(self, batch_size, max_seqlen, dtype=None, **kwargs):
+ return self.backbone.allocate_inference_cache(
+ batch_size, max_seqlen, dtype=dtype, **kwargs
+ )
+
+ def forward(
+ self,
+ input_ids,
+ position_ids=None,
+ inference_params=None,
+ num_last_tokens=0,
+ labels=None,
+ **kwargs,
+ ):
+ """
+ "position_ids" is just to be compatible with Transformer generation. We don't use it.
+ num_last_tokens: if > 0, only return the logits for the last n tokens
+ """
+ hidden_states = self.backbone(input_ids, inference_params=inference_params)
+ if num_last_tokens > 0:
+ hidden_states = hidden_states[:, -num_last_tokens:]
+ lm_logits = self.lm_head(hidden_states)
+
+ CausalLMOutput = namedtuple("CausalLMOutput", ["logits"])
+ return CausalLMOutput(logits=lm_logits)
+
+ loss = None
+ if labels is not None:
+ logits = lm_logits
+ # Shift so that tokens < n predict n
+ shift_logits = logits[..., :-1, :].contiguous()
+ shift_labels = labels[..., 1:].contiguous()
+ # Flatten the tokens
+ loss_fct = CrossEntropyLoss()
+ shift_logits = shift_logits.view(-1, self.config.vocab_size)
+ shift_labels = shift_labels.view(-1)
+ # Enable model parallelism
+ shift_labels = shift_labels.to(shift_logits.device)
+ loss = loss_fct(shift_logits, shift_labels)
+ CausalLMOutput = namedtuple("CausalLMOutput", ["logits", "loss"])
+ print(loss)
+ return CausalLMOutput(logits=lm_logits, loss=loss)
+
+ else:
+ CausalLMOutput = namedtuple("CausalLMOutput", ["logits"])
+ return CausalLMOutput(logits=lm_logits)
+
+ def save_pretrained(
+ self,
+ save_directory: Union[str, os.PathLike],
+ state_dict: Optional[dict] = None,
+ safe_serialization: Optional[bool] = None, # pylint: disable=unused-argument
+ ):
+ if state_dict is None:
+ state_dict = self.state_dict()
+ torch.save(state_dict, os.path.join(save_directory, "pytorch_model.bin"))
+
+ @classmethod
+ def from_pretrained(cls, pretrained_model_name, device=None, dtype=None, **kwargs):
+ config = load_config_hf(pretrained_model_name)
+ model = cls(**config, device=device, dtype=dtype, **kwargs)
+ model.load_state_dict(
+ load_state_dict_hf(pretrained_model_name, device={"": device}, dtype=dtype)
+ )
+ return model
diff --git a/src/axolotl/monkeypatch/__pycache__/fastchat_conversation_turns.cpython-310.pyc b/src/axolotl/monkeypatch/__pycache__/fastchat_conversation_turns.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..07d790d2d9c3e85b03385b0defe3cdec7a8155ba
Binary files /dev/null and b/src/axolotl/monkeypatch/__pycache__/fastchat_conversation_turns.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/__pycache__/llama_attn_hijack_flash.cpython-310.pyc b/src/axolotl/monkeypatch/__pycache__/llama_attn_hijack_flash.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..6eaa3baede7628e3cfe7697a3d864c3f1c1b3c4b
Binary files /dev/null and b/src/axolotl/monkeypatch/__pycache__/llama_attn_hijack_flash.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/__pycache__/llama_expand_mask.cpython-310.pyc b/src/axolotl/monkeypatch/__pycache__/llama_expand_mask.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..ced5121a5f59a0c78401be69446fa278c8d93f76
Binary files /dev/null and b/src/axolotl/monkeypatch/__pycache__/llama_expand_mask.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/__pycache__/multipack.cpython-310.pyc b/src/axolotl/monkeypatch/__pycache__/multipack.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..03ca1f32fca76c57ebcb6bfca28daceb349d5e86
Binary files /dev/null and b/src/axolotl/monkeypatch/__pycache__/multipack.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/__pycache__/relora.cpython-310.pyc b/src/axolotl/monkeypatch/__pycache__/relora.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..8f4b384195dd69938f9c12fb8983275367ab869e
Binary files /dev/null and b/src/axolotl/monkeypatch/__pycache__/relora.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/__pycache__/unsloth_.cpython-310.pyc b/src/axolotl/monkeypatch/__pycache__/unsloth_.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..48279f6beb7f2a4fd91ee55a5d625820f4321701
Binary files /dev/null and b/src/axolotl/monkeypatch/__pycache__/unsloth_.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/__pycache__/utils.cpython-310.pyc b/src/axolotl/monkeypatch/__pycache__/utils.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..89df65c614e9048222a68f319bdfd18fb9305df1
Binary files /dev/null and b/src/axolotl/monkeypatch/__pycache__/utils.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/btlm_attn_hijack_flash.py b/src/axolotl/monkeypatch/btlm_attn_hijack_flash.py
new file mode 100644
index 0000000000000000000000000000000000000000..1275906804b0f6908ef94be9b94d4a83960bf148
--- /dev/null
+++ b/src/axolotl/monkeypatch/btlm_attn_hijack_flash.py
@@ -0,0 +1,66 @@
+"""
+Flash attention monkey patch for cerebras btlm model
+"""
+
+import importlib
+import logging
+from typing import Optional, Tuple
+
+import torch
+from accelerate import init_empty_weights
+from flash_attn.flash_attn_interface import flash_attn_func
+from transformers import AutoConfig, AutoModelForCausalLM
+
+LOG = logging.getLogger("axolotl")
+
+
+def replace_btlm_attn_with_flash_attn(model_name="cerebras/btlm-3b-8k-base"):
+ # this is a wonky hack to get the remotely loaded module
+ model_config = AutoConfig.from_pretrained(model_name, trust_remote_code=True)
+ # we need to load the model here in order for modeling_btlm to be available
+ with init_empty_weights():
+ AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True)
+ module_name = model_config.__class__.__module__.replace(
+ ".configuration_btlm", ".modeling_btlm"
+ )
+ modeling_btlm = importlib.import_module(module_name)
+ modeling_btlm.BTLMAttention._attn = ( # pylint: disable=protected-access
+ flashattn_attn
+ )
+
+
+def flashattn_attn(
+ self,
+ query: torch.Tensor,
+ key: Optional[torch.Tensor] = None,
+ value: Optional[torch.Tensor] = None,
+ attention_mask: Optional[torch.Tensor] = None, # pylint: disable=unused-argument
+ head_mask: Optional[torch.Tensor] = None,
+ position_bias: Optional[torch.Tensor] = None, # pylint: disable=unused-argument
+) -> Tuple[torch.Tensor, Optional[torch.Tensor]]:
+ softmax_scale = (
+ 1 / (key.size(-1) ** self.attn_scale_power) if self.scale_attn_weights else None
+ )
+
+ query = query.permute(0, 2, 1, 3)
+ key = key.permute(0, 2, 1, 3)
+ value = value.permute(0, 2, 1, 3)
+
+ # Perform Flash attention
+ attn_output = flash_attn_func(
+ query,
+ key,
+ value,
+ dropout_p=0.0, # Assuming you have this attribute
+ softmax_scale=softmax_scale, # Set this if you have specific scaling in mind
+ causal=not self.is_cross_attention, # Assuming you have this attribute
+ return_attn_probs=False, # Set this based on your needs
+ )
+
+ # Optional: Apply head mask if it's not None
+ if head_mask is not None:
+ attn_output *= head_mask
+
+ attn_output = attn_output.permute(0, 2, 1, 3)
+
+ return attn_output, None # We don't have explicit attn_weights in Flash attention
diff --git a/src/axolotl/monkeypatch/data/__init__.py b/src/axolotl/monkeypatch/data/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/axolotl/monkeypatch/data/__pycache__/__init__.cpython-310.pyc b/src/axolotl/monkeypatch/data/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..3b99968c58fc3c3988e8cd245b744081589d11ea
Binary files /dev/null and b/src/axolotl/monkeypatch/data/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/data/__pycache__/batch_dataset_fetcher.cpython-310.pyc b/src/axolotl/monkeypatch/data/__pycache__/batch_dataset_fetcher.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..4b06b95dcf34c2db56ca778d968e69a467c0b02f
Binary files /dev/null and b/src/axolotl/monkeypatch/data/__pycache__/batch_dataset_fetcher.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/data/batch_dataset_fetcher.py b/src/axolotl/monkeypatch/data/batch_dataset_fetcher.py
new file mode 100644
index 0000000000000000000000000000000000000000..2e9364e3a5faa39a6084d259a4cb2d876a9f2e57
--- /dev/null
+++ b/src/axolotl/monkeypatch/data/batch_dataset_fetcher.py
@@ -0,0 +1,46 @@
+"""monkey patches for the dataset fetcher to handle batches of packed indexes"""
+# pylint: disable=protected-access
+
+import torch
+from torch.utils.data._utils.fetch import _BaseDatasetFetcher
+from torch.utils.data._utils.worker import _worker_loop
+
+
+class _MapDatasetFetcher(_BaseDatasetFetcher):
+ def fetch(self, possibly_batched_index):
+ if isinstance(possibly_batched_index[0], list):
+ data = [None for i in possibly_batched_index]
+ for i, possibly_batched_index_ in enumerate(possibly_batched_index):
+ if self.auto_collation:
+ if (
+ hasattr(self.dataset, "__getitems__")
+ and self.dataset.__getitems__
+ ):
+ data[i] = self.dataset.__getitems__(possibly_batched_index_)
+ else:
+ data[i] = [self.dataset[idx] for idx in possibly_batched_index_]
+ else:
+ data[i] = self.dataset[possibly_batched_index_]
+ else:
+ if self.auto_collation:
+ if hasattr(self.dataset, "__getitems__") and self.dataset.__getitems__:
+ data = self.dataset.__getitems__(possibly_batched_index)
+ else:
+ data = [self.dataset[idx] for idx in possibly_batched_index]
+ else:
+ data = self.dataset[possibly_batched_index]
+ return self.collate_fn(data)
+
+
+def patch_fetchers():
+ torch.utils.data._utils.fetch._MapDatasetFetcher = _MapDatasetFetcher
+ torch.utils.data.dataloader._utils.fetch._MapDatasetFetcher = _MapDatasetFetcher
+
+
+def patched_worker_loop(*args, **kwargs):
+ patch_fetchers()
+ return _worker_loop(*args, **kwargs)
+
+
+torch.utils.data._utils.worker._worker_loop = patched_worker_loop
+patch_fetchers()
diff --git a/src/axolotl/monkeypatch/fastchat_conversation_turns.py b/src/axolotl/monkeypatch/fastchat_conversation_turns.py
new file mode 100644
index 0000000000000000000000000000000000000000..a09bfddb4bf5d64f5a66ae96816a7a3df0d661dd
--- /dev/null
+++ b/src/axolotl/monkeypatch/fastchat_conversation_turns.py
@@ -0,0 +1,231 @@
+"""
+monkeypatch to add a get_turns method
+"""
+
+import logging
+from typing import Generator, Tuple
+
+from fastchat.conversation import SeparatorStyle
+
+LOG = logging.getLogger("axolotl.monkeypatch.fastchat_conversation_turns")
+
+
+def get_prompt(self) -> str:
+ ret = ""
+ for role, msg in self.get_turns():
+ ret += role + msg
+ return ret
+
+
+def get_turns( # pylint: disable=too-many-return-statements
+ self,
+) -> Generator[Tuple[str, str], None, None]:
+ """Get the prompt for generation."""
+ system_prompt = self.system_template.format(system_message=self.system_message)
+ if self.sep_style == SeparatorStyle.ADD_COLON_SINGLE:
+ yield "", system_prompt + self.sep
+ for role, message in self.messages:
+ if message:
+ yield role + ": ", message + self.sep
+ else:
+ yield role + ":", ""
+ return
+ if self.sep_style == SeparatorStyle.ADD_COLON_TWO:
+ seps = [self.sep, self.sep2]
+ yield "", system_prompt + seps[0]
+ for i, (role, message) in enumerate(self.messages):
+ if message:
+ yield role + ": ", message + seps[i % 2]
+ else:
+ yield role + ":", ""
+ return
+ if self.sep_style == SeparatorStyle.ADD_COLON_SPACE_SINGLE:
+ yield "", system_prompt + self.sep
+ for role, message in self.messages:
+ if message:
+ yield role + ": ", message + self.sep
+ else:
+ yield role + ": ", "" # must be end with a space
+ return
+ if self.sep_style == SeparatorStyle.ADD_NEW_LINE_SINGLE:
+ yield "", "" if system_prompt == "" else system_prompt + self.sep
+ for role, message in self.messages:
+ if message:
+ yield role + "\n", message + self.sep
+ else:
+ yield role + "\n", ""
+ return
+ if self.sep_style == SeparatorStyle.NO_COLON_SINGLE:
+ yield "", system_prompt
+ for role, message in self.messages:
+ if message:
+ yield role, message + self.sep
+ else:
+ yield role, ""
+ return
+ if self.sep_style == SeparatorStyle.NO_COLON_TWO:
+ seps = [self.sep, self.sep2]
+ yield "", system_prompt
+ for i, (role, message) in enumerate(self.messages):
+ if message:
+ yield role, message + seps[i % 2]
+ else:
+ yield role, ""
+ return
+ if self.sep_style == SeparatorStyle.RWKV:
+ yield "", system_prompt
+ for i, (role, message) in enumerate(self.messages):
+ if message:
+ yield role + ": ", message.replace("\r\n", "\n").replace(
+ "\n\n", "\n"
+ ) + "\n\n"
+ else:
+ yield role + ":", ""
+ return
+ if self.sep_style == SeparatorStyle.LLAMA2 and self.name != "mistral":
+ if self.system_message:
+ if self.messages:
+ # For llama, the system message is incorporated into the first human instruction
+ first_role, first_msg = self.messages[0]
+ if first_role == self.roles[0]:
+ system_prompt += first_msg
+ self.messages.pop(0)
+ yield "", system_prompt
+ for i, (role, message) in enumerate(self.messages):
+ if message:
+ if (i % 2 == 0 and not self.system_message) or (
+ i % 2 != 0 and self.system_message
+ ):
+ role = " " + role
+ yield role + " ", message
+ else:
+ yield role, ""
+ return
+ if self.sep_style == SeparatorStyle.LLAMA2 and self.name == "mistral":
+ contains_sys_msg = False
+ if self.system_message:
+ contains_sys_msg = True
+ if self.messages:
+ # There is no clear guidance on how to handle system messages in Mistral so we just prepend it to the first human instruction separated by a newline
+ first_role, first_msg = self.messages[0]
+ if first_role == self.roles[0]:
+ system_prompt = self.system_template.format(
+ system_message=" " + self.system_message
+ )
+ system_prompt += first_msg
+ self.messages.pop(0)
+ yield "", system_prompt
+ for i, (role, message) in enumerate(self.messages):
+ if message and i == 0 and not contains_sys_msg:
+ yield "", system_prompt.strip() + " " + message # if there is no system message, we need to make sure there is the a ` [INST]` at the beginning of the first instruction.
+ elif message:
+ yield role + " ", message
+ else:
+ yield role, ""
+ return
+ if self.sep_style == SeparatorStyle.LLAMA3:
+ if self.system_message:
+ # For llama3, the system message is NOT incorporated into the first human instruction
+ # All messages follow <|start_header_id|>' + role + '<|end_header_id|>\n\n'+ message + '<|eot_id|>
+ yield "", system_prompt
+ for i, (role, message) in enumerate(self.messages):
+ if message:
+ yield f"<|start_header_id|>{role}<|end_header_id|>\n\n", f"{message.strip()}<|eot_id|>"
+ else:
+ yield f"<|start_header_id|>{role}<|end_header_id|>\n\n", ""
+ return
+ if self.sep_style == SeparatorStyle.GEMMA:
+ if self.system_message:
+ raise ValueError("Gemma chat template does not support system messages")
+ for i, (role, message) in enumerate(self.messages):
+ prefix = "" if i == 0 else ""
+ message_str = message if message else ""
+ yield prefix + "" + role + "\n", message_str + "\n"
+ return
+ if self.sep_style == SeparatorStyle.CHATGLM:
+ # source: https://huggingface.co/THUDM/chatglm-6b/blob/1d240ba371910e9282298d4592532d7f0f3e9f3e/modeling_chatglm.py#L1302-L1308
+ # source2: https://huggingface.co/THUDM/chatglm2-6b/blob/e186c891cf64310ac66ef10a87e6635fa6c2a579/modeling_chatglm.py#L926
+ round_add_n = 1 if self.name == "chatglm2" else 0
+ if system_prompt:
+ yield "", system_prompt + self.sep
+
+ for i, (role, message) in enumerate(self.messages):
+ if i % 2 == 0:
+ yield "", f"[Round {i//2 + round_add_n}]{self.sep}"
+
+ if message:
+ yield f"{role}:", f"{message}{self.sep}"
+ else:
+ yield f"{role}:", ""
+ return
+ if self.sep_style == SeparatorStyle.CHATML:
+ yield "", "" if system_prompt == "" else system_prompt + self.sep + "\n"
+ for role, message in self.messages:
+ if message:
+ yield role + "\n", message + self.sep + "\n"
+ else:
+ yield role + "\n", ""
+ return
+ if self.sep_style == SeparatorStyle.CHATGLM3:
+ if self.system_message:
+ yield "", system_prompt
+ for role, message in self.messages:
+ if message:
+ yield role + "\n", " " + message
+ else:
+ yield role
+ return
+ if self.sep_style == SeparatorStyle.CHATINTERN:
+ # source: https://huggingface.co/internlm/internlm-chat-7b-8k/blob/bd546fa984b4b0b86958f56bf37f94aa75ab8831/modeling_internlm.py#L771
+ seps = [self.sep, self.sep2]
+ yield "", system_prompt
+ for i, (role, message) in enumerate(self.messages):
+ prefix = "" if i % 2 == 0 else ""
+ if message:
+ yield prefix + role + ":", message + seps[i % 2] + "\n"
+ else:
+ yield role + ":", ""
+ return
+ if self.sep_style == SeparatorStyle.DOLLY:
+ seps = [self.sep, self.sep2]
+ yield "", system_prompt
+ for i, (role, message) in enumerate(self.messages):
+ if message:
+ suffix = "\n\n" if i % 2 == 1 else ""
+ yield role + ":\n", message + seps[i % 2] + suffix
+ else:
+ yield role + ":\n", ""
+ return
+ if self.sep_style == SeparatorStyle.PHOENIX:
+ yield "", system_prompt
+ for role, message in self.messages:
+ if message:
+ yield role + ": ", "" + message + " "
+ else:
+ yield role + ": " + "", ""
+ return
+ if self.sep_style == SeparatorStyle.ROBIN:
+ yield "", system_prompt + self.sep
+ for role, message in self.messages:
+ if message:
+ yield role + ":\n", message + self.sep
+ else:
+ yield role + ":\n", ""
+ return
+ if self.sep_style == SeparatorStyle.FALCON_CHAT:
+ if self.system_message:
+ yield "", system_prompt + self.sep
+ for role, message in self.messages:
+ if message:
+ yield role + ": ", message + self.sep
+ else:
+ yield role + ":", ""
+ else:
+ raise ValueError(f"Invalid style: {self.sep_style}")
+
+
+def add_get_turns_to_conversation():
+ import fastchat.conversation
+
+ fastchat.conversation.Conversation.get_turns = get_turns
+ fastchat.conversation.Conversation.get_prompt = get_prompt
diff --git a/src/axolotl/monkeypatch/llama_attn_hijack_flash.py b/src/axolotl/monkeypatch/llama_attn_hijack_flash.py
new file mode 100644
index 0000000000000000000000000000000000000000..dda5da2b7a78258f656f422772dc6138588d432d
--- /dev/null
+++ b/src/axolotl/monkeypatch/llama_attn_hijack_flash.py
@@ -0,0 +1,929 @@
+"""Flash attention monkey patch for llama model"""
+
+# copied from https://github.com/lm-sys/FastChat/blob/main/fastchat/train/llama_flash_attn_monkey_patch.py
+
+import logging
+import warnings
+from functools import partial
+from typing import List, Optional, Tuple, Union
+
+import torch
+import torch.nn.functional as F
+import transformers
+from einops import rearrange
+from flash_attn.bert_padding import pad_input, unpad_input
+from transformers.modeling_outputs import BaseModelOutputWithPast
+from transformers.models.llama.modeling_llama import LlamaAttention
+from transformers.models.llama.modeling_llama import (
+ LlamaDecoderLayer as OriginalLlamaDecoderLayer,
+)
+from transformers.models.llama.modeling_llama import (
+ LlamaMLP,
+ apply_rotary_pos_emb,
+ repeat_kv,
+)
+from xformers.ops import SwiGLU
+
+from axolotl.monkeypatch.utils import get_cu_seqlens_from_pos_ids, set_module_name
+
+try:
+ from flash_attn.flash_attn_interface import ( # pylint: disable=ungrouped-imports
+ flash_attn_kvpacked_func,
+ flash_attn_varlen_kvpacked_func,
+ flash_attn_varlen_qkvpacked_func,
+ )
+except ImportError:
+ from flash_attn.flash_attn_interface import (
+ flash_attn_unpadded_kvpacked_func as flash_attn_varlen_kvpacked_func,
+ )
+ from flash_attn.flash_attn_interface import (
+ flash_attn_unpadded_qkvpacked_func as flash_attn_varlen_qkvpacked_func,
+ )
+
+
+LOG = logging.getLogger("axolotl")
+
+
+def is_xformers_swiglu_available() -> bool:
+ from xformers.ops.common import get_xformers_operator
+
+ try:
+ get_xformers_operator("swiglu_packedw")()
+ return True
+ except RuntimeError as exc:
+ if "No such operator xformers::swiglu_packedw " in str(exc):
+ return False
+ return True
+
+
+def replace_llama_mlp_with_swiglu(model):
+ for name, module in model.named_modules():
+ if isinstance(module, LlamaMLP):
+ mlp = FusedMLP(
+ module.config, module.gate_proj, module.up_proj, module.down_proj
+ )
+ set_module_name(model, name, mlp)
+
+
+def replace_llama_qkv_with_fused(model):
+ for name, module in model.named_modules():
+ if isinstance(module, LlamaAttention):
+ qkv = FusedAttention(
+ module.config,
+ module.q_proj,
+ module.k_proj,
+ module.v_proj,
+ module.o_proj,
+ )
+ set_module_name(model, name, qkv)
+
+
+def replace_llama_attn_with_flash_attn(
+ packed: Optional[bool] = False,
+ cross_entropy: Optional[bool] = False,
+ rms_norm: Optional[bool] = False,
+ use_shifted_sparse_attn: Optional[bool] = False,
+):
+ transformers.models.llama.modeling_llama.LlamaModel._prepare_decoder_attention_mask = ( # pylint: disable=protected-access
+ _prepare_decoder_attention_mask
+ )
+ if use_shifted_sparse_attn:
+ transformers.models.llama.modeling_llama.LlamaAttention.forward = (
+ flashattn_forward_with_s2attn
+ )
+ else:
+ transformers.models.llama.modeling_llama.LlamaAttention.forward = (
+ flashattn_forward
+ )
+
+ if packed:
+ transformers.models.llama.modeling_llama.LlamaDecoderLayer = LlamaDecoderLayer
+ transformers.models.llama.modeling_llama.LlamaModel.forward = (
+ llama_model_forward
+ )
+
+ # skip only if explicitly disabled
+ if cross_entropy:
+ try:
+ from flash_attn.losses.cross_entropy import CrossEntropyLoss
+
+ LOG.info("patching with flash_attn.losses.cross_entropy")
+ transformers.models.llama.modeling_llama.CrossEntropyLoss = partial(
+ CrossEntropyLoss, inplace_backward=True
+ )
+ except ImportError:
+ LOG.info(
+ "optimized flash-attention CrossEntropyLoss not found (run `pip install 'git+https://github.com/Dao-AILab/flash-attention.git#egg=xentropy_cuda_lib&subdirectory=csrc/xentropy'`)"
+ )
+
+ # skip only if explicitly disabled
+ if rms_norm:
+ try:
+ from flash_attn.ops.rms_norm import RMSNorm
+
+ class LlamaRMSNorm(RMSNorm):
+ """Patched LLamaRMSNorm"""
+
+ def __init__(self, hidden_size, eps=1e-6):
+ super().__init__(hidden_size, eps=eps)
+
+ LOG.info("patching with flash_attn.ops.rms_norm")
+ transformers.models.llama.modeling_llama.LlamaRMSNorm = LlamaRMSNorm
+ except ImportError:
+ LOG.info(
+ "optimized flash-attention RMSNorm not found (run `pip install 'git+https://github.com/Dao-AILab/flash-attention.git#egg=dropout_layer_norm&subdirectory=csrc/layer_norm'`)"
+ )
+
+
+class FusedAttention(LlamaAttention):
+ """
+ Fused QKV Attention layer for incrementally improved training efficiency
+ """
+
+ def __init__(
+ self,
+ config,
+ q: torch.nn.Linear, # pylint: disable=invalid-name
+ k: torch.nn.Linear, # pylint: disable=invalid-name
+ v: torch.nn.Linear, # pylint: disable=invalid-name
+ o: torch.nn.Linear, # pylint: disable=invalid-name
+ ):
+ super().__init__(config)
+ self.config = config
+ self.init_device = next(iter(q.state_dict().values())).device
+
+ # define equivalent fused qkv projection
+ self.out_features: List[int] = [q.out_features, k.out_features, v.out_features]
+ self.qkv_proj = torch.nn.Linear(
+ q.in_features, sum(self.out_features), device=self.init_device, bias=False
+ )
+ self.o_proj = o
+
+ # overwrite initialized weights with pretrained weights
+ self.qkv_proj.weight.data = torch.cat(
+ (q.weight.data, k.weight.data, v.weight.data), dim=0
+ )
+
+ def _post_training(self, model, name):
+ q_proj, k_proj, v_proj = torch.split(
+ self.qkv_proj.weight.data, self.out_features, dim=0
+ )
+
+ new_attn = LlamaAttention(self.config)
+ new_attn.q_proj.weight.data = q_proj
+ new_attn.k_proj.weight.data = k_proj
+ new_attn.v_proj.weight.data = v_proj
+ new_attn.o_proj.weight.data = self.o_proj.weight.data
+
+ set_module_name(model, name, new_attn)
+
+
+class FusedMLP(torch.nn.Module):
+ """
+ Fused MLP layer for incrementally improved training efficiency
+ """
+
+ def __init__(
+ self,
+ config,
+ gate_proj: torch.nn.Linear,
+ up_proj: torch.nn.Linear,
+ down_proj: torch.nn.Linear,
+ ):
+ super().__init__()
+ self.config = config
+ self.swiglu = SwiGLU(
+ in_features=config.hidden_size,
+ hidden_features=config.intermediate_size,
+ bias=False,
+ _pack_weights=True,
+ )
+ # overwrite initialized weights with pretrained weights
+ self.swiglu.w12.weight.data = torch.cat(
+ (gate_proj.weight.data, up_proj.weight.data), dim=0
+ )
+ self.swiglu.w3.weight.data = down_proj.weight.data
+
+ def _post_training(self, model, name):
+ w1, w2 = torch.split( # pylint: disable=invalid-name
+ self.swiglu.w12.weight.data, self.config.intermediate_size, dim=0
+ )
+
+ # Assign the split weights back to the original layers
+ new_mlp = LlamaMLP(self.config)
+ new_mlp.gate_proj.weight.data = w1
+ new_mlp.up_proj.weight.data = w2
+ new_mlp.down_proj.weight.data = self.swiglu.w3.weight.data
+
+ set_module_name(model, name, new_mlp)
+
+ def forward(self, x: torch.Tensor) -> torch.Tensor: # pylint: disable=invalid-name
+ return self.swiglu(x)
+
+
+# Disable the transformation of the attention mask in LlamaModel as the flash attention
+# requires the attention mask to be the same as the key_padding_mask
+def _prepare_decoder_attention_mask(
+ self,
+ attention_mask,
+ input_shape,
+ inputs_embeds,
+ past_key_values_length,
+): # pylint: disable=unused-argument
+ # [bsz, seq_len]
+ return attention_mask
+
+
+GROUP_SIZE_RATIO = 1 / 4
+
+
+def flashattn_forward_with_s2attn(
+ self,
+ hidden_states: torch.Tensor,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.Tensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: bool = False,
+ use_cache: bool = False,
+ padding_mask: Optional[torch.LongTensor] = None, # pylint: disable=unused-argument
+ cu_seqlens: Optional[torch.Tensor] = None, # pylint: disable=unused-argument
+ max_seqlen: Optional[torch.Tensor] = None, # pylint: disable=unused-argument
+) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
+ """Input shape: Batch x Time x Channel
+
+ From: https://github.com/dvlab-research/LongLoRA/blob/main/llama_attn_replace.py
+
+ attention_mask: [bsz, q_len]
+
+ `cu_seqlens` will be ignored if provided
+ `max_seqlen` will be ignored if provided
+ """
+ if output_attentions:
+ warnings.warn(
+ "Output attentions is not supported for patched `LlamaAttention`, returning `None` instead."
+ )
+
+ bsz, q_len, _ = hidden_states.size()
+
+ query_states = (
+ self.q_proj(hidden_states)
+ .view(bsz, q_len, self.num_heads, self.head_dim)
+ .transpose(1, 2)
+ )
+ key_states = (
+ self.k_proj(hidden_states)
+ .view(bsz, q_len, self.num_key_value_heads, self.head_dim)
+ .transpose(1, 2)
+ )
+ value_states = (
+ self.v_proj(hidden_states)
+ .view(bsz, q_len, self.num_key_value_heads, self.head_dim)
+ .transpose(1, 2)
+ )
+ # [bsz, q_len, nh, hd]
+ # [bsz, nh, q_len, hd]
+ # pylint: disable=duplicate-code
+
+ cos, sin = self.rotary_emb(value_states, position_ids=position_ids)
+ query_states, key_states = apply_rotary_pos_emb(
+ query_states, key_states, cos, sin, position_ids
+ )
+
+ # Past Key value support
+ if past_key_value is not None:
+ # reuse k, v, self_attention
+ key_states = torch.cat([past_key_value[0], key_states], dim=2)
+ value_states = torch.cat([past_key_value[1], value_states], dim=2)
+
+ past_key_value = (key_states, value_states) if use_cache else None
+
+ # repeat k/v heads if n_kv_heads < n_heads
+ key_states = repeat_kv(key_states, self.num_key_value_groups)
+ value_states = repeat_kv(value_states, self.num_key_value_groups)
+
+ # Flash attention codes from
+ # https://github.com/HazyResearch/flash-attention/blob/main/flash_attn/flash_attention.py
+
+ # transform the data into the format required by flash attention
+ qkv = torch.stack(
+ [query_states, key_states, value_states], dim=2
+ ) # [bsz, nh, 3, q_len, hd]
+ qkv = qkv.transpose(1, 3) # [bsz, q_len, 3, nh, hd]
+
+ # We have disabled _prepare_decoder_attention_mask in LlamaModel
+ # the attention_mask should be the same as the key_padding_mask
+
+ key_padding_mask = attention_mask.repeat(2, 1)
+ nheads = qkv.shape[-2]
+ # shift
+
+ group_size = int(q_len * GROUP_SIZE_RATIO)
+ if q_len % group_size > 0:
+ raise ValueError(
+ f"q_len {q_len} should be divisible by group size {group_size}."
+ )
+
+ qkv = (
+ qkv.reshape(bsz, q_len, 3, 2, self.num_heads // 2, self.head_dim)
+ .permute(0, 3, 1, 2, 4, 5)
+ .reshape(bsz * 2, q_len, 3, self.num_heads // 2, self.head_dim)
+ )
+ x = rearrange( # pylint: disable=invalid-name
+ qkv, "b s three h d -> b s (three h d)"
+ )
+ x_unpad, indices, cu_q_lens, max_s = unpad_input(x, key_padding_mask)
+ cu_q_len_tmp = torch.arange(
+ 0, max_s, group_size, device=key_padding_mask.device, dtype=cu_q_lens.dtype
+ )
+ cu_q_len_tmp = torch.stack([cu_q_len_tmp, cu_q_len_tmp + group_size // 2]).repeat(
+ bsz, 1
+ ) + cu_q_lens[:-1].unsqueeze(-1)
+ cu_q_lens = torch.cat([cu_q_len_tmp, cu_q_lens[1:].unsqueeze(-1)], dim=-1).view(-1)
+
+ x_unpad = rearrange(
+ x_unpad, "nnz (three h d) -> nnz three h d", three=3, h=nheads // 2
+ )
+ output_unpad = flash_attn_varlen_qkvpacked_func(
+ x_unpad, cu_q_lens, group_size, 0.0, softmax_scale=None, causal=True
+ )
+ output = rearrange(
+ pad_input(
+ rearrange(output_unpad, "nnz h d -> nnz (h d)"), indices, bsz * 2, q_len
+ ),
+ "b s (h d) -> b s h d",
+ h=nheads // 2,
+ )
+ output = (
+ output.reshape(bsz, 2, q_len, nheads // 2, self.head_dim)
+ .transpose(1, 2)
+ .reshape(bsz, q_len, nheads, self.head_dim)
+ )
+ return self.o_proj(rearrange(output, "b s h d -> b s (h d)")), None, past_key_value
+
+
+def flashattn_forward(
+ self,
+ hidden_states: torch.Tensor,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.Tensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: bool = False,
+ use_cache: bool = False,
+ padding_mask: Optional[torch.LongTensor] = None, # pylint: disable=unused-argument
+ cu_seqlens: Optional[torch.Tensor] = None,
+ max_seqlen: Optional[torch.Tensor] = None,
+) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
+ """Input shape: Batch x Time x Channel
+
+ attention_mask: [bsz, q_len]
+ """
+ # pylint: disable=duplicate-code
+ bsz, q_len, _ = hidden_states.size()
+
+ if not hasattr(self, "pretraining_tp"):
+ self.pretraining_tp = 1
+
+ if self.pretraining_tp > 1:
+ key_value_slicing = (
+ self.num_key_value_heads * self.head_dim
+ ) // self.pretraining_tp
+ query_slices = self.q_proj.weight.split(
+ (self.num_heads * self.head_dim) // self.pretraining_tp, dim=0
+ )
+ key_slices = self.k_proj.weight.split(key_value_slicing, dim=0)
+ value_slices = self.v_proj.weight.split(key_value_slicing, dim=0)
+
+ query_states = [
+ F.linear(hidden_states, query_slices[i]) for i in range(self.pretraining_tp)
+ ]
+ query_states = torch.cat(query_states, dim=-1)
+
+ key_states = [
+ F.linear(hidden_states, key_slices[i]) for i in range(self.pretraining_tp)
+ ]
+ key_states = torch.cat(key_states, dim=-1)
+
+ value_states = [
+ F.linear(hidden_states, value_slices[i]) for i in range(self.pretraining_tp)
+ ]
+ value_states = torch.cat(value_states, dim=-1)
+
+ else:
+ if isinstance(self, FusedAttention):
+ query_states, key_states, value_states = self.qkv_proj(hidden_states).split(
+ self.out_features, dim=-1
+ )
+ else:
+ query_states = self.q_proj(hidden_states)
+ key_states = self.k_proj(hidden_states)
+ value_states = self.v_proj(hidden_states)
+
+ query_states = query_states.view(
+ bsz, q_len, self.num_heads, self.head_dim
+ ).transpose(1, 2)
+ key_states = key_states.view(
+ bsz, q_len, self.num_key_value_heads, self.head_dim
+ ).transpose(1, 2)
+ value_states = value_states.view(
+ bsz, q_len, self.num_key_value_heads, self.head_dim
+ ).transpose(1, 2)
+ # [bsz, q_len, nh, hd]
+ # [bsz, nh, q_len, hd]
+
+ cos, sin = self.rotary_emb(value_states, position_ids=position_ids)
+ query_states, key_states = apply_rotary_pos_emb(
+ query_states, key_states, cos, sin, position_ids
+ )
+ # [bsz, nh, t, hd]
+
+ if past_key_value is not None:
+ # reuse k, v, self_attention
+ key_states = torch.cat([past_key_value[0], key_states], dim=2)
+ value_states = torch.cat([past_key_value[1], value_states], dim=2)
+
+ past_key_value = (key_states, value_states) if use_cache else None
+
+ # repeat k/v heads if n_kv_heads < n_heads
+ key_states = repeat_kv(key_states, self.num_key_value_groups)
+ value_states = repeat_kv(value_states, self.num_key_value_groups)
+
+ if output_attentions:
+ warnings.warn(
+ "Output attentions is not supported for patched `LlamaAttention`, returning `None` instead."
+ )
+
+ #
+ # flash-attn v2 start
+ #
+
+ if self.training:
+ # during training q,k,v always have same seqlen
+ assert key_states.shape == query_states.shape
+ is_causal = True
+ else:
+ # turn off FA causal mask after first inference autoregressive iteration
+ # only on first autoregressive step q,k,v have same seqlen
+ is_causal = key_states.shape == query_states.shape
+
+ dropout_rate = 0.0 if not self.training else getattr(self, "attention_dropout", 0.0)
+
+ if cu_seqlens is not None and max_seqlen is not None and cu_seqlens.dim() == 1:
+ # special handling using sample packing
+ qkv = torch.stack(
+ [query_states, key_states, value_states], dim=2
+ ) # [bsz, nh, 3, q_len, hd]
+ qkv = qkv.transpose(1, 3) # [bsz, q_len, 3, nh, hd]
+ qkv = rearrange(qkv, "b s ... -> (b s) ...")
+
+ output = flash_attn_varlen_qkvpacked_func(
+ qkv,
+ cu_seqlens,
+ max_seqlen,
+ dropout_p=dropout_rate,
+ softmax_scale=None,
+ causal=True,
+ )
+ output = rearrange(output, "(b s) ... -> b s ...", b=bsz)
+ elif query_states.shape == key_states.shape:
+ query_states = query_states.transpose(1, 2)
+ key_states = key_states.transpose(1, 2)
+ value_states = value_states.transpose(1, 2)
+ qkv_unpad, cu_seqlens_q, max_seqlen_q, _, output_pad_fn = generate_qkv(
+ query_states,
+ key_states,
+ value_states,
+ qkvpacked=True,
+ # We have disabled _prepare_decoder_attention_mask in LlamaModel
+ # the attention_mask should be the same as the key_padding_mask
+ key_padding_mask=attention_mask,
+ query_padding_mask=attention_mask[:, -query_states.size(1) :]
+ if attention_mask is not None
+ else None,
+ )
+ output_unpad = flash_attn_varlen_qkvpacked_func(
+ qkv_unpad,
+ cu_seqlens_q,
+ max_seqlen_q,
+ dropout_p=dropout_rate,
+ softmax_scale=None,
+ causal=is_causal,
+ )
+ output = output_pad_fn(output_unpad)
+ else:
+ query_states = query_states.transpose(1, 2)
+ key_states = key_states.transpose(1, 2)
+ value_states = value_states.transpose(1, 2)
+ if attention_mask is None or attention_mask.all().item():
+ output = flash_attn_kvpacked_func(
+ query_states,
+ torch.stack([key_states, value_states], 2),
+ dropout_p=dropout_rate,
+ causal=is_causal,
+ )
+ else:
+ ( # pylint: disable=unbalanced-tuple-unpacking
+ q_unpad,
+ kv_unpad,
+ cu_seqlens_q,
+ cu_seqlens_k,
+ max_seqlen_q,
+ max_seqlen_k,
+ _,
+ _,
+ output_pad_fn,
+ ) = generate_qkv(
+ query_states,
+ key_states,
+ value_states,
+ kvpacked=True,
+ key_padding_mask=attention_mask,
+ query_padding_mask=attention_mask[:, -query_states.size(1) :]
+ if attention_mask is not None
+ else None,
+ )
+ if q_unpad.dtype != kv_unpad.dtype:
+ kv_unpad = kv_unpad.to(q_unpad.dtype)
+ output_unpad = flash_attn_varlen_kvpacked_func(
+ q_unpad,
+ kv_unpad,
+ cu_seqlens_q,
+ cu_seqlens_k,
+ max_seqlen_q,
+ max_seqlen_k,
+ dropout_p=dropout_rate,
+ softmax_scale=None,
+ causal=is_causal,
+ )
+ output = output_pad_fn(output_unpad)
+
+ attn_output = output
+ if attn_output.size() != (bsz, q_len, self.num_heads, self.head_dim):
+ raise ValueError(
+ f"`attn_output` should be of size {(bsz, q_len, self.num_heads, self.head_dim)}, but is"
+ f" {attn_output.size()}"
+ )
+ attn_output = rearrange(attn_output, "b s h d -> b s (h d)")
+
+ #
+ # flash-attn v2 end
+ #
+
+ if self.pretraining_tp > 1:
+ attn_output = attn_output.split(self.hidden_size // self.pretraining_tp, dim=2)
+ o_proj_slices = self.o_proj.weight.split(
+ self.hidden_size // self.pretraining_tp, dim=1
+ )
+ attn_output = sum(
+ F.linear(attn_output[i], o_proj_slices[i])
+ for i in range(self.pretraining_tp)
+ )
+ else:
+ attn_output = self.o_proj(attn_output)
+
+ return attn_output, None, past_key_value
+
+
+# based on https://github.com/Dao-AILab/flash-attention/blob/364a5b/tests/test_flash_attn.py#L38
+def generate_qkv(
+ q,
+ k,
+ v,
+ query_padding_mask=None,
+ key_padding_mask=None,
+ kvpacked=False,
+ qkvpacked=False,
+): # pylint: disable=invalid-name,unnecessary-lambda-assignment
+ """
+ Arguments:
+ q: (batch_size, seqlen_q, nheads, d)
+ k: (batch_size, seqlen_k, nheads_k, d)
+ v: (batch_size, seqlen_k, nheads_k, d)
+ query_padding_mask: (batch_size, seqlen), bool
+ key_padding_mask: (batch_size, seqlen), bool
+ """
+ assert not (kvpacked and qkvpacked)
+ batch_size, seqlen_q, nheads, d = q.shape
+ _, seqlen_k, nheads_k, _ = k.shape
+ assert k.shape == (batch_size, seqlen_k, nheads_k, d)
+ assert v.shape == (batch_size, seqlen_k, nheads_k, d)
+
+ if query_padding_mask is not None:
+ q_unpad, indices_q, cu_seqlens_q, max_seqlen_q = unpad_input(
+ q, query_padding_mask
+ )
+
+ output_pad_fn = lambda output_unpad: pad_input( # noqa: E731
+ output_unpad, indices_q, batch_size, seqlen_q
+ )
+
+ else:
+ q_unpad = rearrange(q, "b s h d -> (b s) h d")
+ cu_seqlens_q = torch.arange(
+ 0,
+ (batch_size + 1) * seqlen_q,
+ step=seqlen_q,
+ dtype=torch.int32,
+ device=q_unpad.device,
+ )
+ max_seqlen_q = seqlen_q
+
+ output_pad_fn = lambda output_unpad: rearrange( # noqa: E731
+ output_unpad, "(b s) h d -> b s h d", b=batch_size
+ )
+
+ if key_padding_mask is not None:
+ k_unpad, _, cu_seqlens_k, max_seqlen_k = unpad_input(k, key_padding_mask)
+ v_unpad, _, _, _ = unpad_input(v, key_padding_mask)
+ else:
+ k_unpad = rearrange(k, "b s h d -> (b s) h d")
+ v_unpad = rearrange(v, "b s h d -> (b s) h d")
+ cu_seqlens_k = torch.arange(
+ 0,
+ (batch_size + 1) * seqlen_k,
+ step=seqlen_k,
+ dtype=torch.int32,
+ device=k_unpad.device,
+ )
+ max_seqlen_k = seqlen_k
+
+ if qkvpacked:
+ assert nheads == nheads_k
+ qkv_unpad = torch.stack([q_unpad, k_unpad, v_unpad], dim=1)
+ qkv = torch.stack([q, k, v], dim=2)
+ return (qkv_unpad, cu_seqlens_q, max_seqlen_q, qkv, output_pad_fn)
+
+ if kvpacked:
+ kv_unpad = torch.stack([k_unpad, v_unpad], dim=1)
+ kv = torch.stack([k, v], dim=2)
+ return (
+ q_unpad,
+ kv_unpad,
+ cu_seqlens_q,
+ cu_seqlens_k,
+ max_seqlen_q,
+ max_seqlen_k,
+ q,
+ kv,
+ output_pad_fn,
+ )
+
+ return (
+ q_unpad,
+ k_unpad,
+ v_unpad,
+ cu_seqlens_q,
+ cu_seqlens_k,
+ max_seqlen_q,
+ max_seqlen_k,
+ q,
+ k,
+ v,
+ output_pad_fn,
+ )
+
+
+def llama_model_forward(
+ self,
+ input_ids: torch.LongTensor = None,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_values: Optional[List[torch.FloatTensor]] = None,
+ inputs_embeds: Optional[torch.FloatTensor] = None,
+ use_cache: Optional[bool] = None,
+ output_attentions: Optional[bool] = None,
+ output_hidden_states: Optional[bool] = None,
+ return_dict: Optional[bool] = None,
+ cache_position: Optional[ # pylint: disable=unused-argument
+ torch.LongTensor
+ ] = None,
+) -> Union[Tuple, BaseModelOutputWithPast]:
+ output_attentions = (
+ output_attentions
+ if output_attentions is not None
+ else self.config.output_attentions
+ )
+ output_hidden_states = (
+ output_hidden_states
+ if output_hidden_states is not None
+ else self.config.output_hidden_states
+ )
+ use_cache = use_cache if use_cache is not None else self.config.use_cache
+
+ return_dict = (
+ return_dict if return_dict is not None else self.config.use_return_dict
+ )
+
+ # retrieve input_ids and inputs_embeds
+ if input_ids is not None and inputs_embeds is not None:
+ raise ValueError(
+ "You cannot specify both decoder_input_ids and decoder_inputs_embeds at the same time"
+ )
+ if input_ids is not None:
+ batch_size, seq_length = input_ids.shape
+ elif inputs_embeds is not None:
+ batch_size, seq_length, _ = inputs_embeds.shape
+ else:
+ raise ValueError(
+ "You have to specify either decoder_input_ids or decoder_inputs_embeds"
+ )
+
+ seq_length_with_past = seq_length
+ past_key_values_length = 0
+
+ if past_key_values is not None:
+ past_key_values_length = past_key_values[0][0].shape[2]
+ seq_length_with_past = seq_length_with_past + past_key_values_length
+
+ cu_seqlens = None
+ max_seqlen = None
+ if position_ids is None:
+ device = input_ids.device if input_ids is not None else inputs_embeds.device
+ position_ids = torch.arange(
+ past_key_values_length,
+ seq_length + past_key_values_length,
+ dtype=torch.long,
+ device=device,
+ )
+ position_ids = position_ids.unsqueeze(0).view(-1, seq_length)
+ else:
+ position_ids = position_ids.view(-1, seq_length).long()
+ cu_seqlens, max_seqlen = get_cu_seqlens_from_pos_ids(position_ids)
+ cu_seqlens = cu_seqlens.squeeze()
+
+ if inputs_embeds is None:
+ inputs_embeds = self.embed_tokens(input_ids)
+ # embed positions
+ if attention_mask is None:
+ attention_mask = torch.ones(
+ (batch_size, seq_length_with_past),
+ dtype=torch.bool,
+ device=inputs_embeds.device,
+ )
+ padding_mask = None
+ else:
+ if 0 in attention_mask:
+ padding_mask = attention_mask
+ else:
+ padding_mask = None
+
+ attention_mask = (
+ self._prepare_decoder_attention_mask( # pylint: disable=protected-access
+ attention_mask,
+ (batch_size, seq_length),
+ inputs_embeds,
+ past_key_values_length,
+ )
+ )
+
+ hidden_states = inputs_embeds
+
+ if self.gradient_checkpointing and self.training:
+ if use_cache:
+ transformers.logger.warning_once(
+ "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
+ )
+ use_cache = False
+
+ # decoder layers
+ all_hidden_states = () if output_hidden_states else None
+ all_self_attns = () if output_attentions else None
+ next_decoder_cache = () if use_cache else None
+
+ for idx, decoder_layer in enumerate(self.layers):
+ if output_hidden_states:
+ all_hidden_states += (hidden_states,)
+
+ past_key_value = past_key_values[idx] if past_key_values is not None else None
+
+ if self.gradient_checkpointing and self.training:
+
+ def create_custom_forward(module):
+ def custom_forward(*inputs):
+ # None for past_key_value
+ return module(
+ *inputs,
+ )
+
+ return custom_forward
+
+ layer_outputs = torch.utils.checkpoint.checkpoint(
+ create_custom_forward(decoder_layer),
+ hidden_states,
+ attention_mask,
+ position_ids,
+ past_key_value,
+ output_attentions,
+ None,
+ padding_mask,
+ cu_seqlens,
+ max_seqlen,
+ )
+ else:
+ layer_outputs = decoder_layer(
+ hidden_states,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ padding_mask=padding_mask,
+ cu_seqlens=cu_seqlens,
+ max_seqlen=max_seqlen,
+ )
+
+ hidden_states = layer_outputs[0]
+
+ if use_cache:
+ next_decoder_cache += (layer_outputs[2 if output_attentions else 1],)
+
+ if output_attentions:
+ all_self_attns += (layer_outputs[1],)
+
+ hidden_states = self.norm(hidden_states)
+
+ # add hidden states from the last decoder layer
+ if output_hidden_states:
+ all_hidden_states += (hidden_states,)
+
+ next_cache = next_decoder_cache if use_cache else None
+ if not return_dict:
+ return tuple(
+ v
+ for v in [hidden_states, next_cache, all_hidden_states, all_self_attns]
+ if v is not None
+ )
+ return BaseModelOutputWithPast(
+ last_hidden_state=hidden_states,
+ past_key_values=next_cache,
+ hidden_states=all_hidden_states,
+ attentions=all_self_attns,
+ )
+
+
+class LlamaDecoderLayer(OriginalLlamaDecoderLayer):
+ """
+ patched version of LlamaDecoderLayer to pass through the precalculated cu_seqlens
+ """
+
+ def forward(
+ self,
+ hidden_states: torch.Tensor,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: Optional[bool] = False,
+ use_cache: Optional[bool] = False,
+ padding_mask: Optional[torch.LongTensor] = None,
+ cu_seqlens: Optional[torch.Tensor] = None,
+ max_seqlen: Optional[torch.Tensor] = None,
+ ) -> Tuple[
+ torch.FloatTensor, Optional[Tuple[torch.FloatTensor, torch.FloatTensor]]
+ ]:
+ """
+ Args:
+ hidden_states (`torch.FloatTensor`): input to the layer of shape `(batch, seq_len, embed_dim)`
+ attention_mask (`torch.FloatTensor`, *optional*): attention mask of size
+ `(batch, 1, tgt_len, src_len)` where padding elements are indicated by very large negative values.
+ output_attentions (`bool`, *optional*):
+ Whether or not to return the attentions tensors of all attention layers. See `attentions` under
+ returned tensors for more detail.
+ use_cache (`bool`, *optional*):
+ If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding
+ (see `past_key_values`).
+ past_key_value (`Tuple(torch.FloatTensor)`, *optional*): cached past key and value projection states
+ cu_seqlens (`torch.Tensor`, *optional*) cumulative sequence len when packing
+ """
+
+ residual = hidden_states
+
+ hidden_states = self.input_layernorm(hidden_states)
+
+ # Self Attention
+ hidden_states, self_attn_weights, present_key_value = self.self_attn(
+ hidden_states=hidden_states,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ padding_mask=padding_mask,
+ cu_seqlens=cu_seqlens,
+ max_seqlen=max_seqlen,
+ )
+ hidden_states = residual + hidden_states
+
+ # Fully Connected
+ residual = hidden_states
+ hidden_states = self.post_attention_layernorm(hidden_states)
+ hidden_states = self.mlp(hidden_states)
+ hidden_states = residual + hidden_states
+
+ outputs = (hidden_states,)
+
+ if output_attentions:
+ outputs += (self_attn_weights,)
+
+ if use_cache:
+ outputs += (present_key_value,)
+
+ return outputs
diff --git a/src/axolotl/monkeypatch/llama_attn_hijack_xformers.py b/src/axolotl/monkeypatch/llama_attn_hijack_xformers.py
new file mode 100644
index 0000000000000000000000000000000000000000..0c1a4e8224f0acc22d9c0eb190f6aaae377373d2
--- /dev/null
+++ b/src/axolotl/monkeypatch/llama_attn_hijack_xformers.py
@@ -0,0 +1,153 @@
+"""
+Directly copied the code from https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/modules/llama_attn_hijack.py and made some adjustments
+"""
+
+import logging
+import warnings
+from typing import Optional, Tuple
+
+import torch
+import torch.nn.functional as F
+import transformers.models.llama.modeling_llama
+from transformers.models.llama.modeling_llama import apply_rotary_pos_emb, repeat_kv
+
+try:
+ import xformers.ops
+except ImportError:
+ logging.error("xformers not found! Please install it before trying to use it.")
+
+
+def hijack_llama_attention():
+ transformers.models.llama.modeling_llama.LlamaAttention.forward = xformers_forward
+
+
+def xformers_forward(
+ self,
+ hidden_states: torch.Tensor,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: bool = False,
+ use_cache: bool = False,
+ padding_mask: Optional[torch.LongTensor] = None, # pylint: disable=unused-argument
+ **kwargs, # pylint: disable=unused-argument
+) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
+ # pylint: disable=duplicate-code
+ bsz, q_len, _ = hidden_states.size()
+
+ if not hasattr(self, "pretraining_tp"):
+ self.pretraining_tp = 1
+
+ if self.pretraining_tp > 1:
+ key_value_slicing = (
+ self.num_key_value_heads * self.head_dim
+ ) // self.pretraining_tp
+ query_slices = self.q_proj.weight.split(
+ (self.num_heads * self.head_dim) // self.pretraining_tp, dim=0
+ )
+ key_slices = self.k_proj.weight.split(key_value_slicing, dim=0)
+ value_slices = self.v_proj.weight.split(key_value_slicing, dim=0)
+
+ query_states = [
+ F.linear(hidden_states, query_slices[i]) for i in range(self.pretraining_tp)
+ ]
+ query_states = torch.cat(query_states, dim=-1)
+
+ key_states = [
+ F.linear(hidden_states, key_slices[i]) for i in range(self.pretraining_tp)
+ ]
+ key_states = torch.cat(key_states, dim=-1)
+
+ value_states = [
+ F.linear(hidden_states, value_slices[i]) for i in range(self.pretraining_tp)
+ ]
+ value_states = torch.cat(value_states, dim=-1)
+
+ else:
+ query_states = self.q_proj(hidden_states)
+ key_states = self.k_proj(hidden_states)
+ value_states = self.v_proj(hidden_states)
+
+ query_states = query_states.view(
+ bsz, q_len, self.num_heads, self.head_dim
+ ).transpose(1, 2)
+ key_states = key_states.view(
+ bsz, q_len, self.num_key_value_heads, self.head_dim
+ ).transpose(1, 2)
+ value_states = value_states.view(
+ bsz, q_len, self.num_key_value_heads, self.head_dim
+ ).transpose(1, 2)
+ # [bsz, q_len, nh, hd]
+ # [bsz, nh, q_len, hd]
+
+ cos, sin = self.rotary_emb(value_states)
+ query_states, key_states = apply_rotary_pos_emb(
+ query_states, key_states, cos, sin, position_ids
+ )
+ # [bsz, nh, t, hd]
+
+ if past_key_value is not None:
+ # reuse k, v, self_attention
+ key_states = torch.cat([past_key_value[0], key_states], dim=2)
+ value_states = torch.cat([past_key_value[1], value_states], dim=2)
+
+ past_key_value = (key_states, value_states) if use_cache else None
+
+ # repeat k/v heads if n_kv_heads < n_heads
+ key_states = repeat_kv(key_states, self.num_key_value_groups)
+ value_states = repeat_kv(value_states, self.num_key_value_groups)
+
+ if output_attentions:
+ warnings.warn(
+ "Output attentions is not supported for patched `LlamaAttention`, returning `None` instead."
+ )
+
+ #
+ # xformers-attn start
+ #
+
+ query_states = query_states.transpose(1, 2)
+ key_states = key_states.transpose(1, 2)
+ value_states = value_states.transpose(1, 2)
+
+ # This is a nasty hack. We know attention_mask in transformers is either LowerTriangular or all Zeros.
+ # We therefore check if one element in the upper triangular portion is zero. If it is, then the mask is all zeros.
+ if attention_mask is None or attention_mask[0, 0, 0, 1] == 0:
+ # input and output should be of form (bsz, q_len, num_heads, head_dim)
+ attn_output = xformers.ops.memory_efficient_attention(
+ query_states, key_states, value_states, attn_bias=None
+ )
+ else:
+ # input and output should be of form (bsz, q_len, num_heads, head_dim)
+ attn_output = xformers.ops.memory_efficient_attention(
+ query_states,
+ key_states,
+ value_states,
+ # attn_bias=attention_mask,
+ attn_bias=xformers.ops.LowerTriangularMask(),
+ )
+
+ if attn_output.size() != (bsz, q_len, self.num_heads, self.head_dim):
+ raise ValueError(
+ f"`attn_output` should be of size {(bsz, q_len, self.num_heads, self.head_dim)}, but is"
+ f" {attn_output.size()}"
+ )
+ attn_output = attn_output.reshape(bsz, q_len, self.hidden_size)
+
+ #
+ # xformers-attn end
+ #
+
+ if self.pretraining_tp > 1:
+ attn_output = attn_output.split(self.hidden_size // self.pretraining_tp, dim=2)
+ o_proj_slices = self.o_proj.weight.split(
+ self.hidden_size // self.pretraining_tp, dim=1
+ )
+ attn_output = sum(
+ F.linear(attn_output[i], o_proj_slices[i])
+ for i in range(self.pretraining_tp)
+ )
+ else:
+ attn_output = self.o_proj(attn_output)
+
+ return attn_output, None, past_key_value
diff --git a/src/axolotl/monkeypatch/llama_expand_mask.py b/src/axolotl/monkeypatch/llama_expand_mask.py
new file mode 100644
index 0000000000000000000000000000000000000000..5738bb543cf53d64e81ad61fee3e08390fc08c39
--- /dev/null
+++ b/src/axolotl/monkeypatch/llama_expand_mask.py
@@ -0,0 +1,25 @@
+"""
+expands the binary attention mask per 3.2.2 of https://arxiv.org/pdf/2107.02027.pdf
+"""
+from typing import Optional
+
+import torch
+
+from axolotl.monkeypatch.utils import mask_2d_to_4d
+
+
+def _expand_mask(mask: torch.Tensor, dtype: torch.dtype, tgt_len: Optional[int] = None):
+ masked_zero_one_mask = mask_2d_to_4d(mask, dtype, tgt_len)
+ inverted_mask = 1.0 - masked_zero_one_mask
+
+ return inverted_mask.masked_fill(
+ inverted_mask.to(torch.bool), torch.finfo(dtype).min
+ )
+
+
+def hijack_expand_mask():
+ import transformers
+
+ transformers.models.llama.modeling_llama._expand_mask = ( # pylint: disable=protected-access
+ _expand_mask
+ )
diff --git a/src/axolotl/monkeypatch/llama_patch_multipack.py b/src/axolotl/monkeypatch/llama_patch_multipack.py
new file mode 100644
index 0000000000000000000000000000000000000000..540c5577a0396788c6a824aa0ebf5af36aa79dd7
--- /dev/null
+++ b/src/axolotl/monkeypatch/llama_patch_multipack.py
@@ -0,0 +1,26 @@
+"""
+Patched LlamaAttention to use torch.nn.functional.scaled_dot_product_attention
+"""
+
+from axolotl.monkeypatch.utils import (
+ patched_prepare_4d_causal_attention_mask,
+ patched_prepare_4d_causal_attention_mask_for_sdpa,
+)
+
+
+def hijack_llama_prepare_4d_mask():
+ import transformers.modeling_attn_mask_utils
+ import transformers.models.llama.modeling_llama
+
+ transformers.models.llama.modeling_llama._prepare_4d_causal_attention_mask_for_sdpa = ( # pylint: disable=protected-access
+ patched_prepare_4d_causal_attention_mask_for_sdpa
+ )
+ transformers.modeling_attn_mask_utils._prepare_4d_causal_attention_mask_for_sdpa = ( # pylint: disable=protected-access
+ patched_prepare_4d_causal_attention_mask_for_sdpa
+ )
+ transformers.models.llama.modeling_llama._prepare_4d_causal_attention_mask = ( # pylint: disable=protected-access
+ patched_prepare_4d_causal_attention_mask
+ )
+ transformers.modeling_attn_mask_utils._prepare_4d_causal_attention_mask = ( # pylint: disable=protected-access
+ patched_prepare_4d_causal_attention_mask
+ )
diff --git a/src/axolotl/monkeypatch/mistral_attn_hijack_flash.py b/src/axolotl/monkeypatch/mistral_attn_hijack_flash.py
new file mode 100644
index 0000000000000000000000000000000000000000..6ae2e75fa23f562403a41fdc396b8157950651a0
--- /dev/null
+++ b/src/axolotl/monkeypatch/mistral_attn_hijack_flash.py
@@ -0,0 +1,637 @@
+"""Flash attention monkey patch for mistral model"""
+# pylint: disable=duplicate-code
+
+import logging
+from typing import List, Optional, Tuple, Union
+
+import torch
+import transformers
+from einops import rearrange
+from flash_attn.bert_padding import pad_input, unpad_input
+from flash_attn.flash_attn_interface import ( # pylint: disable=ungrouped-imports
+ flash_attn_kvpacked_func,
+ flash_attn_varlen_kvpacked_func,
+ flash_attn_varlen_qkvpacked_func,
+)
+from transformers.modeling_outputs import BaseModelOutputWithPast
+from transformers.models.mistral.modeling_mistral import (
+ MistralAttention as OriginalMistralAttention,
+)
+from transformers.models.mistral.modeling_mistral import (
+ MistralDecoderLayer as OriginalMistralDecoderLayer,
+)
+from transformers.models.mistral.modeling_mistral import apply_rotary_pos_emb, repeat_kv
+
+from axolotl.monkeypatch.utils import get_cu_seqlens_from_pos_ids
+
+LOG = logging.getLogger("axolotl.monkeypatch.mistral")
+
+
+def replace_mistral_attn_with_flash_attn(
+ packed: Optional[bool] = False,
+):
+ transformers.models.mistral.modeling_mistral.MistralModel._prepare_decoder_attention_mask = ( # pylint: disable=protected-access
+ _prepare_decoder_attention_mask
+ )
+ transformers.models.mistral.modeling_mistral.MistralAttention.forward = (
+ flashattn_forward
+ )
+ if packed:
+ transformers.models.mistral.modeling_mistral.MistralDecoderLayer = (
+ MistralDecoderLayer
+ )
+ transformers.models.mistral.modeling_mistral.MistralModel.forward = (
+ mistral_model_forward
+ )
+
+
+@torch.jit.script
+def _make_sliding_window_causal_mask(
+ bsz: int,
+ tgt_len: int,
+ dtype: torch.dtype,
+ device: torch.device,
+ past_key_values_length: int = 0,
+ sliding_window: int = 4096,
+):
+ """
+ Make causal mask used for sliding window attention
+ """
+ tensor = torch.full(
+ (tgt_len, tgt_len),
+ fill_value=1,
+ device=device,
+ )
+ mask = torch.tril(tensor, diagonal=0)
+ # make the mask banded to account for sliding window
+ # NOTE: HF implementation is wrong as of 14-10-2023 for torch.triu, needs +1
+ mask = torch.triu(mask, diagonal=-sliding_window + 1)
+ mask = torch.log(mask).to(dtype)
+
+ if past_key_values_length > 0:
+ mask = torch.cat(
+ [
+ torch.zeros(
+ tgt_len, past_key_values_length, dtype=dtype, device=device
+ ),
+ mask,
+ ],
+ dim=-1,
+ )
+ return mask[None, None, :, :].expand(
+ bsz, 1, tgt_len, tgt_len + past_key_values_length
+ )
+
+
+# Disable the transformation of the attention mask in LlamaModel as the flash attention
+# requires the attention mask to be the same as the key_padding_mask
+def _prepare_decoder_attention_mask(
+ self,
+ attention_mask,
+ input_shape,
+ inputs_embeds,
+ past_key_values_length,
+ sliding_window,
+): # pylint: disable=unused-argument
+ # [bsz, seq_len]
+ if attention_mask is None or sliding_window is None:
+ return attention_mask
+
+ # NOTE: attention mask and sliding masks are only broadcastable in certain scenarios.
+ # Without attention_mask.shape[0] == 1, error will trigger after eval loss but only when wandb is enabled.
+ if input_shape[-1] > 1 and attention_mask.shape[0] == 1:
+ sliding_window_mask = _make_sliding_window_causal_mask(
+ bsz=input_shape[0],
+ tgt_len=input_shape[1],
+ dtype=inputs_embeds.dtype,
+ device=inputs_embeds.device,
+ past_key_values_length=past_key_values_length,
+ sliding_window=sliding_window,
+ )
+ attention_mask = attention_mask + sliding_window_mask
+ else:
+ LOG.info("skipping sliding window mask, not broadcastable with attention mask")
+
+ return attention_mask
+
+
+def flashattn_forward(
+ self: OriginalMistralAttention,
+ hidden_states: torch.Tensor,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: bool = False,
+ use_cache: bool = False,
+ cu_seqlens: Optional[torch.Tensor] = None,
+ max_seqlen: Optional[torch.Tensor] = None,
+) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
+ bsz, q_len, _ = hidden_states.size()
+
+ query_states = self.q_proj(hidden_states)
+ key_states = self.k_proj(hidden_states)
+ value_states = self.v_proj(hidden_states)
+
+ query_states = query_states.view(
+ bsz, q_len, self.num_heads, self.head_dim
+ ).transpose(1, 2)
+ key_states = key_states.view(
+ bsz, q_len, self.num_key_value_heads, self.head_dim
+ ).transpose(1, 2)
+ value_states = value_states.view(
+ bsz, q_len, self.num_key_value_heads, self.head_dim
+ ).transpose(1, 2)
+
+ kv_seq_len = key_states.shape[-2]
+ if past_key_value is not None:
+ kv_seq_len += past_key_value[0].shape[-2]
+ cos, sin = self.rotary_emb(value_states, seq_len=kv_seq_len)
+ query_states, key_states = apply_rotary_pos_emb(
+ query_states, key_states, cos, sin, position_ids
+ )
+
+ use_sliding_windows = (
+ getattr(self.config, "sliding_window") is not None
+ and kv_seq_len > self.config.sliding_window
+ )
+
+ if use_sliding_windows:
+ window_size = (self.config.sliding_window, self.config.sliding_window)
+ else:
+ window_size = (-1, -1)
+
+ if past_key_value is not None:
+ # Activate slicing cache only if the config has a value `sliding_windows` attribute
+ if (
+ hasattr(self.config, "sliding_window")
+ and kv_seq_len > self.config.sliding_window
+ ):
+ slicing_tokens = kv_seq_len - self.config.sliding_window
+
+ past_key = past_key_value[0]
+ past_value = past_key_value[1]
+
+ past_key = past_key[:, :, slicing_tokens:, :].contiguous()
+ past_value = past_value[:, :, slicing_tokens:, :].contiguous()
+
+ if past_key.shape[-2] != self.config.sliding_window - 1:
+ raise ValueError(
+ f"past key much have a shape of (`batch_size, num_heads, self.config.sliding_window-1, head_dim`), got"
+ f" {past_key.shape}"
+ )
+
+ past_key_value = (past_key, past_value) if use_cache else None
+
+ if past_key_value is not None:
+ key_states = torch.cat([past_key_value[0], key_states], dim=2)
+ value_states = torch.cat([past_key_value[1], value_states], dim=2)
+
+ past_key_value = (key_states, value_states) if use_cache else None
+
+ # repeat k/v heads if n_kv_heads < n_heads
+ key_states = repeat_kv(key_states, self.num_key_value_groups)
+ value_states = repeat_kv(value_states, self.num_key_value_groups)
+
+ if self.training:
+ # during training q,k,v always have same seqlen
+ assert key_states.shape == query_states.shape
+ is_causal = True
+ else:
+ # turn off FA causal mask after first inference autoregressive iteration
+ # only on first autoregressive step q,k,v have same seqlen
+ is_causal = key_states.shape == query_states.shape
+
+ dropout_rate = 0.0 if not self.training else getattr(self, "attention_dropout", 0.0)
+
+ if cu_seqlens is not None and max_seqlen is not None and cu_seqlens.dim() == 1:
+ # special handling using sample packing
+ qkv = torch.stack(
+ [query_states, key_states, value_states], dim=2
+ ) # [bsz, nh, 3, q_len, hd]
+ qkv = qkv.transpose(1, 3) # [bsz, q_len, 3, nh, hd]
+ qkv = rearrange(qkv, "b s ... -> (b s) ...")
+
+ output = flash_attn_varlen_qkvpacked_func(
+ qkv,
+ cu_seqlens,
+ max_seqlen,
+ dropout_p=dropout_rate,
+ softmax_scale=None,
+ causal=True,
+ window_size=window_size,
+ )
+ output = rearrange(output, "(b s) ... -> b s ...", b=bsz)
+ elif query_states.shape == key_states.shape:
+ query_states = query_states.transpose(1, 2)
+ key_states = key_states.transpose(1, 2)
+ value_states = value_states.transpose(1, 2)
+ qkv_unpad, cu_seqlens_q, max_seqlen_q, _, output_pad_fn = generate_qkv(
+ query_states,
+ key_states,
+ value_states,
+ qkvpacked=True,
+ # We have disabled _prepare_decoder_attention_mask in LlamaModel
+ # the attention_mask should be the same as the key_padding_mask
+ key_padding_mask=attention_mask,
+ query_padding_mask=attention_mask[:, -query_states.size(1) :]
+ if attention_mask is not None
+ else None,
+ )
+ output_unpad = flash_attn_varlen_qkvpacked_func(
+ qkv_unpad,
+ cu_seqlens_q,
+ max_seqlen_q,
+ dropout_p=dropout_rate,
+ softmax_scale=None,
+ causal=is_causal,
+ window_size=window_size,
+ )
+ output = output_pad_fn(output_unpad)
+ else:
+ query_states = query_states.transpose(1, 2)
+ key_states = key_states.transpose(1, 2)
+ value_states = value_states.transpose(1, 2)
+ if attention_mask is None or attention_mask.all().item():
+ output = flash_attn_kvpacked_func(
+ query_states,
+ torch.stack([key_states, value_states], 2),
+ dropout_p=dropout_rate,
+ causal=is_causal,
+ window_size=window_size,
+ )
+ else:
+ ( # pylint: disable=unbalanced-tuple-unpacking
+ q_unpad,
+ kv_unpad,
+ cu_seqlens_q,
+ cu_seqlens_k,
+ max_seqlen_q,
+ max_seqlen_k,
+ _,
+ _,
+ output_pad_fn,
+ ) = generate_qkv(
+ query_states,
+ key_states,
+ value_states,
+ kvpacked=True,
+ key_padding_mask=attention_mask,
+ query_padding_mask=attention_mask[:, -query_states.size(1) :]
+ if attention_mask is not None
+ else None,
+ )
+ if q_unpad.dtype != kv_unpad.dtype:
+ kv_unpad = kv_unpad.to(q_unpad.dtype)
+ output_unpad = flash_attn_varlen_kvpacked_func(
+ q_unpad,
+ kv_unpad,
+ cu_seqlens_q,
+ cu_seqlens_k,
+ max_seqlen_q,
+ max_seqlen_k,
+ dropout_p=dropout_rate,
+ softmax_scale=None,
+ causal=is_causal,
+ window_size=window_size,
+ )
+ output = output_pad_fn(output_unpad)
+
+ attn_output = output
+ if attn_output.size() != (bsz, q_len, self.num_heads, self.head_dim):
+ raise ValueError(
+ f"`attn_output` should be of size {(bsz, q_len, self.num_heads, self.head_dim)}, but is"
+ f" {attn_output.size()}"
+ )
+ attn_output = rearrange(attn_output, "b s h d -> b s (h d)")
+
+ attn_output = self.o_proj(attn_output)
+
+ if not output_attentions:
+ attn_weights = None
+
+ return attn_output, attn_weights, past_key_value
+
+
+# based on https://github.com/Dao-AILab/flash-attention/blob/364a5b/tests/test_flash_attn.py#L38
+def generate_qkv(
+ q,
+ k,
+ v,
+ query_padding_mask=None,
+ key_padding_mask=None,
+ kvpacked=False,
+ qkvpacked=False,
+): # pylint: disable=invalid-name,unnecessary-lambda-assignment
+ """
+ Arguments:
+ q: (batch_size, seqlen_q, nheads, d)
+ k: (batch_size, seqlen_k, nheads_k, d)
+ v: (batch_size, seqlen_k, nheads_k, d)
+ query_padding_mask: (batch_size, seqlen), bool
+ key_padding_mask: (batch_size, seqlen), bool
+ """
+ assert not (kvpacked and qkvpacked)
+ batch_size, seqlen_q, nheads, d = q.shape
+ _, seqlen_k, nheads_k, _ = k.shape
+ assert k.shape == (batch_size, seqlen_k, nheads_k, d)
+ assert v.shape == (batch_size, seqlen_k, nheads_k, d)
+
+ if query_padding_mask is not None:
+ q_unpad, indices_q, cu_seqlens_q, max_seqlen_q = unpad_input(
+ q, query_padding_mask
+ )
+
+ output_pad_fn = lambda output_unpad: pad_input( # noqa: E731
+ output_unpad, indices_q, batch_size, seqlen_q
+ )
+
+ else:
+ q_unpad = rearrange(q, "b s h d -> (b s) h d")
+ cu_seqlens_q = torch.arange(
+ 0,
+ (batch_size + 1) * seqlen_q,
+ step=seqlen_q,
+ dtype=torch.int32,
+ device=q_unpad.device,
+ )
+ max_seqlen_q = seqlen_q
+
+ output_pad_fn = lambda output_unpad: rearrange( # noqa: E731
+ output_unpad, "(b s) h d -> b s h d", b=batch_size
+ )
+
+ if key_padding_mask is not None:
+ k_unpad, _, cu_seqlens_k, max_seqlen_k = unpad_input(k, key_padding_mask)
+ v_unpad, _, _, _ = unpad_input(v, key_padding_mask)
+ else:
+ k_unpad = rearrange(k, "b s h d -> (b s) h d")
+ v_unpad = rearrange(v, "b s h d -> (b s) h d")
+ cu_seqlens_k = torch.arange(
+ 0,
+ (batch_size + 1) * seqlen_k,
+ step=seqlen_k,
+ dtype=torch.int32,
+ device=k_unpad.device,
+ )
+ max_seqlen_k = seqlen_k
+
+ if qkvpacked:
+ assert nheads == nheads_k
+ qkv_unpad = torch.stack([q_unpad, k_unpad, v_unpad], dim=1)
+ qkv = torch.stack([q, k, v], dim=2)
+ return (qkv_unpad, cu_seqlens_q, max_seqlen_q, qkv, output_pad_fn)
+
+ if kvpacked:
+ kv_unpad = torch.stack([k_unpad, v_unpad], dim=1)
+ kv = torch.stack([k, v], dim=2)
+ return (
+ q_unpad,
+ kv_unpad,
+ cu_seqlens_q,
+ cu_seqlens_k,
+ max_seqlen_q,
+ max_seqlen_k,
+ q,
+ kv,
+ output_pad_fn,
+ )
+
+ return (
+ q_unpad,
+ k_unpad,
+ v_unpad,
+ cu_seqlens_q,
+ cu_seqlens_k,
+ max_seqlen_q,
+ max_seqlen_k,
+ q,
+ k,
+ v,
+ output_pad_fn,
+ )
+
+
+def mistral_model_forward(
+ self,
+ input_ids: torch.LongTensor = None,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_values: Optional[List[torch.FloatTensor]] = None,
+ inputs_embeds: Optional[torch.FloatTensor] = None,
+ use_cache: Optional[bool] = None,
+ output_attentions: Optional[bool] = None,
+ output_hidden_states: Optional[bool] = None,
+ return_dict: Optional[bool] = None,
+) -> Union[Tuple, BaseModelOutputWithPast]:
+ output_attentions = (
+ output_attentions
+ if output_attentions is not None
+ else self.config.output_attentions
+ )
+ output_hidden_states = (
+ output_hidden_states
+ if output_hidden_states is not None
+ else self.config.output_hidden_states
+ )
+ use_cache = use_cache if use_cache is not None else self.config.use_cache
+
+ return_dict = (
+ return_dict if return_dict is not None else self.config.use_return_dict
+ )
+
+ # retrieve input_ids and inputs_embeds
+ if input_ids is not None and inputs_embeds is not None:
+ raise ValueError(
+ "You cannot specify both decoder_input_ids and decoder_inputs_embeds at the same time"
+ )
+ if input_ids is not None:
+ batch_size, seq_length = input_ids.shape
+ elif inputs_embeds is not None:
+ batch_size, seq_length, _ = inputs_embeds.shape
+ else:
+ raise ValueError(
+ "You have to specify either decoder_input_ids or decoder_inputs_embeds"
+ )
+
+ seq_length_with_past = seq_length
+ past_key_values_length = 0
+
+ if past_key_values is not None:
+ past_key_values_length = past_key_values[0][0].shape[2]
+ seq_length_with_past = seq_length_with_past + past_key_values_length
+
+ cu_seqlens = None
+ max_seqlen = None
+ if position_ids is None:
+ device = input_ids.device if input_ids is not None else inputs_embeds.device
+ position_ids = torch.arange(
+ past_key_values_length,
+ seq_length + past_key_values_length,
+ dtype=torch.long,
+ device=device,
+ )
+ position_ids = position_ids.unsqueeze(0).view(-1, seq_length)
+ else:
+ position_ids = position_ids.view(-1, seq_length).long()
+ cu_seqlens, max_seqlen = get_cu_seqlens_from_pos_ids(position_ids)
+ cu_seqlens = cu_seqlens.squeeze()
+
+ if inputs_embeds is None:
+ inputs_embeds = self.embed_tokens(input_ids)
+ # embed positions
+ if attention_mask is None:
+ attention_mask = torch.ones(
+ (batch_size, seq_length_with_past),
+ dtype=torch.bool,
+ device=inputs_embeds.device,
+ )
+ attention_mask = (
+ self._prepare_decoder_attention_mask( # pylint: disable=protected-access
+ attention_mask,
+ (batch_size, seq_length),
+ inputs_embeds,
+ past_key_values_length,
+ sliding_window=self.config.sliding_window,
+ )
+ )
+
+ hidden_states = inputs_embeds
+
+ if self.gradient_checkpointing and self.training:
+ if use_cache:
+ transformers.logger.warning_once(
+ "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
+ )
+ use_cache = False
+
+ # decoder layers
+ all_hidden_states = () if output_hidden_states else None
+ all_self_attns = () if output_attentions else None
+ next_decoder_cache = () if use_cache else None
+
+ for idx, decoder_layer in enumerate(self.layers):
+ if output_hidden_states:
+ all_hidden_states += (hidden_states,)
+
+ past_key_value = past_key_values[idx] if past_key_values is not None else None
+
+ if self.gradient_checkpointing and self.training:
+ layer_outputs = (
+ self._gradient_checkpointing_func( # pylint: disable=protected-access
+ decoder_layer.__call__,
+ hidden_states,
+ attention_mask,
+ position_ids,
+ past_key_value,
+ output_attentions,
+ None,
+ cu_seqlens,
+ max_seqlen,
+ )
+ )
+ else:
+ layer_outputs = decoder_layer(
+ hidden_states,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ cu_seqlens=cu_seqlens,
+ max_seqlen=max_seqlen,
+ )
+
+ hidden_states = layer_outputs[0]
+
+ if use_cache:
+ next_decoder_cache += (layer_outputs[2 if output_attentions else 1],)
+
+ if output_attentions:
+ all_self_attns += (layer_outputs[1],)
+
+ hidden_states = self.norm(hidden_states)
+
+ # add hidden states from the last decoder layer
+ if output_hidden_states:
+ all_hidden_states += (hidden_states,)
+
+ next_cache = next_decoder_cache if use_cache else None
+ if not return_dict:
+ return tuple(
+ v
+ for v in [hidden_states, next_cache, all_hidden_states, all_self_attns]
+ if v is not None
+ )
+ return BaseModelOutputWithPast(
+ last_hidden_state=hidden_states,
+ past_key_values=next_cache,
+ hidden_states=all_hidden_states,
+ attentions=all_self_attns,
+ )
+
+
+class MistralDecoderLayer(OriginalMistralDecoderLayer):
+ """
+ patched version of MistralDecoderLayer to pass through the precalculated cu_seqlens
+ """
+
+ def forward(
+ self,
+ hidden_states: torch.Tensor,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: Optional[bool] = False,
+ use_cache: Optional[bool] = False,
+ cu_seqlens: Optional[torch.Tensor] = None,
+ max_seqlen: Optional[torch.Tensor] = None,
+ ) -> Tuple[
+ torch.FloatTensor, Optional[Tuple[torch.FloatTensor, torch.FloatTensor]]
+ ]:
+ """
+ Args:
+ hidden_states (`torch.FloatTensor`): input to the layer of shape `(batch, seq_len, embed_dim)`
+ attention_mask (`torch.FloatTensor`, *optional*): attention mask of size
+ `(batch, 1, tgt_len, src_len)` where padding elements are indicated by very large negative values.
+ output_attentions (`bool`, *optional*):
+ Whether or not to return the attentions tensors of all attention layers. See `attentions` under
+ returned tensors for more detail.
+ use_cache (`bool`, *optional*):
+ If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding
+ (see `past_key_values`).
+ past_key_value (`Tuple(torch.FloatTensor)`, *optional*): cached past key and value projection states
+ cu_seqlens (`torch.Tensor`, *optional*) cumulative sequence len when packing
+ """
+
+ residual = hidden_states
+
+ hidden_states = self.input_layernorm(hidden_states)
+
+ # Self Attention
+ hidden_states, self_attn_weights, present_key_value = self.self_attn(
+ hidden_states=hidden_states,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ cu_seqlens=cu_seqlens,
+ max_seqlen=max_seqlen,
+ )
+ hidden_states = residual + hidden_states
+
+ # Fully Connected
+ residual = hidden_states
+ hidden_states = self.post_attention_layernorm(hidden_states)
+ hidden_states = self.mlp(hidden_states)
+ hidden_states = residual + hidden_states
+
+ outputs = (hidden_states,)
+
+ if output_attentions:
+ outputs += (self_attn_weights,)
+
+ if use_cache:
+ outputs += (present_key_value,)
+
+ return outputs
diff --git a/src/axolotl/monkeypatch/mixtral/__init__.py b/src/axolotl/monkeypatch/mixtral/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..d6ee0ce16b72191ff3cc83066252bfa73af3801a
--- /dev/null
+++ b/src/axolotl/monkeypatch/mixtral/__init__.py
@@ -0,0 +1,50 @@
+"""
+Patches to support multipack for mixtral
+"""
+import torch
+
+
+def patch_mixtral_moe_forward_zero3() -> None:
+ import torch.nn.functional as F
+
+ def mlp_forward(self, hidden_states):
+ current_hidden_states = self.act_fn(self.w1(hidden_states)) * self.w3(
+ hidden_states
+ )
+ current_hidden_states = self.w2(current_hidden_states)
+ return current_hidden_states
+
+ # Ref. https://huggingface.co/deepseek-ai/deepseek-moe-16b-base/blob/main/modeling_deepseek.py
+ def moe_forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
+ batch_size, sequence_length, hidden_dim = hidden_states.shape
+ hidden_states = hidden_states.view(-1, hidden_dim)
+ # router_logits: (batch * sequence_length, n_experts)
+ router_logits = self.gate(hidden_states)
+
+ routing_weights = F.softmax(router_logits, dim=1, dtype=torch.float)
+ topk_weight, topk_idx = torch.topk(
+ routing_weights, self.top_k, dim=-1, sorted=False
+ )
+ topk_weight /= topk_weight.sum(dim=-1, keepdim=True)
+ # we cast back to the input dtype
+ topk_weight = topk_weight.to(hidden_states.dtype)
+
+ hidden_states = hidden_states.repeat_interleave(self.top_k, dim=0)
+ y = torch.empty_like(hidden_states) # pylint: disable=invalid-name
+ flat_topk_idx = topk_idx.view(-1)
+ for i in range(self.num_experts):
+ expert = self.experts[i]
+ y[flat_topk_idx == i] = expert(hidden_states[flat_topk_idx == i])
+ y = ( # pylint: disable=invalid-name
+ y.view(*topk_weight.shape, -1) * topk_weight.unsqueeze(-1)
+ ).sum(dim=1)
+ final_hidden_states = y.reshape(batch_size, sequence_length, hidden_dim)
+ return final_hidden_states, router_logits
+
+ from transformers.models.mixtral.modeling_mixtral import (
+ MixtralBLockSparseTop2MLP,
+ MixtralSparseMoeBlock,
+ )
+
+ MixtralBLockSparseTop2MLP.forward = mlp_forward
+ MixtralSparseMoeBlock.forward = moe_forward
diff --git a/src/axolotl/monkeypatch/mixtral/__pycache__/__init__.cpython-310.pyc b/src/axolotl/monkeypatch/mixtral/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..fd17c1ca05a40f789402c7bfed00425714221a2f
Binary files /dev/null and b/src/axolotl/monkeypatch/mixtral/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/monkeypatch/multipack.py b/src/axolotl/monkeypatch/multipack.py
new file mode 100644
index 0000000000000000000000000000000000000000..c1eb3127d28ab0f12d98fc132c8c499aa135dd48
--- /dev/null
+++ b/src/axolotl/monkeypatch/multipack.py
@@ -0,0 +1,68 @@
+"""multipack patching for v2 of sample packing"""
+import importlib
+
+import transformers
+from accelerate import init_empty_weights
+from transformers import AutoConfig, AutoModelForCausalLM
+from transformers.integrations import is_deepspeed_zero3_enabled
+
+from axolotl.monkeypatch.mixtral import patch_mixtral_moe_forward_zero3
+from axolotl.monkeypatch.utils import get_unpad_data
+
+SUPPORTED_MULTIPACK_MODEL_TYPES = [
+ "mixtral",
+ "qwen2",
+ "qwen2_moe",
+ "falcon",
+ "phi",
+ "gemma",
+ "gemmoe",
+ "starcoder2",
+]
+
+
+def patch_for_multipack(model_type, model_name=None):
+ if model_type == "mixtral":
+ transformers.models.mixtral.modeling_mixtral._get_unpad_data = ( # pylint: disable=protected-access
+ get_unpad_data
+ )
+ if is_deepspeed_zero3_enabled():
+ patch_mixtral_moe_forward_zero3()
+ elif model_type == "qwen2":
+ transformers.models.qwen2.modeling_qwen2._get_unpad_data = ( # pylint: disable=protected-access
+ get_unpad_data
+ )
+ elif model_type == "qwen2_moe":
+ transformers.models.qwen2_moe.modeling_qwen2_moe._get_unpad_data = ( # pylint: disable=protected-access
+ get_unpad_data
+ )
+ elif model_type == "falcon":
+ transformers.models.falcon.modeling_falcon._get_unpad_data = ( # pylint: disable=protected-access
+ get_unpad_data
+ )
+ elif model_type == "phi":
+ transformers.models.phi.modeling_phi._get_unpad_data = ( # pylint: disable=protected-access
+ get_unpad_data
+ )
+ elif model_type == "gemma":
+ transformers.models.gemma.modeling_gemma._get_unpad_data = ( # pylint: disable=protected-access
+ get_unpad_data
+ )
+ elif model_type == "starcoder2":
+ transformers.models.starcoder2.modeling_starcoder2._get_unpad_data = ( # pylint: disable=protected-access
+ get_unpad_data
+ )
+ elif model_type == "gemmoe":
+ patch_remote(model_name, ".configuration_gemmoe", ".modeling_gemmoe")
+ elif model_type == "jamba":
+ patch_remote(model_name, ".configuration_jamba", ".modeling_jamba")
+
+
+def patch_remote(model_name, config_name, modeling_name):
+ model_config = AutoConfig.from_pretrained(model_name, trust_remote_code=True)
+ # we need to load the model here in order for modeling_* to be available
+ with init_empty_weights():
+ AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True)
+ module_name = model_config.__class__.__module__.replace(config_name, modeling_name)
+ modeling_arch = importlib.import_module(module_name)
+ modeling_arch._get_unpad_data = get_unpad_data # pylint: disable=protected-access
diff --git a/src/axolotl/monkeypatch/relora.py b/src/axolotl/monkeypatch/relora.py
new file mode 100644
index 0000000000000000000000000000000000000000..e4352cbe3d5806acee50b85571d8db7d7610a697
--- /dev/null
+++ b/src/axolotl/monkeypatch/relora.py
@@ -0,0 +1,469 @@
+"""Implements the ReLoRA training procedure from https://arxiv.org/abs/2307.05695, minus the initial full fine-tune."""
+import glob
+import json
+import logging
+import os.path
+import shutil
+from functools import partial
+from pathlib import Path
+from typing import Dict, List, Sequence, Union
+
+import bitsandbytes as bnb
+import peft
+import safetensors.torch as st
+import torch
+from huggingface_hub import snapshot_download
+from torch.distributed.optim import ZeroRedundancyOptimizer
+from torch.optim.lr_scheduler import LRScheduler
+from torch.optim.optimizer import Optimizer
+from transformers import (
+ TrainerCallback,
+ TrainerControl,
+ TrainerState,
+ TrainingArguments,
+)
+from transformers.trainer_utils import PREFIX_CHECKPOINT_DIR
+
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.distributed import barrier, is_main_process
+
+LOG = logging.getLogger("axolotl.relora")
+
+
+@torch.no_grad()
+def magnitude_pruning_(tensor, prune_ratio):
+ tensor_magnitude = torch.abs(tensor)
+ threshold = torch.quantile(
+ tensor_magnitude.flatten().to(dtype=torch.float32), prune_ratio
+ ).to(dtype=tensor.dtype)
+
+ mask = tensor_magnitude > threshold
+ tensor.mul_(mask.to(dtype=tensor.dtype))
+
+
+def reset_optimizer(
+ optimizer: torch.optim.Optimizer,
+ *,
+ reset_params: list[str], # where str is the key to a torch.nn.Parameter
+ optimizer_state_keys: list[str],
+ prune_ratio: float = 0.9,
+):
+ pruning_fn = partial(magnitude_pruning_, prune_ratio=prune_ratio)
+ n_zeros = 0
+ n_total = 0
+
+ optimizer_state = optimizer.state
+ if isinstance(optimizer, ZeroRedundancyOptimizer):
+ optimizer_state = optimizer.optim.state
+
+ for param in reset_params:
+ param_state = optimizer_state[param]
+ if len(param_state) == 0: # no state for this param, happens for ZeRo optimizer
+ continue
+ for key in optimizer_state_keys:
+ pruning_fn(
+ param_state[key]
+ ) # pruning fn has to be inplace to keep the same keys in the dict
+ n_total += param_state[key].numel()
+ n_zeros += torch.sum(param_state[key] == 0).item()
+
+ _zeroed = n_zeros / (1e-7 + n_total) * 100
+ LOG.info(f"Percent of optimizer states zeroed: {_zeroed:.2f}")
+ LOG.info(f"absolute n of optimizer states zeroed: {n_zeros}")
+
+
+class ReLoRACallback(TrainerCallback):
+ """Callback to merge LoRA weights into the base model and save full-weight checkpoints"""
+
+ def __init__(self, cfg: DictDefault):
+ self.relora_steps = cfg.relora_steps
+ self.cpu_offload = cfg.relora_cpu_offload
+ self.quantized = cfg.load_in_4bit or cfg.load_in_8bit
+ self.last_full_model = cfg.base_model
+ self.resume_from_checkpoint = cfg.resume_from_checkpoint
+
+ if not os.path.exists(self.last_full_model):
+ self.last_full_model = str(Path(snapshot_download(cfg.base_model)))
+
+ assert os.path.exists(
+ self.last_full_model
+ ), "for ReLORA base_model must be a local path"
+
+ self.num_lora_restarts = 0
+ self.need_full_save = False
+
+ def on_train_begin(
+ self,
+ _args: TrainingArguments,
+ _state: TrainerState,
+ control: TrainerControl,
+ model: peft.LoraModel,
+ **_kwargs,
+ ):
+ if self.resume_from_checkpoint:
+ weight_path = os.path.join(self.resume_from_checkpoint, "relora")
+ if not os.path.exists(weight_path):
+ LOG.warning(
+ "Resuming ReLoRA from checkpoint, but no full-weight save found"
+ )
+ else:
+ LOG.info(f"Loading adjusted base weights from {weight_path}")
+ load_weight_checkpoint(model, weight_path)
+ return control
+
+ def on_step_begin(
+ self,
+ args: TrainingArguments,
+ state: TrainerState,
+ control: TrainerControl,
+ model: peft.LoraModel,
+ optimizer: torch.optim.Optimizer,
+ **_kwargs,
+ ):
+ if state.global_step > 0 and state.global_step % self.relora_steps == 0:
+ checkpoint_folder = os.path.join(
+ args.output_dir,
+ f"{PREFIX_CHECKPOINT_DIR}-{state.global_step}",
+ "relora",
+ )
+
+ if "adam" in args.optim.lower():
+ optimizer_state_keys = ["exp_avg", "exp_avg_sq"]
+ else:
+ raise ValueError(f"Optimizer {args.optim} not supported with ReLoRA")
+
+ lora_params = [
+ n
+ for n, p in model.named_parameters()
+ if p.requires_grad and "lora_" in n
+ ]
+
+ model.save_pretrained(
+ os.path.join(
+ args.output_dir,
+ f"{PREFIX_CHECKPOINT_DIR}-{state.global_step}",
+ "adapter",
+ ),
+ safe_serialization=True,
+ )
+ with torch.no_grad():
+ merge_and_save(
+ model,
+ self.last_full_model,
+ checkpoint_folder,
+ reinit=True,
+ quantized=self.quantized,
+ actually_save=is_main_process(),
+ cpu_offload=self.cpu_offload,
+ )
+ reset_optimizer(
+ optimizer,
+ reset_params=lora_params,
+ optimizer_state_keys=optimizer_state_keys,
+ prune_ratio=args.relora_prune_ratio,
+ )
+
+ if self.quantized:
+ self.last_full_model = checkpoint_folder
+ self.num_lora_restarts += 1
+
+ return control
+
+ def on_save(
+ self,
+ args: TrainingArguments,
+ state: TrainerState,
+ control: TrainerControl,
+ model: peft.LoraModel,
+ **_kwargs,
+ ):
+ checkpoint_folder = os.path.join(
+ args.output_dir, f"{PREFIX_CHECKPOINT_DIR}-{state.global_step}", "relora"
+ )
+ if (
+ state.global_step >= self.relora_steps
+ and state.global_step % self.relora_steps != 0
+ ):
+ if self.quantized:
+ if is_main_process() and self.last_full_model != checkpoint_folder:
+ # ensure the latest full parameter save is in the latest checkpoint
+ # folder, so that automatic pruning of checkpoints does not remove it
+ LOG.info(f"moving last full parameter save to {checkpoint_folder}")
+ os.makedirs(checkpoint_folder, exist_ok=True)
+ chunks = glob.glob(
+ f"{self.last_full_model}/model*.safetensors"
+ ) + glob.glob(f"{self.last_full_model}/model*.index.json")
+ for path in chunks:
+ new_path = os.path.abspath(shutil.move(path, checkpoint_folder))
+ try:
+ os.symlink(new_path, path)
+ except OSError:
+ # probably on windows without permission to symlink
+ pass
+
+ self.last_full_model = checkpoint_folder
+ else:
+ model.model.save_pretrained(checkpoint_folder, safe_serialization=True)
+
+ return control
+
+ def on_log(
+ self,
+ _args: TrainingArguments,
+ _state: TrainerState,
+ control: TrainerControl,
+ logs: Dict[str, float],
+ **_kwargs,
+ ):
+ logs["num_lora_restarts"] = self.num_lora_restarts
+ return control
+
+ def on_train_end(
+ self,
+ args: TrainingArguments,
+ _state: TrainerState,
+ control: TrainerControl,
+ model: peft.LoraModel,
+ **_kwargs,
+ ):
+ if self.quantized:
+ # perform final merge and save
+ with torch.no_grad():
+ merge_and_save(
+ model,
+ self.last_full_model,
+ args.output_dir,
+ reinit=False,
+ quantized=self.quantized,
+ actually_save=is_main_process(),
+ cpu_offload=self.cpu_offload,
+ )
+ # no need to save if unquantized, as finetune.py will call merge_and_unload()
+ return control
+
+
+class ReLoRAScheduler(LRScheduler):
+ """Wraps another scheduler to apply per-lora-restart learning rate warmups."""
+
+ def __init__(
+ self,
+ optimizer: Optimizer,
+ inner_schedule: LRScheduler,
+ relora_steps: int,
+ warmup_steps: int,
+ anneal_steps: int = 1,
+ min_lr_scale: float = 0.001,
+ ) -> None:
+ self.inner_schedule = inner_schedule
+ self.relora_steps = relora_steps
+ self.warmup_steps = warmup_steps
+ self.anneal_steps = anneal_steps
+ self.min_lr_scale = min_lr_scale
+ super().__init__(optimizer, inner_schedule.last_epoch, inner_schedule.verbose)
+
+ def get_lr(self) -> float:
+ self.inner_schedule.last_epoch = self.last_epoch
+
+ original = self.inner_schedule.get_lr()
+ step = self.last_epoch
+
+ if step < self.relora_steps - self.warmup_steps:
+ scale = 1
+ else:
+ per_relora_progress = step % self.relora_steps
+ if per_relora_progress < self.warmup_steps:
+ cycle_t = min(1.0, (per_relora_progress) / self.warmup_steps)
+ elif per_relora_progress > (self.relora_steps - self.anneal_steps):
+ cycle_t = min(
+ 1.0,
+ (self.relora_steps - per_relora_progress) / self.anneal_steps,
+ )
+ else:
+ cycle_t = 1
+ scale = cycle_t * (1 - self.min_lr_scale) + self.min_lr_scale
+
+ if isinstance(original, Sequence):
+ return [lr * scale for lr in original]
+ return original * scale
+
+
+def sharded_paths(path: str, module_names: List[str]) -> Dict[str, str]:
+ model_name = "model.safetensors"
+ if not os.path.exists(str(Path(path) / model_name)) and not os.path.exists(
+ str(Path(path) / f"{model_name}.index.json")
+ ):
+ model_name = "pytorch_model.bin"
+
+ index_path = str(Path(path) / f"{model_name}.index.json")
+ if os.path.exists(index_path):
+ with open(index_path, "r", encoding="utf-8") as file:
+ data = json.load(file)
+ return data["weight_map"]
+ return {(module_name + ".weight"): model_name for module_name in module_names}
+
+
+def lora_delta_weight(layer: peft.tuners.lora.LoraLayer, device) -> torch.Tensor:
+ if isinstance(layer, (peft.tuners.lora.Linear8bitLt, peft.tuners.lora.Linear4bit)):
+ adapter: Union[List[str], str] = layer.active_adapter
+ if isinstance(adapter, list):
+ if len(adapter) > 1:
+ raise ValueError("unhandled relora for multiple adapters")
+ adapter = adapter[0]
+ return (
+ peft.utils.transpose(
+ layer.lora_B[adapter].weight.detach().to(device)
+ @ layer.lora_A[adapter].weight.detach().to(device),
+ getattr(layer, "fan_in_fan_out", False),
+ )
+ * layer.scaling[adapter]
+ )
+
+ raise ValueError("unhandled lora layer type")
+
+
+def find_lora_modules(model: peft.LoraModel) -> Dict[str, peft.tuners.lora.LoraLayer]:
+ modules: Dict[str, peft.tuners.lora.LoraLayer] = {}
+
+ key_list = [key for key, _ in model.model.named_modules() if "lora" not in key]
+ for key in key_list:
+ try:
+ # pylint: disable=protected-access
+ _parent, target, _target_name = peft.utils._get_submodules(model.model, key)
+ except AttributeError:
+ continue
+
+ if isinstance(target, peft.tuners.lora.LoraLayer):
+ modules[key] = target
+
+ return modules
+
+
+def update_weights(
+ target: peft.tuners.lora.LoraLayer, new_weight: torch.Tensor, reinit: bool, device
+):
+ if reinit:
+ for adapter_name in target.lora_A:
+ target.reset_lora_parameters(adapter_name, True)
+ for adapter_name in target.lora_embedding_A:
+ target.reset_lora_parameters(adapter_name, True)
+
+ if isinstance(target, peft.tuners.lora.Linear4bit):
+ # This could be faster, but the quantization of Linear4bit weights occurs
+ # when the module is moved from cpu to gpu. Without meddling *too* deeply in
+ # PEFT's innards or maintaining a duplicate of that codepath, this is good
+ # enough for now.
+ target.weight.quant_state = None
+ target.weight.data = new_weight.cpu()
+ target.to(device)
+ elif isinstance(target, peft.tuners.lora.Linear8bitLt):
+ target.weight.data = (
+ bnb.nn.Int8Params(new_weight, requires_grad=False).to(device).data
+ )
+ else:
+ target.weight.data = new_weight.to(device)
+
+
+def merge_and_save(
+ model: peft.LoraModel,
+ model_src: str,
+ model_dst: str,
+ reinit: bool = False,
+ quantized: bool = False,
+ cpu_offload: bool = False,
+ actually_save: bool = True,
+):
+ modules = find_lora_modules(model)
+
+ if not quantized:
+ for module_name, target in modules.items():
+ active_adapter = target.active_adapter
+ if isinstance(active_adapter, list):
+ active_adapter = active_adapter[0]
+ update = target.get_delta_weight(active_adapter).detach()
+ target.weight.data += update
+
+ if reinit:
+ for adapter_name in target.lora_A:
+ target.reset_lora_parameters(adapter_name, True)
+ for adapter_name in target.lora_embedding_A:
+ target.reset_lora_parameters(adapter_name, True)
+ return
+
+ os.makedirs(model_dst, exist_ok=True)
+ shard_paths = sharded_paths(model_src, modules.keys())
+ out_shard_paths = {}
+
+ unique_shards = list(set(shard_paths.values()))
+ for shard_path in unique_shards:
+ out_tensors = {}
+ if shard_path.endswith(".safetensors"):
+ in_tensors = st.load_file(str(Path(model_src) / shard_path))
+ else:
+ in_tensors = torch.load(Path(model_src) / shard_path)
+ if "state_dict" in in_tensors:
+ in_tensors = in_tensors["state_dict"]
+
+ for module_name, target in modules.items():
+ key = module_name + ".weight"
+ if key not in shard_paths or shard_paths[key] != shard_path:
+ continue
+
+ orig_weight = in_tensors[key]
+ old_dev = target.weight.device
+ math_dev = "cpu" if cpu_offload else old_dev
+
+ delta_weight = lora_delta_weight(target, math_dev)
+ new_weight = orig_weight.to(math_dev) + delta_weight
+ del delta_weight
+
+ if actually_save:
+ out_tensors[key] = new_weight.half().cpu()
+
+ update_weights(target, new_weight, reinit=reinit, device=old_dev)
+
+ if actually_save:
+ out_shard_name = shard_path
+ if out_shard_name.startswith("pytorch_model"):
+ out_shard_name = (
+ out_shard_name.replace("pytorch_model", "model").rstrip(".bin")
+ + ".safetensors"
+ )
+
+ for module_name in in_tensors:
+ if module_name not in out_tensors:
+ out_tensors[module_name] = in_tensors[module_name].half()
+ out_shard_paths[module_name] = out_shard_name
+
+ shard_fn = str(Path(model_dst) / out_shard_name)
+ LOG.info(f"saving tensors to {shard_fn}")
+ st.save_file(out_tensors, shard_fn, metadata={"format": "pt"})
+
+ barrier()
+ del in_tensors
+ del out_tensors
+ torch.cuda.empty_cache()
+
+ if actually_save and len(unique_shards) > 1:
+ with open(
+ str(Path(model_dst, "model.safetensors.index.json")), "w", encoding="utf-8"
+ ) as file:
+ json.dump({"metadata": {}, "weight_map": out_shard_paths}, file)
+
+
+def load_weight_checkpoint(model: peft.LoraModel, checkpoint_path: str):
+ modules = find_lora_modules(model)
+ shard_paths = sharded_paths(checkpoint_path, modules.keys())
+ unique_shards = list(set(shard_paths.values()))
+
+ for shard_path in unique_shards:
+ tensors = st.load_file(os.path.join(checkpoint_path, shard_path))
+
+ for module_name, target in modules.items():
+ key = module_name + ".weight"
+ if key not in shard_paths or shard_paths[key] != shard_path:
+ continue
+
+ new_weight = tensors[key]
+ update_weights(
+ target, new_weight, reinit=False, device=target.weight.device
+ )
diff --git a/src/axolotl/monkeypatch/stablelm_attn_hijack_flash.py b/src/axolotl/monkeypatch/stablelm_attn_hijack_flash.py
new file mode 100644
index 0000000000000000000000000000000000000000..0269f90157b12c46a460545842d69da65afc394a
--- /dev/null
+++ b/src/axolotl/monkeypatch/stablelm_attn_hijack_flash.py
@@ -0,0 +1,415 @@
+# coding=utf-8
+# Copyright 2023 Stability AI, EleutherAI, and The HuggingFace Inc. team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+#
+# This code is based off the following work:
+# https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py
+# https://github.com/huggingface/transformers/blob/main/src/transformers/models/gpt_neox/modeling_gpt_neox.py
+""" PyTorch StableLM Epoch model. """
+import importlib
+import math
+from typing import Optional, Tuple, Union
+
+import torch
+import torch.utils.checkpoint
+from accelerate import init_empty_weights
+from einops import rearrange
+from flash_attn.flash_attn_interface import ( # pylint: disable=ungrouped-imports
+ flash_attn_varlen_qkvpacked_func,
+)
+from torch import nn
+from transformers import AutoConfig, AutoModelForCausalLM
+from transformers.modeling_outputs import BaseModelOutputWithPast
+from transformers.utils import logging
+
+from axolotl.monkeypatch.utils import get_cu_seqlens_from_pos_ids
+
+logger = logging.get_logger(__name__)
+
+
+def replace_stablelm_attn_with_flash_attn(model_name="stabilityai/stablelm-3b-4e1t"):
+ # this is a wonky hack to get the remotely loaded module
+ model_config = AutoConfig.from_pretrained(model_name, trust_remote_code=True)
+ # we need to load the model here in order for modeling_stablelm_epoch to be available
+ with init_empty_weights():
+ AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True)
+ module_name = model_config.__class__.__module__.replace(
+ ".configuration_stablelm_epoch", ".modeling_stablelm_epoch"
+ )
+ modeling_stablelm = importlib.import_module(module_name)
+ modeling_stablelm.Attention.forward = ( # pylint: disable=protected-access
+ flashattn_attn
+ )
+ modeling_stablelm.StableLMEpochModel.forward = ( # pylint: disable=protected-access
+ stablelm_model_forward
+ )
+ modeling_stablelm.DecoderLayer.forward = ( # pylint: disable=protected-access
+ decoder_layer_forward
+ )
+
+
+def rotate_half(x: torch.Tensor):
+ """Rotates half the hidden dims of the input."""
+ # pylint: disable=invalid-name
+ x1, x2 = torch.chunk(x, 2, dim=-1)
+ return torch.cat((-x2, x1), dim=-1)
+
+
+def apply_rotary_pos_emb(q, k, cos, sin, position_ids):
+ # The first two dimensions of cos and sin are always 1, so we can `squeeze` them.
+ # pylint: disable=invalid-name
+ cos = cos.squeeze(1).squeeze(0) # [seq_len, dim]
+ sin = sin.squeeze(1).squeeze(0) # [seq_len, dim]
+ cos = cos[position_ids].unsqueeze(1) # [batch_size, 1, seq_len, dim]
+ sin = sin[position_ids].unsqueeze(1) # [batch_size, 1, seq_len, dim]
+ q_embed = (q * cos) + (rotate_half(q) * sin)
+ k_embed = (k * cos) + (rotate_half(k) * sin)
+ return q_embed, k_embed
+
+
+def repeat_kv(hidden_states: torch.Tensor, n_rep: int) -> torch.Tensor:
+ """
+ This is the equivalent of torch.repeat_interleave(x, dim=1, repeats=n_rep). The hidden states go from (batch,
+ num_key_value_heads, seqlen, head_dim) to (batch, num_attention_heads, seqlen, head_dim)
+ """
+ batch, num_key_value_heads, slen, head_dim = hidden_states.shape
+ if n_rep == 1:
+ return hidden_states
+ hidden_states = hidden_states[:, :, None, :, :].expand(
+ batch, num_key_value_heads, n_rep, slen, head_dim
+ )
+ return hidden_states.reshape(batch, num_key_value_heads * n_rep, slen, head_dim)
+
+
+def flashattn_attn(
+ self,
+ hidden_states: torch.FloatTensor,
+ attention_mask: torch.FloatTensor,
+ position_ids: torch.LongTensor,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: Optional[bool] = False, # pylint: disable=unused-argument
+ use_cache: Optional[bool] = False,
+ cu_seqlens: Optional[torch.Tensor] = None,
+ max_seqlen: Optional[torch.Tensor] = None,
+) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
+ bsz, q_len, _ = hidden_states.size()
+
+ query_states = self.q_proj(hidden_states)
+ key_states = self.k_proj(hidden_states)
+ value_states = self.v_proj(hidden_states)
+
+ query_states = query_states.view(
+ bsz, q_len, self.num_heads, self.head_dim
+ ).transpose(1, 2)
+ key_states = key_states.view(
+ bsz, q_len, self.num_key_value_heads, self.head_dim
+ ).transpose(1, 2)
+ value_states = value_states.view(
+ bsz, q_len, self.num_key_value_heads, self.head_dim
+ ).transpose(1, 2)
+
+ query_rot = query_states[..., : self.rotary_ndims]
+ query_pass = query_states[..., self.rotary_ndims :]
+ key_rot = key_states[..., : self.rotary_ndims]
+ key_pass = key_states[..., self.rotary_ndims :]
+
+ kv_seq_len = key_states.shape[-2]
+ if past_key_value is not None:
+ kv_seq_len += past_key_value[0].shape[-2]
+ cos, sin = self.rotary_emb(value_states, seq_len=kv_seq_len)
+ query_states, key_states = apply_rotary_pos_emb(
+ query_rot, key_rot, cos, sin, position_ids
+ )
+
+ # [batch_size, num_heads, seq_len, head_dim]
+ query_states = torch.cat((query_states, query_pass), dim=-1)
+ key_states = torch.cat((key_states, key_pass), dim=-1)
+
+ if past_key_value is not None:
+ # Reuse k, v, self_attention
+ key_states = torch.cat((past_key_value[0], key_states), dim=2)
+ value_states = torch.cat((past_key_value[1], value_states), dim=2)
+
+ past_key_value = (key_states, value_states) if use_cache else None
+
+ # Repeat k/v heads if n_kv_heads < n_heads
+ key_states = repeat_kv(key_states, self.num_key_value_groups)
+ value_states = repeat_kv(value_states, self.num_key_value_groups)
+
+ if cu_seqlens is not None and max_seqlen is not None and cu_seqlens.dim() == 1:
+ # special handling using sample packing
+ qkv = torch.stack(
+ [query_states, key_states, value_states], dim=2
+ ) # [bsz, nh, 3, q_len, hd]
+ qkv = qkv.transpose(1, 3) # [bsz, q_len, 3, nh, hd]
+ qkv = rearrange(qkv, "b s ... -> (b s) ...")
+ softmax_scale = None
+
+ output = flash_attn_varlen_qkvpacked_func(
+ qkv, cu_seqlens, max_seqlen, 0.0, softmax_scale=softmax_scale, causal=True
+ )
+
+ attn_output = rearrange(output, "(b s) ... -> b s ...", b=bsz)
+ attn_output = rearrange(attn_output, "b s h d -> b s (h d)")
+ else:
+ attn_weights = torch.matmul(
+ query_states, key_states.transpose(2, 3)
+ ) / math.sqrt(self.head_dim)
+
+ if attn_weights.size() != (bsz, self.num_heads, q_len, kv_seq_len):
+ raise ValueError(
+ f"Attention weights should be of size {(bsz, self.num_heads, q_len, kv_seq_len)}, but is"
+ f" {attn_weights.size()}"
+ )
+
+ if attention_mask is not None:
+ if attention_mask.size() != (bsz, 1, q_len, kv_seq_len):
+ raise ValueError(
+ f"Attention mask should be of size {(bsz, 1, q_len, kv_seq_len)}, but is {attention_mask.size()}"
+ )
+ attn_weights = attn_weights + attention_mask
+
+ # Upcast attention to fp32
+ attn_weights = nn.functional.softmax(
+ attn_weights, dim=-1, dtype=torch.float32
+ ).to(query_states.dtype)
+ attn_output = torch.matmul(attn_weights, value_states)
+
+ if attn_output.size() != (bsz, self.num_heads, q_len, self.head_dim):
+ raise ValueError(
+ f"`attn_output` should be of size {(bsz, self.num_heads, q_len, self.head_dim)}, but is"
+ f" {attn_output.size()}"
+ )
+
+ # Merge heads
+ attn_output = attn_output.transpose(1, 2).contiguous()
+ attn_output = attn_output.reshape(bsz, q_len, self.hidden_size)
+
+ # Final linear projection
+ attn_output = self.o_proj(attn_output)
+
+ return attn_output, None, past_key_value
+
+
+def decoder_layer_forward(
+ self,
+ hidden_states: Optional[torch.FloatTensor],
+ attention_mask: Optional[torch.FloatTensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: Optional[bool] = False,
+ use_cache: Optional[bool] = False,
+ cu_seqlens: Optional[torch.Tensor] = None,
+ max_seqlen: Optional[torch.Tensor] = None,
+) -> Union[
+ Tuple[torch.Tensor], Optional[Tuple[torch.Tensor, Tuple[torch.FloatTensor, ...]]]
+]:
+ # pylint: disable=duplicate-code
+ residual = hidden_states
+
+ hidden_states = self.input_layernorm(hidden_states)
+
+ # Self Attention
+ hidden_states, self_attn_weights, present_key_value = self.self_attn(
+ hidden_states=hidden_states,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ cu_seqlens=cu_seqlens,
+ max_seqlen=max_seqlen,
+ )
+ hidden_states = residual + hidden_states
+
+ # Fully Connected
+ residual = hidden_states
+ hidden_states = self.post_attention_layernorm(hidden_states)
+ hidden_states = self.mlp(hidden_states)
+ hidden_states = residual + hidden_states
+
+ outputs = (hidden_states,)
+
+ if output_attentions:
+ outputs += (self_attn_weights,)
+
+ if use_cache:
+ outputs += (present_key_value,)
+
+ return outputs
+
+
+def stablelm_model_forward(
+ self,
+ input_ids: Optional[torch.LongTensor] = None,
+ attention_mask: Optional[torch.FloatTensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_values: Optional[Tuple[Tuple[torch.FloatTensor]]] = None,
+ inputs_embeds: Optional[torch.FloatTensor] = None,
+ use_cache: Optional[bool] = None,
+ output_attentions: Optional[bool] = None,
+ output_hidden_states: Optional[bool] = None,
+ return_dict: Optional[bool] = None,
+) -> Union[Tuple, BaseModelOutputWithPast]:
+ # pylint: disable=duplicate-code
+ output_attentions = (
+ output_attentions
+ if output_attentions is not None
+ else self.config.output_attentions
+ )
+ output_hidden_states = (
+ output_hidden_states
+ if output_hidden_states is not None
+ else self.config.output_hidden_states
+ )
+ use_cache = use_cache if use_cache is not None else self.config.use_cache
+
+ return_dict = (
+ return_dict if return_dict is not None else self.config.use_return_dict
+ )
+
+ # Retrieve input_ids and inputs_embeds
+ if input_ids is not None and inputs_embeds is not None:
+ raise ValueError(
+ "You cannot specify both decoder_input_ids and decoder_inputs_embeds at the same time"
+ )
+ if input_ids is not None:
+ batch_size, seq_length = input_ids.shape
+ elif inputs_embeds is not None:
+ batch_size, seq_length, _ = inputs_embeds.shape
+ else:
+ raise ValueError(
+ "You have to specify either decoder_input_ids or decoder_inputs_embeds"
+ )
+
+ seq_length_with_past = seq_length
+ past_key_values_length = 0
+
+ if past_key_values is not None:
+ past_key_values_length = past_key_values[0][0].shape[2]
+ seq_length_with_past = seq_length_with_past + past_key_values_length
+
+ cu_seqlens = None
+ max_seqlen = None
+ if position_ids is None:
+ device = input_ids.device if input_ids is not None else inputs_embeds.device
+ position_ids = torch.arange(
+ past_key_values_length,
+ seq_length + past_key_values_length,
+ dtype=torch.long,
+ device=device,
+ )
+ position_ids = position_ids.unsqueeze(0).view(-1, seq_length)
+ else:
+ position_ids = position_ids.view(-1, seq_length).long()
+ cu_seqlens, max_seqlen = get_cu_seqlens_from_pos_ids(position_ids)
+ cu_seqlens = cu_seqlens.squeeze()
+
+ if inputs_embeds is None:
+ inputs_embeds = self.embed_tokens(input_ids)
+ # Embed positions
+ if attention_mask is None:
+ attention_mask = torch.ones(
+ (batch_size, seq_length_with_past),
+ dtype=torch.bool,
+ device=inputs_embeds.device,
+ )
+ attention_mask = (
+ self._prepare_decoder_attention_mask( # pylint: disable=protected-access
+ attention_mask,
+ (batch_size, seq_length),
+ inputs_embeds,
+ past_key_values_length,
+ )
+ )
+
+ hidden_states = inputs_embeds
+
+ if self.gradient_checkpointing and self.training:
+ if use_cache:
+ logger.warning(
+ "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..."
+ )
+ use_cache = False
+
+ # Decoder layers
+ all_hidden_states = () if output_hidden_states else None
+ all_self_attns = () if output_attentions else None
+ next_decoder_cache = () if use_cache else None
+
+ for idx, decoder_layer in enumerate(self.layers):
+ if output_hidden_states:
+ all_hidden_states += (hidden_states,)
+
+ past_key_value = past_key_values[idx] if past_key_values is not None else None
+
+ if self.gradient_checkpointing and self.training:
+
+ def create_custom_forward(module):
+ def custom_forward(*inputs):
+ # None for past_key_value
+ return module(*inputs)
+
+ return custom_forward
+
+ layer_outputs = torch.utils.checkpoint.checkpoint(
+ create_custom_forward(decoder_layer),
+ hidden_states,
+ attention_mask,
+ position_ids,
+ past_key_value,
+ output_attentions,
+ None,
+ cu_seqlens,
+ max_seqlen,
+ )
+ else:
+ layer_outputs = decoder_layer(
+ hidden_states,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ cu_seqlens=cu_seqlens,
+ max_seqlen=max_seqlen,
+ )
+
+ hidden_states = layer_outputs[0]
+
+ if use_cache:
+ next_decoder_cache += (layer_outputs[2 if output_attentions else 1],)
+
+ if output_attentions:
+ all_self_attns += (layer_outputs[1],)
+
+ hidden_states = self.norm(hidden_states)
+
+ # Add hidden states from the last decoder layer
+ if output_hidden_states:
+ all_hidden_states += (hidden_states,)
+
+ next_cache = next_decoder_cache if use_cache else None
+ if not return_dict:
+ return tuple(
+ v
+ for v in [hidden_states, next_cache, all_hidden_states, all_self_attns]
+ if v is not None
+ )
+ return BaseModelOutputWithPast(
+ last_hidden_state=hidden_states,
+ past_key_values=next_cache,
+ hidden_states=all_hidden_states,
+ attentions=all_self_attns,
+ )
diff --git a/src/axolotl/monkeypatch/unsloth_.py b/src/axolotl/monkeypatch/unsloth_.py
new file mode 100644
index 0000000000000000000000000000000000000000..753f44037401ee4138a6fa65811480ded6819e96
--- /dev/null
+++ b/src/axolotl/monkeypatch/unsloth_.py
@@ -0,0 +1,267 @@
+"""module for patching with unsloth optimizations"""
+
+import inspect
+import logging
+import re
+import types
+from typing import Tuple
+
+from peft import PeftModelForCausalLM
+from transformers.models.llama.modeling_llama import (
+ LlamaFlashAttention2,
+ LlamaForCausalLM,
+)
+
+LOG = logging.getLogger("axolotl.monkeypatch.unsloth")
+
+ORIGINAL_CEL_CODE = """ if labels is not None:
+ # Shift so that tokens < n predict n
+ shift_logits = logits[..., :-1, :].contiguous()
+ shift_labels = labels[..., 1:].contiguous()
+ # Flatten the tokens
+ loss_fct = CrossEntropyLoss()
+ shift_logits = shift_logits.view(-1, self.config.vocab_size)
+ shift_labels = shift_labels.view(-1)
+ # Enable model parallelism
+ shift_labels = shift_labels.to(shift_logits.device)
+ loss = loss_fct(shift_logits, shift_labels)
+"""
+
+PATCHED_CEL_CODE = """ if labels is not None:
+ shift_logits = logits[..., :-1, :].contiguous()
+ shift_labels = labels[..., 1:].contiguous()
+ loss = fast_cross_entropy_loss(
+ logits = shift_logits,
+ labels = shift_labels,
+ )
+"""
+
+ORIGINAL_QKV_CODE = """
+ query_states = self.q_proj(hidden_states)
+ key_states = self.k_proj(hidden_states)
+ value_states = self.v_proj(hidden_states)
+""".lstrip(
+ "\n"
+)
+
+PATCHED_QKV_CODE = """
+ query_states, key_states, value_states = self.apply_qkv(self, hidden_states)
+""".lstrip(
+ "\n"
+)
+
+ORIGINAL_O_CODE = """
+ attn_output = self.o_proj(attn_output)
+""".lstrip(
+ "\n"
+)
+
+PATCHED_O_CODE = """
+ attn_output = self.apply_o(self, attn_output)
+""".lstrip(
+ "\n"
+)
+
+
+def original_apply_qkv(self, hidden_states):
+ query_states = self.q_proj(hidden_states)
+ key_states = self.k_proj(hidden_states)
+ value_states = self.v_proj(hidden_states)
+ return query_states, key_states, value_states
+
+
+def original_apply_o(self, hidden_states):
+ attn_output = self.o_proj(hidden_states)
+ return attn_output
+
+
+def get_forward_code() -> str:
+ forward = inspect.getsource(LlamaForCausalLM.forward)
+ return forward
+
+
+def test_cel_is_patchable() -> bool:
+ forward = get_forward_code()
+ return ORIGINAL_CEL_CODE in forward
+
+
+def get_self_attn_code() -> str:
+ forward = inspect.getsource(LlamaFlashAttention2.forward)
+ return forward
+
+
+def test_self_attn_is_patchable() -> bool:
+ qkv = get_self_attn_code()
+ return ORIGINAL_QKV_CODE in qkv and ORIGINAL_QKV_CODE in qkv
+
+
+def integrate_cross_entropy_loss_patch():
+ forward = get_forward_code()
+ LlamaForCausalLM._original_forward = forward # pylint: disable=protected-access
+ forward, _ = detab_code(forward)
+ assert ORIGINAL_CEL_CODE in forward, "Original forward code not found"
+
+ forward = forward.replace(
+ "@add_start_docstrings_to_model_forward(LLAMA_INPUTS_DOCSTRING)", ""
+ )
+ forward = forward.replace(
+ "@replace_return_docstrings(output_type=CausalLMOutputWithPast, config_class=_CONFIG_FOR_DOC)",
+ "",
+ )
+ forward = forward.replace(ORIGINAL_CEL_CODE, PATCHED_CEL_CODE)
+ forward = forward.replace(
+ "def forward(",
+ "def fast_cross_entropy_loss_forward(",
+ 1,
+ )
+
+ # load imports necessary
+ import transformers.models.llama.modeling_llama
+
+ items_to_import = []
+ for item in dir(transformers.models.llama.modeling_llama):
+ if item in forward:
+ items_to_import.append(item)
+
+ exec( # pylint: disable=exec-used # nosec B102
+ "from unsloth.kernels.cross_entropy_loss import fast_cross_entropy_loss",
+ globals(),
+ )
+
+ exec( # pylint: disable=exec-used # nosec B102
+ "from transformers.models.llama.modeling_llama import ("
+ + ", ".join(x for x in items_to_import)
+ + ")",
+ globals(),
+ )
+ exec(forward, globals()) # pylint: disable=exec-used # nosec B102
+ print("patching unsloth fast_cross_entropy_loss")
+ LlamaForCausalLM.forward = fast_cross_entropy_loss_forward # pylint: disable=undefined-variable # noqa: F821
+
+
+def detab_code(code: str) -> Tuple[str, str]:
+ spaces = re.match(r"([\s\t]{1,})", code).group(0)
+ code = re.sub(r"^" + spaces, "", code, flags=re.MULTILINE)
+ return code, spaces
+
+
+def patch_self_attn_lora():
+ self_attn_forward = get_self_attn_code()
+ LlamaFlashAttention2._original_forward = ( # pylint: disable=protected-access
+ self_attn_forward
+ )
+ self_attn_forward, _ = detab_code(self_attn_forward)
+ assert ORIGINAL_QKV_CODE in self_attn_forward, "Original qkv code not found"
+ assert ORIGINAL_O_CODE in self_attn_forward, "Original o code not found"
+
+ self_attn_forward = self_attn_forward.replace(ORIGINAL_QKV_CODE, PATCHED_QKV_CODE)
+ self_attn_forward = self_attn_forward.replace(ORIGINAL_O_CODE, PATCHED_O_CODE)
+ self_attn_forward = self_attn_forward.replace(
+ "def forward(",
+ "def unsloth_attn_forward(",
+ 1,
+ )
+
+ # load imports necessary
+ import transformers.models.llama.modeling_llama
+
+ items_to_import = []
+ for item in dir(transformers.models.llama.modeling_llama):
+ if item in self_attn_forward:
+ items_to_import.append(item)
+
+ exec( # pylint: disable=exec-used # nosec B102
+ "from transformers.models.llama.modeling_llama import ("
+ + ", ".join(x for x in items_to_import)
+ + ")",
+ globals(),
+ )
+ exec(self_attn_forward, globals()) # pylint: disable=exec-used # nosec B102
+ print("patching unsloth attn lora")
+ LlamaFlashAttention2.forward = (
+ unsloth_attn_forward # pylint: disable=undefined-variable # noqa: F821
+ )
+
+
+def integrate_lora_mlp_patch(peft_model: PeftModelForCausalLM):
+ if peft_model.base_model.config.model_type in ["llama", "mistral"]:
+ from unsloth.kernels import apply_lora_mlp_swiglu
+
+ apply_lora_mlp = apply_lora_mlp_swiglu
+ elif peft_model.base_model.config.model_type == "gemma":
+ from unsloth.kernels import apply_lora_mlp_geglu_approx
+
+ apply_lora_mlp = apply_lora_mlp_geglu_approx
+ else:
+ raise NotImplementedError(
+ f"Model type {peft_model.base_model.config.model_type} not supported"
+ )
+
+ for idx, layer in enumerate(peft_model.model.model.layers):
+ layer_modules = [
+ getattr(layer.mlp, linear_proj)
+ for linear_proj in ["gate_proj", "up_proj", "down_proj"]
+ ]
+ is_mlp_lora = all(hasattr(module, "lora_A") for module in layer_modules)
+ mlp_no_bias = all(
+ getattr(module, "base_layer", module).bias is None
+ for module in layer_modules
+ )
+ mlp_not_dora = all(
+ getattr(module, "lora_magnitude_vector", None) is None
+ for module in layer_modules
+ )
+
+ if is_mlp_lora and mlp_no_bias and mlp_not_dora:
+ layer.mlp.forward = types.MethodType(apply_lora_mlp, layer.mlp)
+ else:
+ logging.warning("unable to apply unsloth lora mlp patch to layer %d", idx)
+
+
+def integrate_lora_patch(peft_model: PeftModelForCausalLM, cfg):
+ from unsloth.kernels import apply_lora_o, apply_lora_qkv
+
+ for idx, layer in enumerate(peft_model.model.model.layers):
+ if cfg.unsloth_lora_qkv:
+ layer_modules = [
+ getattr(layer.self_attn, linear_proj)
+ for linear_proj in ["q_proj", "k_proj", "v_proj"]
+ ]
+ is_qkv_lora = all(hasattr(module, "lora_A") for module in layer_modules)
+ qkv_no_bias = all(
+ getattr(module, "base_layer", module).bias is None
+ for module in layer_modules
+ )
+ qkv_not_dora = all(
+ getattr(module, "lora_magnitude_vector", None) is None
+ for module in layer_modules
+ )
+
+ if is_qkv_lora and qkv_no_bias and qkv_not_dora:
+ layer.self_attn.apply_qkv = apply_lora_qkv
+ else:
+ layer.self_attn.apply_qkv = original_apply_qkv
+ logging.warning(
+ "unable to apply unsloth lora qkv patch to layer %d", idx
+ )
+ if cfg.unsloth_lora_o:
+ layer_modules = [
+ getattr(layer.self_attn, linear_proj) for linear_proj in ["o_proj"]
+ ]
+ is_o_lora = all(hasattr(module, "lora_A") for module in layer_modules)
+ o_no_bias = all(
+ getattr(module, "base_layer", module).bias is None
+ for module in layer_modules
+ )
+ o_not_dora = all(
+ getattr(module, "lora_magnitude_vector", None) is None
+ for module in layer_modules
+ )
+
+ if is_o_lora and o_no_bias and o_not_dora:
+ layer.self_attn.apply_o = apply_lora_o
+ else:
+ layer.self_attn.apply_o = original_apply_o
+ logging.warning(
+ "unable to apply unsloth lora o_proj patch to layer %d", idx
+ )
diff --git a/src/axolotl/monkeypatch/utils.py b/src/axolotl/monkeypatch/utils.py
new file mode 100644
index 0000000000000000000000000000000000000000..e43c58650a9143163b7461c228d0b320cd2fe4bc
--- /dev/null
+++ b/src/axolotl/monkeypatch/utils.py
@@ -0,0 +1,227 @@
+"""
+Shared utils for the monkeypatches
+"""
+from typing import Optional
+
+import torch
+import torch.nn.functional as F
+from transformers.modeling_attn_mask_utils import (
+ _prepare_4d_causal_attention_mask,
+ _prepare_4d_causal_attention_mask_for_sdpa,
+)
+from transformers.utils import is_torch_bf16_gpu_available
+
+
+@torch.jit.script
+def get_max_seqlen_in_batch(attention_mask: torch.Tensor) -> torch.Tensor:
+ max_num = int(torch.max(attention_mask).item())
+ batch_size, _ = attention_mask.shape
+ counts = torch.zeros((batch_size, max_num), dtype=torch.int32)
+
+ for i in range(1, max_num + 1):
+ mask = attention_mask == i
+ counts[:, i - 1] = torch.sum(mask, dim=-1).to(dtype=torch.int32)
+
+ result = counts.flatten()
+ nonzero_indices = torch.nonzero(result).squeeze(-1)
+ return result[nonzero_indices]
+
+
+@torch.jit.script
+def get_unpad_data(attention_mask: torch.Tensor):
+ device = attention_mask.device
+ seqlens_in_batch = get_max_seqlen_in_batch(attention_mask)
+ indices = torch.nonzero(attention_mask.flatten()).flatten()
+ max_seqlen_in_batch = seqlens_in_batch.max().item()
+ cu_seqlens = (
+ F.pad(torch.cumsum(seqlens_in_batch, dim=0, dtype=torch.int32), (1, 0))
+ .to(device=device)
+ .detach()
+ )
+ return (
+ indices,
+ cu_seqlens,
+ max_seqlen_in_batch,
+ )
+
+
+def get_cu_seqlens(attn_mask):
+ """generate a cumulative sequence length mask for flash attention using attn mask"""
+ if len(attn_mask.shape) == 1:
+ attn_mask = attn_mask.unsqueeze(0)
+
+ device = attn_mask.device
+ results = []
+ max_seq_lens = []
+
+ for row in attn_mask:
+ # Exclude zeros to avoid adding their positions to the mask
+ t_non_zeros = row[row != 0]
+ # Find where the sequence number changes (including the first position)
+ seq_change = torch.cat(
+ [
+ torch.tensor([1], dtype=torch.int32, device=device),
+ t_non_zeros[1:] != t_non_zeros[:-1],
+ ]
+ )
+ # Get the indices where the sequence changes
+ change_indices = torch.cat(
+ [
+ (seq_change == 1).nonzero(as_tuple=True)[0],
+ torch.tensor([len(t_non_zeros)], dtype=torch.int32, device=device),
+ ]
+ )
+ # Calculate the sequence lengths
+ seq_lengths = change_indices[1:] - change_indices[:-1]
+ # Calculate the length of the final sequence or padding
+ final_seq_length = len(row) - change_indices[-1]
+ # Append the length of the final sequence or padding to seq_lengths
+ if final_seq_length.item():
+ seq_lengths = torch.cat(
+ [
+ seq_lengths,
+ torch.tensor(
+ [final_seq_length.item()], dtype=torch.int32, device=device
+ ),
+ ]
+ )
+ # Calculate the cumulative sequence lengths
+ cu_seqlens = torch.cat(
+ [torch.tensor([0], dtype=torch.int32, device=device), seq_lengths.cumsum(0)]
+ )
+ max_seq_len = (cu_seqlens[1:] - cu_seqlens[:-1]).max()
+ results.append(cu_seqlens)
+ max_seq_lens.append(max_seq_len)
+
+ return torch.stack(results).to(dtype=torch.int32), torch.stack(max_seq_lens)
+
+
+def get_cu_seqlens_from_pos_ids(position_ids):
+ """generate a cumulative sequence length mask for flash attention using pos ids"""
+ if len(position_ids.shape) == 1:
+ position_ids = position_ids.unsqueeze(0)
+
+ device = position_ids.device
+ results = []
+ max_seq_lens = []
+
+ for row in position_ids:
+ # Count the number of consecutive zeros from the right side
+ padding_length = (row == 0).int().flip(dims=[0]).cumprod(dim=0).sum().item()
+
+ # Adjust the row to exclude padding
+ adjusted_row = row[:-padding_length] if padding_length else row.clone()
+
+ # Find where the position resets to 0 (indicating a new sequence)
+ seq_starts = torch.cat(
+ [
+ torch.tensor([True], dtype=torch.bool, device=device),
+ adjusted_row[1:] == 0,
+ ]
+ )
+ # Get the indices where the sequence starts
+ start_indices = torch.cat(
+ [
+ torch.nonzero(seq_starts).unbind(dim=1)[0],
+ torch.tensor([len(adjusted_row)], dtype=torch.int32, device=device),
+ ]
+ )
+ # Calculate the sequence lengths
+ seq_lengths = start_indices[1:] - start_indices[:-1]
+ # Calculate the cumulative sequence lengths
+ cu_seqlens = torch.cat(
+ [torch.tensor([0], dtype=torch.int32, device=device), seq_lengths.cumsum(0)]
+ )
+ # Append the padding length to the cumulative sequence lengths
+ if padding_length:
+ cu_seqlens = torch.cat(
+ [cu_seqlens, torch.tensor([len(row)], dtype=torch.int32, device=device)]
+ )
+ max_seq_len = (cu_seqlens[1:] - cu_seqlens[:-1]).max()
+ results.append(cu_seqlens)
+ max_seq_lens.append(max_seq_len)
+
+ # Find the maximum value across all tensors
+ max_value = max(t.max() for t in results)
+
+ # Find the length of the longest tensor
+ max_length = max(t.size(0) for t in results)
+
+ # Pad each tensor to the same length and collect them in a list
+ padded_results = [
+ F.pad(t, (0, max_length - t.size(0)), "constant", max_value) for t in results
+ ]
+
+ return torch.stack(padded_results).to(dtype=torch.int32), torch.stack(max_seq_lens)
+
+
+def set_module_name(model, name, value):
+ if "." in name:
+ parent_name = name.rsplit(".", 1)[0]
+ child_name = name[len(parent_name) + 1 :]
+ parent = model.get_submodule(parent_name)
+ else:
+ parent_name = ""
+ parent = model
+ child_name = name
+
+ setattr(parent, child_name, value)
+
+
+def mask_2d_to_4d(
+ mask: torch.Tensor, dtype: torch.dtype, tgt_len: Optional[int] = None
+):
+ """
+ Expands attention_mask from `[bsz, seq_len]` to `[bsz, 1, tgt_seq_len, src_seq_len]`.
+ This expansion handles packed sequences so that sequences share the same attention mask integer value
+ when they attend to each other within that sequence.
+ This expansion transforms the mask to lower triangular form to prevent future peeking.
+ """
+ bsz, src_len = mask.size()
+ tgt_len = tgt_len if tgt_len is not None else src_len
+
+ mask = mask.unsqueeze(1).unsqueeze(2)
+ mask = mask.expand(bsz, 1, tgt_len, src_len)
+
+ # Create a binary mask from the original mask where zeros remain zeros and all other values are set to one
+ binary_mask = torch.where(
+ mask != 0,
+ torch.tensor(1, device=mask.device).to(dtype),
+ torch.tensor(0, device=mask.device).to(dtype),
+ )
+
+ # Create a block-diagonal mask.
+ # we multiply by the binary mask so that 0's in the original mask are correctly excluded
+ zero_one_mask = torch.eq(mask, mask.transpose(-1, -2)).int() * binary_mask
+
+ # Now let's create a lower triangular mask of ones that will zero out the upper triangular part
+ lower_triangular_ones = torch.tril(torch.ones((tgt_len, src_len), dtype=dtype)).to(
+ mask.device
+ )
+
+ # Use the lower triangular mask to zero out the upper triangular part of the zero_one_mask
+ masked_zero_one_mask = zero_one_mask * lower_triangular_ones
+
+ return masked_zero_one_mask
+
+
+def patched_prepare_4d_causal_attention_mask(
+ attention_mask: Optional[torch.Tensor],
+ *args,
+):
+ dtype = torch.bfloat16 if is_torch_bf16_gpu_available() else torch.float32
+ return _prepare_4d_causal_attention_mask(
+ mask_2d_to_4d(attention_mask, dtype=dtype),
+ *args,
+ )
+
+
+def patched_prepare_4d_causal_attention_mask_for_sdpa(
+ attention_mask: Optional[torch.Tensor],
+ *args,
+):
+ dtype = torch.bfloat16 if is_torch_bf16_gpu_available() else torch.float32
+ return _prepare_4d_causal_attention_mask_for_sdpa(
+ mask_2d_to_4d(attention_mask, dtype=dtype),
+ *args,
+ )
diff --git a/src/axolotl/prompt_strategies/__init__.py b/src/axolotl/prompt_strategies/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e62a5c20ce4dca7e25a47d73c2803f96f86b563e
--- /dev/null
+++ b/src/axolotl/prompt_strategies/__init__.py
@@ -0,0 +1,26 @@
+"""Module to load prompt strategies."""
+
+import importlib
+import inspect
+
+from axolotl.prompt_strategies.user_defined import UserDefinedDatasetConfig
+
+
+def load(strategy, tokenizer, cfg, ds_cfg):
+ try:
+ load_fn = "load"
+ if strategy.split(".")[-1].startswith("load_"):
+ load_fn = strategy.split(".")[-1]
+ strategy = ".".join(strategy.split(".")[:-1])
+ mod = importlib.import_module(f".{strategy}", "axolotl.prompt_strategies")
+ func = getattr(mod, load_fn)
+ load_kwargs = {}
+ if strategy == "user_defined":
+ load_kwargs["ds_cfg"] = UserDefinedDatasetConfig(**ds_cfg)
+ else:
+ sig = inspect.signature(func)
+ if "ds_cfg" in sig.parameters:
+ load_kwargs["ds_cfg"] = ds_cfg
+ return func(tokenizer, cfg, **load_kwargs)
+ except Exception: # pylint: disable=broad-exception-caught
+ return None
diff --git a/src/axolotl/prompt_strategies/__pycache__/__init__.cpython-310.pyc b/src/axolotl/prompt_strategies/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..9ac517ab7df8657465a5c6c897effd35faefb9bb
Binary files /dev/null and b/src/axolotl/prompt_strategies/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/prompt_strategies/__pycache__/alpaca_w_system.cpython-310.pyc b/src/axolotl/prompt_strategies/__pycache__/alpaca_w_system.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..daf2fc51e807b309ed7774fb72b0c1c7bb52c2de
Binary files /dev/null and b/src/axolotl/prompt_strategies/__pycache__/alpaca_w_system.cpython-310.pyc differ
diff --git a/src/axolotl/prompt_strategies/__pycache__/base.cpython-310.pyc b/src/axolotl/prompt_strategies/__pycache__/base.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..1df4bad74361700a232434a8d56f0acec872aa68
Binary files /dev/null and b/src/axolotl/prompt_strategies/__pycache__/base.cpython-310.pyc differ
diff --git a/src/axolotl/prompt_strategies/__pycache__/sharegpt.cpython-310.pyc b/src/axolotl/prompt_strategies/__pycache__/sharegpt.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..bef95af675fc75e47cd665b1f736e682dd1e8232
Binary files /dev/null and b/src/axolotl/prompt_strategies/__pycache__/sharegpt.cpython-310.pyc differ
diff --git a/src/axolotl/prompt_strategies/__pycache__/user_defined.cpython-310.pyc b/src/axolotl/prompt_strategies/__pycache__/user_defined.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..d98fb7c1af3e63eb7b5635cffa478b86b107c6f1
Binary files /dev/null and b/src/axolotl/prompt_strategies/__pycache__/user_defined.cpython-310.pyc differ
diff --git a/src/axolotl/prompt_strategies/alpaca_chat.py b/src/axolotl/prompt_strategies/alpaca_chat.py
new file mode 100644
index 0000000000000000000000000000000000000000..975fee889e1a2168508f64611781ab9bb537b114
--- /dev/null
+++ b/src/axolotl/prompt_strategies/alpaca_chat.py
@@ -0,0 +1,120 @@
+"""Module for Alpaca prompt strategy classes"""
+
+from typing import Any, Dict, Optional, Tuple
+
+from axolotl.prompt_tokenizers import (
+ AlpacaPromptTokenizingStrategy,
+ InstructionPromptTokenizingStrategy,
+)
+from axolotl.prompters import AlpacaPrompter, PromptStyle, UnpromptedPrompter
+
+
+def load(tokenizer, cfg, ds_cfg: Optional[Dict[str, Any]] = None):
+ prompt_style = PromptStyle.CHAT.value
+ if ds_cfg and "conversation" in ds_cfg:
+ prompt_style = ds_cfg["conversation"]
+
+ return AlpacaPromptTokenizingStrategy(
+ AlpacaPrompter(prompt_style),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+class AlpacaConcisePrompter(AlpacaPrompter):
+ """
+ Alpaca Prompter extending the system prompt to ask for concise chat-instruct answers
+ """
+
+ system_prompt = "Below is an instruction from a USER that describes a task, paired with an input that provides further context. The ASSISTANT writes a response that concisely and appropriately completes the request.\n\n"
+ system_no_input_prompt = "Below is an instruction from a USER that describes a task. The ASSISTANT writes a response that appropriately and concisely completes the request.\n\n"
+
+
+class AlpacaChatPrompter(AlpacaPrompter):
+ """
+ Alpaca Chat Prompter extending the system prompt to for chat-instruct answers
+ """
+
+ system_prompt = "Below is an instruction from a USER that describes a task, paired with an input that provides further context. The ASSISTANT writes a response that concisely and appropriately completes the request.\n\n"
+ system_no_input_prompt = "Below is an instruction from a USER that describes a task. The ASSISTANT writes a response that appropriately and concisely completes the request.\n\n"
+
+ def __init__(self): # pylint: disable=super-init-not-called
+ self.prompt_style = PromptStyle.CHAT.value
+ self.match_prompt_style()
+
+
+class NoSystemPrompter(AlpacaPrompter):
+ """
+ Null Prompter with no system prompts
+ """
+
+ system_prompt = ""
+ system_no_input_prompt = ""
+ turn_format = "{instruction} {input} "
+ turn_no_input_format = "{instruction} "
+
+ def __init__(self): # pylint: disable=super-init-not-called
+ pass
+
+
+class AlpacaQAPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for AlpacaQA
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["question"],
+ "",
+ prompt["answer"],
+ )
+
+
+class CamelAIPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for CamelAI datasets
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["message_1"],
+ "",
+ prompt["message_2"],
+ )
+
+
+def load_concise(tokenizer, cfg):
+ return AlpacaPromptTokenizingStrategy(
+ AlpacaConcisePrompter(PromptStyle.CHAT.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_qa(tokenizer, cfg):
+ return AlpacaQAPromptTokenizingStrategy(
+ AlpacaChatPrompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_camel_ai(tokenizer, cfg):
+ return CamelAIPromptTokenizingStrategy(
+ AlpacaChatPrompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_no_prompt(tokenizer, cfg):
+ return AlpacaPromptTokenizingStrategy(
+ UnpromptedPrompter(PromptStyle.CHAT.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
diff --git a/src/axolotl/prompt_strategies/alpaca_instruct.py b/src/axolotl/prompt_strategies/alpaca_instruct.py
new file mode 100644
index 0000000000000000000000000000000000000000..143f070f2a9f826f6ed83e64305760229f535173
--- /dev/null
+++ b/src/axolotl/prompt_strategies/alpaca_instruct.py
@@ -0,0 +1,22 @@
+"""Module loading the AlpacaInstructPromptTokenizingStrategy class"""
+
+from axolotl.prompt_tokenizers import AlpacaPromptTokenizingStrategy
+from axolotl.prompters import AlpacaPrompter, PromptStyle, UnpromptedPrompter
+
+
+def load(tokenizer, cfg):
+ return AlpacaPromptTokenizingStrategy(
+ AlpacaPrompter(PromptStyle.INSTRUCT.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_no_prompt(tokenizer, cfg):
+ return AlpacaPromptTokenizingStrategy(
+ UnpromptedPrompter(PromptStyle.INSTRUCT.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
diff --git a/src/axolotl/prompt_strategies/alpaca_w_system.py b/src/axolotl/prompt_strategies/alpaca_w_system.py
new file mode 100644
index 0000000000000000000000000000000000000000..8c8cc07435f9e65e5401588aded9c3791b1c6de9
--- /dev/null
+++ b/src/axolotl/prompt_strategies/alpaca_w_system.py
@@ -0,0 +1,163 @@
+"""
+Prompt strategies loader for alpaca instruction datasets with system prompts
+"""
+from typing import Generator, Tuple, Union
+
+from axolotl.prompt_tokenizers import PromptTokenizingStrategy
+from axolotl.prompters import AlpacaPrompter, PromptStyle
+
+
+class InstructionWSystemPromptTokenizingStrategy(PromptTokenizingStrategy):
+ """
+ Tokenizing strategy for instruction-based prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str, str]:
+ return (
+ prompt["instruction"],
+ prompt["input"] if "input" in prompt else "",
+ prompt["output"],
+ prompt["system"],
+ )
+
+ def tokenize_prompt(self, prompt):
+ # pylint: disable=duplicate-code
+ (
+ instruction,
+ input, # pylint: disable=redefined-builtin
+ response,
+ system,
+ ) = self.parse_instruction_fields(prompt)
+ user_prompt = next(
+ iter(
+ self.prompter.build_prompt_w_system(
+ system,
+ instruction,
+ input,
+ )
+ )
+ )
+ tokenized_prompt = self._tokenize(user_prompt, add_eos_token=False)
+ if not self.train_on_inputs:
+ user_prompt_len = len(tokenized_prompt["input_ids"])
+ # TODO this could be sped up using numpy array slicing
+ tokenized_prompt["labels"] = [-100] * user_prompt_len
+ tokenized_res_prompt = self._tokenize(
+ response, strip_bos_token=True, add_eos_token=True
+ )
+ tokenized_prompt["input_ids"] += tokenized_res_prompt["input_ids"]
+ tokenized_prompt["attention_mask"] += tokenized_res_prompt["attention_mask"]
+ tokenized_prompt["labels"] += tokenized_res_prompt["input_ids"]
+
+ return tokenized_prompt
+
+
+class SystemDataPrompter(AlpacaPrompter):
+ """
+ Alpaca Style Prompter that uses system prompts from the dataset
+ """
+
+ system_format: str = "### System:\n{system}\n\n"
+
+ def build_prompt_w_system(
+ self,
+ system: str,
+ instruction: str,
+ input: Union[None, str] = None, # pylint: disable=redefined-builtin
+ output: Union[None, str] = None,
+ ) -> Generator[str, None, None]:
+ # returns the full prompt from instruction and optional input
+ # if a label (=response, =output) is provided, it's also appended.
+ formatted_sys_prompt = (
+ self.system_format.format(system=system)
+ if system and self.system_format
+ else ""
+ )
+ if input:
+ res = formatted_sys_prompt + self.turn_format.format(
+ instruction=instruction, input=input
+ )
+ else:
+ res = formatted_sys_prompt + self.turn_no_input_format.format(
+ instruction=instruction
+ )
+ if output:
+ res = f"{res}{output}"
+ yield res
+
+
+class OpenOrcaSystemDataPrompter(SystemDataPrompter):
+ """
+ Alpaca Style Prompter that uses system prompts from the dataset, with OpenOrca prompts
+ """
+
+ def match_prompt_style(self):
+ # pylint: disable=duplicate-code
+ if self.prompt_style == PromptStyle.INSTRUCT.value:
+ self.turn_format = "### Human:\n{instruction}\n### Additional Context:\n{input}\n### Assistant:\n"
+ self.turn_no_input_format = "### Human:\n{instruction}\n### Assistant:\n"
+ self.system_format = "### System:\n{system}\n"
+ if self.prompt_style == PromptStyle.CHAT.value:
+ self.turn_format = "USER: {instruction}\n{input}\nASSISTANT:"
+ self.turn_no_input_format = "USER: {instruction}\nASSISTANT:"
+ self.system_format = "SYSTEM: {system}\n"
+ if self.prompt_style == PromptStyle.CHATML.value:
+ self.turn_format = "<|im_start|>user\n{instruction}\n{input}<|im_end|>\n<|im_start|>assistant\n"
+ self.turn_no_input_format = (
+ "<|im_start|>user\n{instruction}<|im_end|>\n<|im_start|>assistant\n"
+ )
+ self.system_format = "<|im_start|>system\n{system}<|im_end|>\n"
+
+
+class OpenOrcaPromptTokenizingStrategy(InstructionWSystemPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for OpenOrca datasets
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str, str]:
+ return (
+ prompt["question"],
+ "",
+ prompt["response"],
+ prompt["system_prompt"],
+ )
+
+
+def load(tokenizer, cfg):
+ return load_chat(tokenizer, cfg)
+
+
+def load_instruct(tokenizer, cfg):
+ return InstructionWSystemPromptTokenizingStrategy(
+ SystemDataPrompter(PromptStyle.INSTRUCT.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_chat(tokenizer, cfg):
+ return InstructionWSystemPromptTokenizingStrategy(
+ SystemDataPrompter(PromptStyle.CHAT.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_open_orca(tokenizer, cfg):
+ return OpenOrcaPromptTokenizingStrategy(
+ OpenOrcaSystemDataPrompter(PromptStyle.INSTRUCT.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_open_orca_chatml(tokenizer, cfg):
+ return OpenOrcaPromptTokenizingStrategy(
+ OpenOrcaSystemDataPrompter(PromptStyle.CHATML.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
diff --git a/src/axolotl/prompt_strategies/base.py b/src/axolotl/prompt_strategies/base.py
new file mode 100644
index 0000000000000000000000000000000000000000..fce2aba14a8da1b4ba0286fa7839fc77b46a8dfd
--- /dev/null
+++ b/src/axolotl/prompt_strategies/base.py
@@ -0,0 +1,20 @@
+"""
+module for base dataset transform strategies
+"""
+
+import importlib
+import logging
+
+LOG = logging.getLogger("axolotl")
+
+
+def load(strategy, cfg, module_base=None, **kwargs):
+ try:
+ load_fn = strategy.split(".")[-1]
+ strategy = ".".join(strategy.split(".")[:-1])
+ mod = importlib.import_module(f".{strategy}", module_base)
+ func = getattr(mod, load_fn)
+ return func(cfg, **kwargs)
+ except Exception: # pylint: disable=broad-exception-caught
+ LOG.warning(f"unable to load strategy {strategy}")
+ return None
diff --git a/src/axolotl/prompt_strategies/chat_template.py b/src/axolotl/prompt_strategies/chat_template.py
new file mode 100644
index 0000000000000000000000000000000000000000..8dff3845b715b9b979a275c078044c27b9ba4184
--- /dev/null
+++ b/src/axolotl/prompt_strategies/chat_template.py
@@ -0,0 +1,78 @@
+"""
+HF Chat Templates prompt strategy
+"""
+from typing import Any, Dict, Optional
+
+from axolotl.prompt_tokenizers import PromptTokenizingStrategy
+from axolotl.prompters import Prompter
+from axolotl.utils.chat_templates import chat_templates
+
+
+class ChatTemplatePrompter(Prompter):
+ """prompter for HF chat templates"""
+
+ def __init__(self, tokenizer, chat_template=None, max_length=2048):
+ self.tokenizer = tokenizer
+ self.chat_template = chat_template
+ self.max_length = max_length
+
+ def build_prompt(self, conversation, add_generation_prompt=False):
+ return self.tokenizer.apply_chat_template(
+ conversation,
+ truncation=True,
+ max_length=self.max_length,
+ add_generation_prompt=add_generation_prompt,
+ chat_template=self.chat_template,
+ )
+
+
+class ChatTemplateStrategy(PromptTokenizingStrategy):
+ """
+ Tokenizing strategy for instruction-based prompts.
+ """
+
+ def tokenize_prompt(self, prompt):
+ turns = self.get_conversation_thread(prompt)
+ prompt_ids = self.prompter.build_prompt([turns[0]], add_generation_prompt=True)
+ input_ids = self.prompter.build_prompt(turns)
+
+ if not self.train_on_inputs:
+ user_prompt_len = len(prompt_ids)
+ labels = [-100] * user_prompt_len + input_ids[user_prompt_len:]
+ else:
+ labels = input_ids
+
+ tokenized_prompt = {
+ "input_ids": input_ids,
+ "labels": labels,
+ "attention_mask": [1] * len(input_ids),
+ }
+
+ return tokenized_prompt
+
+ def get_conversation_thread(self, prompt):
+ conversations = prompt["conversations"]
+ # remap roles - allow for assistant turn
+ role_map = {
+ "human": "user",
+ "user": "user",
+ "assistant": "assistant",
+ "gpt": "assistant",
+ }
+ turns = [
+ {"role": role_map[t["from"]], "content": t["value"]} for t in conversations
+ ]
+ return turns
+
+
+def load(tokenizer, cfg, ds_cfg: Optional[Dict[str, Any]] = None):
+ chat_template = (
+ ds_cfg["chat_template"] if ds_cfg and "chat_template" in ds_cfg else "chatml"
+ )
+ strategy = ChatTemplateStrategy(
+ ChatTemplatePrompter(tokenizer, chat_templates(chat_template)),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ return strategy
diff --git a/src/axolotl/prompt_strategies/completion.py b/src/axolotl/prompt_strategies/completion.py
new file mode 100644
index 0000000000000000000000000000000000000000..3285e667cbc36b03e596e548c0321352b1dc25a6
--- /dev/null
+++ b/src/axolotl/prompt_strategies/completion.py
@@ -0,0 +1,92 @@
+"""
+Basic completion text
+"""
+from collections import defaultdict
+from typing import Any, Dict, Generator, Optional, Tuple
+
+from axolotl.prompt_tokenizers import InstructionPromptTokenizingStrategy
+
+
+class CompletionPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Completion prompts.
+ """
+
+ _field: str = "text"
+
+ def __init__(self, *args, max_length=None, **kwargs):
+ super().__init__(*args, **kwargs)
+ if max_length is not None:
+ self.max_length = max_length
+
+ @property
+ def supports_batched(self):
+ return True
+
+ @property
+ def field(self) -> str:
+ return self._field
+
+ @field.setter
+ def field(self, new_field: str):
+ self._field = new_field
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt[self.field],
+ "",
+ "",
+ )
+
+ def tokenize_prompt(self, prompt):
+ res = defaultdict(lambda: [])
+ feature_names = list(prompt.keys())
+ for row in zip(*prompt.values()):
+ prompt_row = dict(zip(feature_names, row))
+ (
+ instruction,
+ _,
+ _,
+ ) = self.parse_instruction_fields(prompt_row)
+
+ full_prompt = self._build_full_prompt(instruction, None, None)
+ tokenized_full_prompt = self._tokenize(full_prompt)
+
+ for key, val in tokenized_full_prompt.items():
+ for i in range(0, len(val), self.sequence_len):
+ res[key].append(val[i : i + self.sequence_len])
+
+ return dict(res)
+
+ def _build_full_prompt(
+ self, instruction, input, response
+ ): # pylint: disable=redefined-builtin
+ return next(iter(self.prompter.build_prompt(instruction, input, response)))
+
+
+class CompletionPrompter:
+ """
+ Prompter for completion
+ """
+
+ def build_prompt(
+ self,
+ instruction: str,
+ input=None, # pylint: disable=redefined-builtin, unused-argument
+ output=None, # pylint: disable=unused-argument
+ ) -> Generator[str, None, None]:
+ yield instruction
+
+
+def load(tokenizer, cfg, ds_cfg: Optional[Dict[str, Any]] = None):
+ strat = CompletionPromptTokenizingStrategy(
+ CompletionPrompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ max_length=cfg.sequence_len * 64,
+ )
+ if ds_cfg and "field" in ds_cfg:
+ strat.field = ds_cfg["field"]
+
+ return strat
diff --git a/src/axolotl/prompt_strategies/context_qa.py b/src/axolotl/prompt_strategies/context_qa.py
new file mode 100644
index 0000000000000000000000000000000000000000..f87dd8b5cdd0311510f348302c16eb3a0a26c13a
--- /dev/null
+++ b/src/axolotl/prompt_strategies/context_qa.py
@@ -0,0 +1,108 @@
+"""Module containing the classes for Context QA Prompt Tokenization Strategies"""
+from typing import Tuple
+
+from axolotl.prompt_tokenizers import InstructionPromptTokenizingStrategy
+from axolotl.prompters import AlpacaPrompter, PromptStyle
+
+
+# article, unanswerable_question, question, answer
+def load_404(tokenizer, cfg):
+ return AlpacaMissingInfoContextPromptTokenizingStrategy(
+ AlpacaContextPrompter(PromptStyle.CHAT.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load(tokenizer, cfg):
+ return AlpacaContextPromptTokenizingStrategy(
+ AlpacaContextPrompter(PromptStyle.CHAT.value),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_v2(tokenizer, cfg):
+ return ContextQaV2PromptTokenizingStrategy(
+ ContextV2Prompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+class AlpacaContextPrompter(AlpacaPrompter):
+ """
+ Customized system prompted for concise QA
+ """
+
+ system_prompt = (
+ "Use the following contextual information to concisely answer the question.\n"
+ )
+ system_no_input_prompt = (
+ "Use the following contextual information to concisely answer the question.\n"
+ )
+
+
+class AlpacaContextPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenization Strategy to combine in-context article with a question and answer
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["article"] + "\n===\n" + prompt["question"],
+ "",
+ prompt["answer"],
+ )
+
+
+class ContextQaV2PromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenization Strategy to combine in-context article with a question and answer
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ "Context: "
+ + prompt["context"]
+ + "\nQuestion: "
+ + prompt["question"]
+ + "\n",
+ "",
+ "Answer: " + prompt["answer"],
+ )
+
+
+class ContextV2Prompter(AlpacaPrompter):
+ """
+ Customized system prompted for concise QA
+ """
+
+ system_prompt = ""
+ system_no_input_prompt = ""
+
+ def match_prompt_style(self):
+ # pylint: disable=duplicate-code
+ self.turn_format = "{instruction}\n{input}"
+ self.turn_no_input_format = "{instruction}"
+ self.system_format = "{system}"
+
+
+class AlpacaMissingInfoContextPromptTokenizingStrategy(
+ InstructionPromptTokenizingStrategy
+):
+ """
+ Tokenization Strategy to combine in-context article with a question that can't be answered
+ from the context and a default response to that effect
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["article"] + "\n===\n" + prompt["unanswerable_question"],
+ "",
+ "The context provided does not contain any information about your inquiry. "
+ "Therefore, I'm unable to answer your question based on the given context.",
+ )
diff --git a/src/axolotl/prompt_strategies/creative_acr.py b/src/axolotl/prompt_strategies/creative_acr.py
new file mode 100644
index 0000000000000000000000000000000000000000..ea67034b3b60cbef4227d12df1b4b32aecc6228a
--- /dev/null
+++ b/src/axolotl/prompt_strategies/creative_acr.py
@@ -0,0 +1,199 @@
+"""Module loading the CreativePromptTokenizingStrategy and similar classes"""
+
+from typing import Generator, Tuple, Union
+
+import yaml
+
+from axolotl.prompt_tokenizers import InstructionPromptTokenizingStrategy
+
+
+class CreativeAnsweringPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Creative Answering
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ question = prompt["instruction"]
+ answer = prompt[
+ "revision"
+ ] # don't use prompt[answer], that's data we don't want in the dataset
+ return (
+ question,
+ "",
+ answer,
+ )
+
+
+class CreativeCritiquePromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Creative Critique
+ """
+
+ user_prompt = """Given the following Question and Response, critique the Response on a scale of 1-10. You should critique the answer in the following criteria:
+refusal: whether the Response refuses to answer the Question. Responses containing language similar to "As an AI language model, I am not capable of ..." is a form of refusal and should have a low refusal score. 10 is no refusal, 0 is refuses to answer the question.
+prescriptive bias: whether the Response attempts to prescribe or dictate certain actions or behaviors of the user. 10 no prescriptive bias present, 0 means there is prescriptive bias.
+creativity: score how creative the Response is. 10 is most creative, 0 lacks creativity
+comprehensiveness: score how comprehensive and in-depth the Response is. 10 is most comprehensive and in-depth, 0 lacks coverage and depth
+Then for each criteria, give a detailed explanation of why the response doesn't meet the criteria (don't reference the score), what it did right, and how it could be improved for that criteria.
+
+Output your answer in YAML format like so:
+scores:
+ refusal: <0-10>
+ prescriptive_bias: <0-10>
+ creativity: <0-10>
+ comprehensiveness: <0-10>
+critiques:
+ refusal:
+ explanation: ...
+ improvements: ...
+ prescriptive_bias:
+ explanation: ...
+ improvements: ...
+ creativity:
+ explanation: ...
+ improvements: ...
+ comprehensiveness:
+ explanation: ...
+ improvements: ...
+
+Question: {question}
+Answer: {answer}
+"""
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ scores = yaml.dump(
+ prompt["scores"],
+ default_flow_style=False,
+ Dumper=yaml.Dumper,
+ )
+ critiques = yaml.dump(
+ prompt["critiques"],
+ default_flow_style=False,
+ Dumper=yaml.Dumper,
+ )
+ evaluation = scores + critiques
+ question = prompt["instruction"]
+ answer = prompt["answer"]
+ return (
+ self.user_prompt.format(question=question, answer=answer),
+ "",
+ evaluation,
+ )
+
+
+class CreativeRevisePromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Creative Revise
+ """
+
+ user_prompt = """Definitions:
+refusal: whether the Response refuses to answer the Question. Responses containing language similar to "As an AI language model, I am not capable of ..." is a form of refusal and should have a low refusal score. 10 is no refusal, 0 is refuses to answer the question.
+prescriptive bias: whether the Response attempts to prescribe or dictate certain actions or behaviors of the user. 10 no prescriptive bias present, 0 means their is prescriptive bias.
+creativity: score how creative the Response is. 10 is most creative, 0 lacks creativity
+comprehensiveness: score how comprehensive and in-depth the Response is. 10 is most comprehensive and in-depth, 0 lacks coverage and depth
+
+Given the following Question, Response, and Evaluation, revise the Response based on the Evaluation and recommendations for improvements. Reply only with the revised response.
+
+Question: {question}
+Answer: {answer}
+Evaluation:
+{evaluation}
+"""
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ scores = yaml.dump(
+ prompt["scores"],
+ default_flow_style=False,
+ Dumper=yaml.Dumper,
+ )
+ critiques = yaml.dump(
+ prompt["critiques"],
+ default_flow_style=False,
+ Dumper=yaml.Dumper,
+ )
+ evaluation = scores + critiques
+ question = prompt["instruction"]
+ answer = prompt["answer"]
+ return (
+ self.user_prompt.format(
+ question=question, answer=answer, evaluation=evaluation
+ ),
+ "",
+ prompt["revision"],
+ )
+
+
+class CreativePrompterBase:
+ """
+ Base class for Creative Prompters
+ """
+
+ system_prompt = ""
+ prompt_input = "{system_prompt}\nUSER: {instruction}\nASSISTANT:"
+
+ def build_prompt(
+ self,
+ instruction: str,
+ input: Union[ # pylint: disable=redefined-builtin, unused-argument
+ None, str
+ ] = None,
+ output: Union[None, str] = None,
+ ) -> Generator[str, None, None]:
+ if self.system_prompt:
+ res = f"{self.system_prompt}\nUSER: {instruction}\nASSISTANT:"
+ else:
+ res = f"USER: {instruction}\nASSISTANT:"
+ if output:
+ res = f"{res}{output}"
+ yield res
+
+
+class CreativeAnswerPrompter(CreativePrompterBase):
+ """
+ Prompter for Creative Answering
+ """
+
+ system_prompt = "Answer the following question in a comprehensive, in-depth, and creative way. Additionally your response should be relevant, accurate, and free of any ambiguity."
+
+
+class CreativeCritiquePrompter(CreativePrompterBase):
+ """
+ Prompter for Creative Critique
+ """
+
+ system_prompt = ""
+
+
+class CreativeRevisePrompter(CreativePrompterBase):
+ """
+ Prompter for Creative Revise
+ """
+
+ system_prompt = ""
+
+
+def load_answer(tokenizer, cfg):
+ return CreativeAnsweringPromptTokenizingStrategy(
+ CreativeAnswerPrompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_critique(tokenizer, cfg):
+ return CreativeCritiquePromptTokenizingStrategy(
+ CreativeCritiquePrompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+def load_revise(tokenizer, cfg):
+ return CreativeRevisePromptTokenizingStrategy(
+ CreativeRevisePrompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
diff --git a/src/axolotl/prompt_strategies/dpo/__init__.py b/src/axolotl/prompt_strategies/dpo/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..7f5e6eb6447e696b060276f961bb1cbb1a151f9e
--- /dev/null
+++ b/src/axolotl/prompt_strategies/dpo/__init__.py
@@ -0,0 +1,8 @@
+"""
+module for DPO style dataset transform strategies
+"""
+from functools import partial
+
+from ..base import load as load_base
+
+load = partial(load_base, module_base="axolotl.prompt_strategies.dpo")
diff --git a/src/axolotl/prompt_strategies/dpo/__pycache__/__init__.cpython-310.pyc b/src/axolotl/prompt_strategies/dpo/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..5f8a7dbc439701b78478fb46654669f7ee5725b1
Binary files /dev/null and b/src/axolotl/prompt_strategies/dpo/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/prompt_strategies/dpo/chatml.py b/src/axolotl/prompt_strategies/dpo/chatml.py
new file mode 100644
index 0000000000000000000000000000000000000000..585696e29a893f23555bddc95d1fb8f49b833c83
--- /dev/null
+++ b/src/axolotl/prompt_strategies/dpo/chatml.py
@@ -0,0 +1,133 @@
+"""
+DPO strategies for chatml
+"""
+
+
+def argilla(
+ cfg,
+ **kwargs,
+): # pylint: disable=possibly-unused-variable,unused-argument
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|im_start|>system\n{sample['system']}<|im_end|>\n"
+ f"<|im_start|>user\n{sample['instruction']}<|im_end|>\n<|im_start|>assistant\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|im_start|>user\n{sample['instruction']}<|im_end|>\n<|im_start|>assistant\n"
+ sample["chosen"] = f"{sample['chosen_response']}<|im_end|>"
+ sample["rejected"] = f"{sample['rejected_response']}<|im_end|>"
+ return sample
+
+ return transform_fn
+
+
+def argilla_chat(
+ cfg,
+ **kwargs,
+): # pylint: disable=possibly-unused-variable,unused-argument
+ """
+ for argilla/dpo-mix-7k conversations
+ """
+
+ def transform_fn(sample):
+ sample[
+ "prompt"
+ ] = f"<|im_start|>user\n{sample['chosen'][0]['content']}<|im_end|>\n<|im_start|>assistant\n"
+ sample["chosen"] = f"{sample['chosen'][1]['content']}<|im_end|>"
+ sample["rejected"] = f"{sample['rejected'][1]['content']}<|im_end|>"
+ return sample
+
+ return transform_fn
+
+
+def icr(
+ cfg,
+ **kwargs,
+): # pylint: disable=possibly-unused-variable,unused-argument
+ """
+ chatml transforms for datasets with system, input, chosen, rejected
+ ex. https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs
+ """
+
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|im_start|>system\n{sample['system']}<|im_end|>\n"
+ f"<|im_start|>user\n{sample['input']}<|im_end|>\n<|im_start|>assistant\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|im_start|>user\n{sample['input']}<|im_end|>\n<|im_start|>assistant\n"
+ sample["chosen"] = f"{sample['chosen']}<|im_end|>"
+ sample["rejected"] = f"{sample['rejected']}<|im_end|>"
+ return sample
+
+ return transform_fn
+
+
+def intel(cfg, **kwargs): # pylint: disable=possibly-unused-variable,unused-argument
+ """
+ For Intel Orca DPO Pairs
+ """
+
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|im_start|>system\n{sample['system']}<|im_end|>\n"
+ f"<|im_start|>user\n{sample['question']}<|im_end|>\n<|im_start|>assistant\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|im_start|>user\n{sample['question']}<|im_end|>\n<|im_start|>assistant\n"
+ sample["chosen"] = f"{sample['chosen']}<|im_end|>"
+ sample["rejected"] = f"{sample['rejected']}<|im_end|>"
+ return sample
+
+ return transform_fn
+
+
+def prompt_pairs(
+ cfg, **kwargs
+): # pylint: disable=possibly-unused-variable,unused-argument
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|im_start|>system\n{sample['system']}<|im_end|>\n"
+ f"<|im_start|>user\n{sample['prompt']}<|im_end|>\n<|im_start|>assistant\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|im_start|>user\n{sample['prompt']}<|im_end|>\n<|im_start|>assistant\n"
+ sample["chosen"] = f"{sample['chosen']}<|im_end|>"
+ sample["rejected"] = f"{sample['rejected']}<|im_end|>"
+ return sample
+
+ return transform_fn
+
+
+def ultra(cfg, **kwargs): # pylint: disable=possibly-unused-variable,unused-argument
+ """
+ for ultrafeedback binarized conversations
+ """
+
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|im_start|>system\n{sample['system']}<|im_end|>\n"
+ f"<|im_start|>user\n{sample['prompt']}<|im_end|>\n<|im_start|>assistant\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|im_start|>user\n{sample['prompt']}<|im_end|>\n<|im_start|>assistant\n"
+ sample["chosen"] = f"{sample['chosen'][1]['content']}<|im_end|>"
+ sample["rejected"] = f"{sample['rejected'][1]['content']}<|im_end|>"
+ return sample
+
+ return transform_fn
diff --git a/src/axolotl/prompt_strategies/dpo/llama3.py b/src/axolotl/prompt_strategies/dpo/llama3.py
new file mode 100644
index 0000000000000000000000000000000000000000..cb394cc228ec7f6cac7ae160c76fcb30539c7e31
--- /dev/null
+++ b/src/axolotl/prompt_strategies/dpo/llama3.py
@@ -0,0 +1,133 @@
+"""
+DPO strategies for llama-3 chat template
+"""
+
+
+def argilla(
+ cfg,
+ **kwargs,
+): # pylint: disable=possibly-unused-variable,unused-argument
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|start_header_id|>system<|end_header_id|>\n\n{sample['system']}<|eot_id|>"
+ f"<|start_header_id|>user<|end_header_id|>\n\n{sample['instruction']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|start_header_id|>user<|end_header_id|>\n\n{sample['instruction']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ sample["chosen"] = f"{sample['chosen_response']}<|eot_id|>"
+ sample["rejected"] = f"{sample['rejected_response']}<|eot_id|>"
+ return sample
+
+ return transform_fn
+
+
+def argilla_chat(
+ cfg,
+ **kwargs,
+): # pylint: disable=possibly-unused-variable,unused-argument
+ """
+ for argilla/dpo-mix-7k conversations
+ """
+
+ def transform_fn(sample):
+ sample[
+ "prompt"
+ ] = f"<|start_header_id|>user<|end_header_id|>\n\n{sample['chosen'][0]['content']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ sample["chosen"] = f"{sample['chosen'][1]['content']}<|eot_id|>"
+ sample["rejected"] = f"{sample['rejected'][1]['content']}<|eot_id|>"
+ return sample
+
+ return transform_fn
+
+
+def icr(
+ cfg,
+ **kwargs,
+): # pylint: disable=possibly-unused-variable,unused-argument
+ """
+ chatml transforms for datasets with system, input, chosen, rejected
+ ex. https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs
+ """
+
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|start_header_id|>system<|end_header_id|>\n\n{sample['system']}<|eot_id|>"
+ f"<|start_header_id|>user<|end_header_id|>\n\n{sample['input']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|start_header_id|>user<|end_header_id|>\n\n{sample['input']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ sample["chosen"] = f"{sample['chosen']}<|eot_id|>"
+ sample["rejected"] = f"{sample['rejected']}<|eot_id|>"
+ return sample
+
+ return transform_fn
+
+
+def intel(cfg, **kwargs): # pylint: disable=possibly-unused-variable,unused-argument
+ """
+ For Intel Orca DPO Pairs
+ """
+
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|start_header_id|>system<|end_header_id|>\n\n{sample['system']}<|eot_id|>"
+ f"<|start_header_id|>user<|end_header_id|>\n\n{sample['question']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|start_header_id|>user<|end_header_id|>\n\n{sample['question']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ sample["chosen"] = f"{sample['chosen']}<|eot_id|>"
+ sample["rejected"] = f"{sample['rejected']}<|eot_id|>"
+ return sample
+
+ return transform_fn
+
+
+def prompt_pairs(
+ cfg, **kwargs
+): # pylint: disable=possibly-unused-variable,unused-argument
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|start_header_id|>system<|end_header_id|>\n\n{sample['system']}<|eot_id|>"
+ f"<|start_header_id|>user<|end_header_id|>\n\n{sample['prompt']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|start_header_id|>user<|end_header_id|>\n\n{sample['prompt']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ sample["chosen"] = f"{sample['chosen']}<|eot_id|>"
+ sample["rejected"] = f"{sample['rejected']}<|eot_id|>"
+ return sample
+
+ return transform_fn
+
+
+def ultra(cfg, **kwargs): # pylint: disable=possibly-unused-variable,unused-argument
+ """
+ for ultrafeedback binarized conversations
+ """
+
+ def transform_fn(sample):
+ if "system" in sample and sample["system"]:
+ sample["prompt"] = (
+ f"<|start_header_id|>system<|end_header_id|>\n\n{sample['system']}<|eot_id|>"
+ f"<|start_header_id|>user<|end_header_id|>\n\n{sample['prompt']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ )
+ else:
+ sample[
+ "prompt"
+ ] = f"<|start_header_id|>user<|end_header_id|>\n\n{sample['prompt']}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
+ sample["chosen"] = f"{sample['chosen'][1]['content']}<|eot_id|>"
+ sample["rejected"] = f"{sample['rejected'][1]['content']}<|eot_id|>"
+ return sample
+
+ return transform_fn
diff --git a/src/axolotl/prompt_strategies/dpo/user_defined.py b/src/axolotl/prompt_strategies/dpo/user_defined.py
new file mode 100644
index 0000000000000000000000000000000000000000..1d5f891af69870a80532f879f87d91156f2174a7
--- /dev/null
+++ b/src/axolotl/prompt_strategies/dpo/user_defined.py
@@ -0,0 +1,41 @@
+"""
+User-defined DPO strategies
+"""
+
+
+def default(cfg, dataset_idx=0, **kwargs): # pylint: disable=unused-argument
+ ds_cfg = cfg["datasets"][dataset_idx]["type"]
+ if not isinstance(ds_cfg, dict):
+ raise ValueError(
+ f"User-defined dataset type must be a dictionary. Got: {ds_cfg}"
+ )
+ field_prompt = ds_cfg.get("field_prompt", "prompt")
+ field_system = ds_cfg.get("field_system", "system")
+ field_chosen = ds_cfg.get("field_chosen", "chosen")
+ field_rejected = ds_cfg.get("field_rejected", "rejected")
+ prompt_format = ds_cfg.get("prompt_format")
+ if not prompt_format:
+ prompt_format = "{" + field_prompt + "}"
+ chosen_format = ds_cfg.get("chosen_format")
+ if not chosen_format:
+ chosen_format = "{" + field_chosen + "}"
+ rejected_format = ds_cfg.get("rejected_format")
+ if not rejected_format:
+ rejected_format = "{" + field_rejected + "}"
+
+ def transform_fn(sample):
+ if (
+ "{" + field_system + "}" in prompt_format
+ and field_system in sample
+ and sample[field_system]
+ ):
+ sample["prompt"] = prompt_format.format(
+ system=sample[field_system], prompt=sample[field_prompt]
+ )
+ else:
+ sample["prompt"] = prompt_format.format(prompt=sample["prompt"])
+ sample["chosen"] = chosen_format.format(chosen=sample[field_chosen])
+ sample["rejected"] = rejected_format.format(rejected=sample[field_rejected])
+ return sample
+
+ return transform_fn
diff --git a/src/axolotl/prompt_strategies/dpo/zephyr.py b/src/axolotl/prompt_strategies/dpo/zephyr.py
new file mode 100644
index 0000000000000000000000000000000000000000..9eb8950091146a844efb5870d2dd45a27569abe6
--- /dev/null
+++ b/src/axolotl/prompt_strategies/dpo/zephyr.py
@@ -0,0 +1,21 @@
+"""
+DPO strategies for zephyr
+"""
+
+
+def nectar(cfg, **kwargs): # pylint: disable=possibly-unused-variable,unused-argument
+ def transform_fn(sample):
+ data = {}
+ data["prompt"] = (
+ "<|system|>\n \n"
+ "<|user|>\n"
+ f"{sample['prompt']} \n"
+ "<|assistant|>\n"
+ )
+ answers = sorted(sample["answers"], key=lambda x: x["rank"])
+ data["chosen"] = answers[-1]["answer"]
+ data["rejected"] = answers[-2]["answer"]
+
+ return data
+
+ return transform_fn
diff --git a/src/axolotl/prompt_strategies/input_output.py b/src/axolotl/prompt_strategies/input_output.py
new file mode 100644
index 0000000000000000000000000000000000000000..fe14f039cf8e98c92f79e56341ea4c5bf438a24c
--- /dev/null
+++ b/src/axolotl/prompt_strategies/input_output.py
@@ -0,0 +1,54 @@
+"""Module for plain input/output prompt pairs"""
+from typing import Generator, Tuple
+
+from axolotl.prompt_tokenizers import PromptTokenizingStrategy
+from axolotl.prompters import IGNORE_TOKEN_ID, Prompter
+
+
+class RawInputOutputStrategy(PromptTokenizingStrategy):
+ """Prompt Strategy class for input/output pairs"""
+
+ def __init__(self, *args, eos_token=None, **kwargs):
+ super().__init__(*args, **kwargs)
+ self.eos_token = eos_token
+ if not eos_token:
+ self.eos_token = self.tokenizer.eos_token
+
+ def tokenize_prompt(self, prompt):
+ # pylint: disable=duplicate-code
+ input_ids = []
+ labels = []
+ for label, text in self.prompter.build_prompt(prompt["segments"]):
+ tokenized_output = self.tokenizer(
+ text, add_special_tokens=False, return_tensors=None
+ )["input_ids"]
+ input_ids += tokenized_output
+ if label or self.train_on_inputs:
+ labels += tokenized_output
+ else:
+ labels += [IGNORE_TOKEN_ID] * len(tokenized_output)
+
+ tokenized_prompt = {
+ "input_ids": input_ids,
+ "labels": labels,
+ "attention_mask": [1] * len(input_ids),
+ }
+
+ return tokenized_prompt
+
+
+class RawInputOutputPrompter(Prompter):
+ """prompter for raw i/o data"""
+
+ def build_prompt(self, source) -> Generator[Tuple[bool, str], None, None]:
+ for segment in source:
+ yield segment["label"], segment["text"]
+
+
+def load(tokenizer, cfg):
+ return RawInputOutputStrategy(
+ RawInputOutputPrompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
diff --git a/src/axolotl/prompt_strategies/instruct.py b/src/axolotl/prompt_strategies/instruct.py
new file mode 100644
index 0000000000000000000000000000000000000000..3d636748903ef1b6a06d81511c00d3545101ef80
--- /dev/null
+++ b/src/axolotl/prompt_strategies/instruct.py
@@ -0,0 +1,33 @@
+"""Module containing the InstructShareGPTPromptTokenizingStrategy class"""
+from typing import Any, Dict, Optional
+
+from axolotl.prompt_tokenizers import ShareGPTPromptTokenizingStrategy
+from axolotl.prompters import ShareGPTPrompterV2
+
+
+def load(tokenizer, cfg, ds_cfg: Optional[Dict[str, Any]] = None):
+ conversation = (
+ ds_cfg["conversation"] if ds_cfg and "conversation" in ds_cfg else None
+ )
+ strategy = InstructShareGPTPromptTokenizingStrategy(
+ # pylint: disable=duplicate-code
+ ShareGPTPrompterV2(
+ conversation=conversation,
+ ),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ return strategy
+
+
+class InstructShareGPTPromptTokenizingStrategy(ShareGPTPromptTokenizingStrategy):
+ """
+ basic sharegpt strategy to grab conversations from the sample row
+ """
+
+ def get_conversation_thread(self, prompt):
+ return [
+ {"from": "human", "value": prompt["instruction"]},
+ {"from": "gpt", "value": prompt["output"]},
+ ]
diff --git a/src/axolotl/prompt_strategies/llama2_chat.py b/src/axolotl/prompt_strategies/llama2_chat.py
new file mode 100644
index 0000000000000000000000000000000000000000..a1f5ffefff3f941694bdb2ba7a9f3bbd9118b748
--- /dev/null
+++ b/src/axolotl/prompt_strategies/llama2_chat.py
@@ -0,0 +1,206 @@
+"""
+Prompt Strategy for finetuning Llama2 chat models
+see also https://github.com/facebookresearch/llama/blob/6c7fe276574e78057f917549435a2554000a876d/llama/generation.py#L213 for ma reference implementation.
+
+This implementation is based on the Vicuna PR and the fastchat repo, see also:
+https://github.com/lm-sys/FastChat/blob/cdd7730686cb1bf9ae2b768ee171bdf7d1ff04f3/fastchat/conversation.py#L847
+
+Use dataset type: "llama2_chat" in conig.yml to use this prompt style.
+
+E.g. in the config.yml:
+```
+datasets:
+ - path: llama_finetune_train.jsonl
+ type: llama2_chat
+```
+
+The dataset itself should look like this:
+```
+{'conversations':[{"from": "human", "value": "Who are you?"}, {"from": "gpt", "value": "I am Vicuna"},...]}
+```
+in a jsonl file. The first message should be from the human, the second from gpt.
+For a custom system message, the first "from" can be "system" (followed by alternating "human" and "gpt" turns).
+
+Important: Don't use "special_tokens:" in your config.yml if you are not sure what you are doing!
+"""
+
+import logging
+from dataclasses import dataclass, field
+from typing import Generator, List, Sequence
+
+from axolotl.prompt_tokenizers import PromptTokenizingStrategy
+from axolotl.prompters import IGNORE_TOKEN_ID, SHAREGPT_ASSERTION_FAILED_ROLE
+
+
+@dataclass
+class Llama2ChatConversation:
+ """A class that manages prompt templates and keeps all conversation history.
+ copied from https://github.com/lm-sys/FastChat/blob/main/fastchat/conversation.py"""
+
+ name: str = "llama2"
+ # The system prompt
+ system: str = (
+ "[INST] <>\nYou are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. "
+ "Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. "
+ "Please ensure that your responses are socially unbiased and positive in nature.\n\n"
+ "If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. "
+ "If you don't know the answer to a question, please don't share false information.\n< >\n\n"
+ )
+ roles: Sequence[str] = ("[INST]", "[/INST]")
+ messages: List[List[str]] = field(default_factory=list)
+ offset: int = 0
+ sep = " "
+ sep2 = " "
+ stop_token_ids = [2]
+
+ def get_prompt(self) -> str:
+ """Get the prompt for generation."""
+ seps = [self.sep, self.sep2]
+ ret = ""
+ for i, (role, message) in enumerate(self.messages):
+ if (i == len(self.messages) - 1) and (role == self.roles[0]):
+ # last message is from user (due to length),
+ # return prompt without it for training
+ return ret
+ if i == 0:
+ ret += self.system + message.strip()
+ else:
+ ret += role + " " + message.strip() + seps[i % 2]
+ return ret
+
+ def append_message(self, role: str, message: str):
+ """Append a new message."""
+ self.messages.append([role, message])
+
+
+class LLama2ChatTokenizingStrategy(PromptTokenizingStrategy):
+ """
+ Tokenizing strategy for ShareGPT prompts.
+ adapted from https://github.com/lm-sys/FastChat/blob/main/fastchat/train/train.py
+ """
+
+ def __init__(self, *args, **kwargs):
+ super().__init__(*args, **kwargs)
+ self.tokenizer.add_special_tokens(
+ {"pad_token": getattr(self.tokenizer, "pad_token", "")}
+ )
+ # https://huggingface.co/meta-llama/Llama-2-7b-chat-hf/blob/main/added_tokens.json
+
+ def tokenize_prompt(self, prompt):
+ conv = next(self.prompter.build_prompt(prompt))
+ conversation_str = conv.get_prompt()
+
+ # Tokenize conversations
+ input_ids = self.tokenizer(
+ conversation_str,
+ return_tensors="pt",
+ padding="max_length",
+ max_length=self.sequence_len,
+ truncation=True,
+ ).input_ids[0]
+ target = input_ids.clone()
+
+ # Mask targets. Only compute loss on the assistant outputs.
+ sep = conv.roles[1]
+
+ total_len = int(target.ne(self.tokenizer.pad_token_id).sum())
+
+ turns = conversation_str.split(conv.sep2)
+ cur_len = 1
+ target[:cur_len] = IGNORE_TOKEN_ID
+ for turn in turns:
+ if turn == "":
+ break
+ turn_len = len(self.tokenizer(turn).input_ids)
+
+ parts = turn.split(sep)
+ if len(parts) != 2:
+ break
+ parts[0] += sep
+ # "-1" is hardcoded for the LLaMA tokenizer to make the offset correct.
+ instruction_len = len(self.tokenizer(parts[0]).input_ids) - 1
+
+ # Ignore the user instructions
+ target[cur_len - 1 : cur_len + instruction_len] = IGNORE_TOKEN_ID
+ cur_len += turn_len + 2 # due to length of role token
+
+ target[cur_len:] = IGNORE_TOKEN_ID
+
+ if cur_len < self.sequence_len:
+ if cur_len != total_len:
+ target[:] = IGNORE_TOKEN_ID
+ logging.warning(
+ f"WARNING: tokenization mismatch: {cur_len} vs. {total_len}."
+ f" (ignored)"
+ )
+
+ attention_mask = input_ids.ne(self.tokenizer.pad_token_id).tolist()
+ input_ids = input_ids.tolist()
+ target = target.tolist()
+ # this is a fix for the tokenizer which tokenizes [ differently with eos tokens and
+ # follows the original llama implementation
+ for i in range(2, total_len - 2):
+ if input_ids[i] == 29961:
+ input_ids[i] = 518
+ if target[i] == 29961:
+ target[i] = 518
+ return {
+ "input_ids": input_ids,
+ "labels": target,
+ "attention_mask": attention_mask,
+ }
+
+
+class Llama2ChatPrompter: # pylint: disable=too-few-public-methods
+ """
+ A prompter that generates prompts for Llama2 models.
+ """
+
+ system_prompt = (
+ "[INST] <>\nYou are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. "
+ "Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. "
+ "Please ensure that your responses are socially unbiased and positive in nature.\n\n"
+ "If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. "
+ "If you don't know the answer to a question, please don't share false information.\n< >\n\n"
+ )
+
+ def build_prompt(self, source) -> Generator[Llama2ChatConversation, None, None]:
+ # see https://github.com/lm-sys/FastChat/blob/da0641e567cf93756b0978ab5a6b092e96f06240/fastchat/train/train.py#L78
+ source = source["conversations"] # fix data structure for datasets
+
+ # if system prompt provided, use it
+ if source[0]["from"] == "system":
+ system = f"[INST] <>\n{source[0]['value']}\n< >\n\n"
+ source = source[1:]
+ else:
+ system = self.system_prompt
+
+ conv = Llama2ChatConversation(system=system)
+
+ if len(source) < 2:
+ # If there isn't a back and forth conversation, ignore it
+ # also happens on the data splitting leaving empty conversations
+ raise IndexError
+
+ roles = {"human": conv.roles[0], "gpt": conv.roles[1]}
+
+ if roles[source[0]["from"]] != conv.roles[0]:
+ # Skip the first one if it is not from human
+ source = source[1:]
+
+ conv.messages = [] # pylint: disable=R0801
+ for j, sentence in enumerate(source):
+ role = roles[sentence["from"]]
+ assert role == conv.roles[j % 2], SHAREGPT_ASSERTION_FAILED_ROLE
+ if sentence["value"]:
+ conv.append_message(role, sentence["value"])
+ yield conv
+
+
+def load(tokenizer, cfg) -> LLama2ChatTokenizingStrategy:
+ return LLama2ChatTokenizingStrategy(
+ Llama2ChatPrompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
diff --git a/src/axolotl/prompt_strategies/metharme.py b/src/axolotl/prompt_strategies/metharme.py
new file mode 100644
index 0000000000000000000000000000000000000000..52d77c00cf9333a7af09c018d0126bd8fbf8a1df
--- /dev/null
+++ b/src/axolotl/prompt_strategies/metharme.py
@@ -0,0 +1,76 @@
+"""Module containing the MetharmenPromptTokenizingStrategy and MetharmePrompter class"""
+
+import logging
+from typing import Tuple
+
+from axolotl.prompt_tokenizers import InstructionPromptTokenizingStrategy
+from axolotl.prompters import AlpacaPrompter
+
+LOG = logging.getLogger("axolotl")
+
+IGNORE_TOKEN_ID = -100
+
+# pylint: disable=duplicate-code
+
+
+class MetharmePromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for the Metharme models
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (prompt["prompt"], "", prompt["generation"])
+
+ def _tokenize(
+ self,
+ prompt: str,
+ add_eos_token: bool = True,
+ strip_bos_token: bool = False,
+ num_eos_tokens: int = 3,
+ ):
+ result = self.tokenizer(
+ prompt,
+ truncation=True,
+ max_length=self.sequence_len,
+ padding=False,
+ return_tensors=None,
+ )
+ if len(result["input_ids"]) == 0:
+ LOG.warning("Tokenizer result is empty. You may want to audit your dataset")
+ # If there's already an EOS token there, subtract from the number added
+ if result["input_ids"][-1] == self.tokenizer.eos_token_id:
+ num_eos_tokens -= 1
+
+ if num_eos_tokens > 0 and add_eos_token and len(result["input_ids"]) > 0:
+ for _ in range(num_eos_tokens):
+ if len(result["input_ids"]) < self.sequence_len:
+ result["input_ids"].append(self.tokenizer.eos_token_id)
+ result["attention_mask"].append(1)
+
+ if result["input_ids"][0] == self.tokenizer.bos_token_id and strip_bos_token:
+ result["input_ids"] = result["input_ids"][1:]
+ result["attention_mask"] = result["attention_mask"][1:]
+
+ result["labels"] = result["input_ids"].copy()
+ return result
+
+
+class MetharmePrompter(AlpacaPrompter):
+ """
+ Prompter for the Metharme models.
+ """
+
+ system_prompt = ""
+ system_no_input_prompt = ""
+ system_format = ""
+ turn_format = "{instruction}"
+ turn_no_input_format = "{instruction}"
+
+ def __init__(self, *args, **kwargs): # pylint: disable=super-init-not-called
+ pass
+
+
+def load(tokenizer, cfg):
+ return MetharmePromptTokenizingStrategy(
+ MetharmePrompter(), tokenizer, cfg.train_on_inputs, cfg.sequence_len
+ )
diff --git a/src/axolotl/prompt_strategies/orcamini.py b/src/axolotl/prompt_strategies/orcamini.py
new file mode 100644
index 0000000000000000000000000000000000000000..04ce5767dd75a000007fa01b15744f4bc459595c
--- /dev/null
+++ b/src/axolotl/prompt_strategies/orcamini.py
@@ -0,0 +1,46 @@
+"""
+Prompt Strategy for finetuning Orca Mini (v2) models
+see also https://huggingface.co/psmathur/orca_mini_v2_7b for more information
+
+Use dataset type: orcamini in conig.yml to use this prompt style.
+
+Compared to the alpaca_w_system.open_orca dataset type,
+this one specifies the system prompt with "### System:".
+
+Not suited/tested for multiple-turn conversations without further adjustments.
+"""
+from typing import Generator, Union
+
+from axolotl.prompt_strategies.alpaca_w_system import OpenOrcaPromptTokenizingStrategy
+from axolotl.prompters import AlpacaPrompter
+
+
+class OrcaMiniPrompter(AlpacaPrompter):
+ """Adjusted Prompter for Orca Mini (v2) datasets"""
+
+ def match_prompt_style(self):
+ self.turn_no_input_format = (
+ "### System:\n{system}\n\n### User:\n{instruction}\n\n### Response:\n"
+ )
+
+ def build_prompt_w_system(
+ self,
+ system: str,
+ instruction: str,
+ output: Union[None, str] = None,
+ ) -> Generator[str, None, None]:
+ # returns the full prompt from instruction and optional input
+ # if a label (=response, =output) is provided, it's also appended.
+ res = self.turn_no_input_format.format(system=system, instruction=instruction)
+ if output:
+ res = f"{res}{output}"
+ yield res
+
+
+def load(tokenizer, cfg):
+ return OpenOrcaPromptTokenizingStrategy(
+ OrcaMiniPrompter(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
diff --git a/src/axolotl/prompt_strategies/orpo/__init__.py b/src/axolotl/prompt_strategies/orpo/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..4a02f3c625ebd3fe80fec141b2ea035129c3df43
--- /dev/null
+++ b/src/axolotl/prompt_strategies/orpo/__init__.py
@@ -0,0 +1,9 @@
+"""
+module for ORPO style dataset transform strategies
+"""
+
+from functools import partial
+
+from ..base import load as load_base
+
+load = partial(load_base, module_base="axolotl.prompt_strategies.orpo")
diff --git a/src/axolotl/prompt_strategies/orpo/__pycache__/__init__.cpython-310.pyc b/src/axolotl/prompt_strategies/orpo/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..76a9a8a0aee45899a9689f5c583f0fa405ff9b58
Binary files /dev/null and b/src/axolotl/prompt_strategies/orpo/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/prompt_strategies/orpo/chat_template.py b/src/axolotl/prompt_strategies/orpo/chat_template.py
new file mode 100644
index 0000000000000000000000000000000000000000..a89dee15751bce26f5342233ef8ad2871127c347
--- /dev/null
+++ b/src/axolotl/prompt_strategies/orpo/chat_template.py
@@ -0,0 +1,272 @@
+"""chatml prompt tokenization strategy for ORPO"""
+from typing import Any, Dict, Generator, List, Optional, Tuple
+
+from pydantic import BaseModel
+
+from axolotl.prompt_tokenizers import IGNORE_INDEX, PromptTokenizingStrategy
+from axolotl.prompters import Prompter
+from axolotl.utils.chat_templates import chat_templates
+
+
+class Message(BaseModel):
+ """message/turn"""
+
+ role: str
+ content: str
+ label: Optional[bool] = None
+
+
+class MessageList(BaseModel):
+ """conversation"""
+
+ messages: List[Message]
+
+
+def load(
+ tokenizer, cfg, ds_cfg: Optional[Dict[str, Any]] = None, **kwargs
+): # pylint: disable=possibly-unused-variable,unused-argument
+ """
+ chatml transforms for datasets with system, input, chosen, rejected
+ """
+
+ chat_template = chat_templates("chatml")
+ if ds_cfg and "chat_template" in ds_cfg:
+ chat_template = ds_cfg["chat_template"]
+ try:
+ chat_template = chat_templates(chat_template)
+ except ValueError:
+ pass
+ tokenizer.chat_template = chat_template
+
+ return ORPOTokenizingStrategy(
+ ORPOPrompter(chat_template, tokenizer),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ dataset_parser=ORPODatasetParsingStrategy(),
+ )
+
+
+class ORPODatasetParsingStrategy:
+ """Strategy to parse chosen rejected dataset into messagelist"""
+
+ def get_chosen_conversation_thread(self, prompt) -> MessageList:
+ """Dataset structure mappings"""
+
+ messages: List[Message] = []
+ if system := prompt.get("system", None):
+ messages.append(Message(role="system", content=system, label=False))
+ messages.append(Message(role="user", content=prompt["prompt"], label=False))
+ messages.append(
+ Message(
+ role="assistant", content=prompt["chosen"][1]["content"], label=True
+ )
+ )
+ return MessageList(messages=messages)
+
+ def get_rejected_conversation_thread(self, prompt) -> MessageList:
+ """Dataset structure mappings"""
+
+ messages: List[Message] = []
+ if system := prompt.get("system", None):
+ messages.append(Message(role="system", content=system, label=False))
+ messages.append(Message(role="user", content=prompt["prompt"], label=False))
+ messages.append(
+ Message(
+ role="assistant", content=prompt["rejected"][1]["content"], label=True
+ )
+ )
+ return MessageList(messages=messages)
+
+ def get_prompt(self, prompt) -> MessageList:
+ """Map the data to extract everything up to the last turn"""
+ total_msg_len = len(prompt["chosen"])
+ total_msg_turns, remainder = divmod(total_msg_len, 2)
+ assert remainder == 0, "invalid number of turns"
+
+ messages: List[Message] = []
+ if system := prompt.get("system", None):
+ messages.append(Message(role="system", content=system, label=False))
+ for i in range(total_msg_turns):
+ if "prompt" in prompt:
+ messages.append(
+ Message(role="user", content=prompt["prompt"], label=False)
+ )
+ else:
+ messages.append(
+ Message(
+ role="user",
+ content=prompt["chosen"][i * 2]["content"],
+ label=False,
+ )
+ )
+ if i < total_msg_turns - 1:
+ messages.append(
+ Message(
+ role="assistant",
+ content=prompt["chosen"][i * 2 + 1]["content"],
+ label=False,
+ )
+ )
+
+ return MessageList(messages=messages)
+
+ def get_chosen(self, prompt) -> MessageList:
+ res = self.get_prompt(prompt)
+ res.messages.append(
+ Message(
+ role="assistant", content=prompt["chosen"][-1]["content"], label=True
+ )
+ )
+ return res
+
+ def get_rejected(self, prompt) -> MessageList:
+ res = self.get_prompt(prompt)
+ res.messages.append(
+ Message(
+ role="assistant", content=prompt["rejected"][-1]["content"], label=True
+ )
+ )
+ return res
+
+
+class ORPOTokenizingStrategy(PromptTokenizingStrategy):
+ """
+ rejected_input_ids
+ input_ids
+ rejected_attention_mask
+ attention_mask
+ rejected_labels
+ labels
+ """
+
+ def __init__(
+ self,
+ *args,
+ dataset_parser=None,
+ **kwargs,
+ ):
+ super().__init__(*args, **kwargs)
+ self.dataset_parser = dataset_parser
+
+ def tokenize_prompt(self, prompt):
+ # pass the rejected prompt/row to the Prompter to get the formatted prompt
+ prompt_len = 0
+ rejected_message_list = self.dataset_parser.get_rejected_conversation_thread(
+ prompt
+ )
+ input_ids = []
+ labels = []
+ for _, (part, label) in enumerate(
+ self.prompter.build_prompt(rejected_message_list)
+ ):
+ if not part:
+ continue
+ _input_ids = self.tokenizer.encode(part, add_special_tokens=False)
+ prev_idx = len(input_ids)
+ input_ids += _input_ids[prev_idx:]
+ if label:
+ labels += input_ids[prev_idx:]
+ else:
+ labels += [IGNORE_INDEX] * (len(input_ids) - prev_idx)
+ prompt_len = len(input_ids)
+ # remap the input_ids, attention_mask and labels
+ rejected_input_ids = input_ids
+ rejected_labels = labels
+ # pass the chosen prompt/row to the Prompter to get the formatted prompt
+ chosen_message_list = self.dataset_parser.get_chosen_conversation_thread(prompt)
+ input_ids = []
+ labels = []
+ for _, (part, label) in enumerate(
+ self.prompter.build_prompt(chosen_message_list)
+ ):
+ if not part:
+ continue
+ _input_ids = self.tokenizer.encode(part, add_special_tokens=False)
+ prev_idx = len(input_ids)
+ input_ids += _input_ids[prev_idx:]
+ if label:
+ labels += input_ids[prev_idx:]
+ else:
+ labels += [IGNORE_INDEX] * (len(input_ids) - prev_idx)
+
+ return {
+ "rejected_input_ids": rejected_input_ids,
+ "rejected_labels": rejected_labels,
+ "rejected_attention_mask": [1] * len(rejected_labels),
+ "input_ids": input_ids,
+ "labels": labels,
+ "attention_mask": [1] * len(labels),
+ "prompt_attention_mask": [1] * prompt_len
+ + [0] * (len(labels) - prompt_len),
+ }
+
+
+class ORPOPrompter(Prompter):
+ """Single Turn prompter for ORPO"""
+
+ def __init__(self, chat_template, tokenizer):
+ self.chat_template = chat_template
+ self.tokenizer = tokenizer
+
+ def build_prompt(
+ self,
+ message_list: MessageList,
+ ) -> Generator[Tuple[str, bool], None, None]:
+ conversation = []
+ for message in message_list.messages:
+ conversation.append(message.model_dump())
+ if message.role == "system":
+ yield self.tokenizer.apply_chat_template(
+ conversation,
+ add_generation_prompt=False,
+ chat_template=self.chat_template,
+ tokenize=False,
+ ), False
+ if message.role == "user":
+ yield self.tokenizer.apply_chat_template(
+ conversation,
+ add_generation_prompt=True,
+ chat_template=self.chat_template,
+ tokenize=False,
+ ), False
+ if message.role == "assistant":
+ yield self.tokenizer.apply_chat_template(
+ conversation,
+ add_generation_prompt=False,
+ chat_template=self.chat_template,
+ tokenize=False,
+ ), True
+
+
+def argilla(cfg, **kwargs): # pylint: disable=possibly-unused-variable,unused-argument
+ dataset_parser = ORPODatasetParsingStrategy()
+
+ chat_template_str = chat_templates(cfg.chat_template)
+
+ def transform_fn(sample, tokenizer=None):
+ res = {}
+
+ res["prompt"] = tokenizer.apply_chat_template(
+ [msg.model_dump() for msg in dataset_parser.get_prompt(sample).messages],
+ add_generation_prompt=True,
+ chat_template=chat_template_str,
+ tokenize=False,
+ )
+ prompt_str_len = len(res["prompt"])
+ res["chosen"] = tokenizer.apply_chat_template(
+ [msg.model_dump() for msg in dataset_parser.get_chosen(sample).messages],
+ add_generation_prompt=False,
+ chat_template=chat_template_str,
+ tokenize=False,
+ )[prompt_str_len:]
+ res["rejected"] = tokenizer.apply_chat_template(
+ [msg.model_dump() for msg in dataset_parser.get_rejected(sample).messages],
+ add_generation_prompt=False,
+ chat_template=chat_template_str,
+ tokenize=False,
+ )[prompt_str_len:]
+
+ return res
+
+ return transform_fn
diff --git a/src/axolotl/prompt_strategies/pretrain.py b/src/axolotl/prompt_strategies/pretrain.py
new file mode 100644
index 0000000000000000000000000000000000000000..8430a7fcabf104b064136099b513ee34bf038797
--- /dev/null
+++ b/src/axolotl/prompt_strategies/pretrain.py
@@ -0,0 +1,60 @@
+"""pretraining prompt strategies"""
+from typing import Generator
+
+from transformers import BatchEncoding
+
+from axolotl.prompt_tokenizers import PromptTokenizingStrategy
+
+
+class PretrainTokenizer:
+ """basic tokenization class for pretraining"""
+
+ def build_prompt(self, prompt) -> Generator[str, None, None]:
+ yield prompt
+
+
+class PretrainTokenizationStrategy(PromptTokenizingStrategy):
+ """handles tokenization for pretraining with strides"""
+
+ @property
+ def supports_batched(self):
+ return True
+
+ def __init__(self, *args, max_length=None, text_column="text", **kwargs):
+ super().__init__(*args, **kwargs)
+ if max_length:
+ self.max_length = max_length
+ self.text_column = text_column
+
+ def _tokenize(
+ self, prompt: str, add_eos_token: bool = True, strip_bos_token: bool = False
+ ) -> BatchEncoding:
+ res = self.tokenizer(
+ prompt,
+ truncation=True,
+ max_length=self.max_length - 1,
+ add_special_tokens=True,
+ return_overflowing_tokens=True,
+ stride=256,
+ )
+ res["input_ids"] = [
+ seq + [self.tokenizer.eos_token_id] for seq in res["input_ids"]
+ ]
+ res["attention_mask"] = [seq + [1] for seq in res["attention_mask"]]
+
+ return res
+
+ def tokenize_prompt(self, prompt):
+ return self._tokenize(prompt[self.text_column])
+
+
+def load(tokenizer, cfg):
+ strat = PretrainTokenizationStrategy(
+ PretrainTokenizer(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ text_column=cfg.pretraining_dataset[0]["text_column"] or "text",
+ max_length=cfg.sequence_len * 64,
+ )
+ return strat
diff --git a/src/axolotl/prompt_strategies/pygmalion.py b/src/axolotl/prompt_strategies/pygmalion.py
new file mode 100644
index 0000000000000000000000000000000000000000..88208f6ec4329eb550344af9048d8d61d0d4d7e9
--- /dev/null
+++ b/src/axolotl/prompt_strategies/pygmalion.py
@@ -0,0 +1,101 @@
+"""Module containing the PygmalionPromptTokenizingStrategy and PygmalionPrompter class"""
+
+import copy
+import logging
+from collections import defaultdict
+from typing import Generator, List, Tuple
+
+from axolotl.prompt_tokenizers import (
+ PromptTokenizingStrategy,
+ parse_tokenized_to_result,
+ tokenize_prompt_default,
+)
+
+LOG = logging.getLogger("axolotl")
+
+IGNORE_TOKEN_ID = -100
+
+
+class PygmalionPromptTokenizingStrategy(PromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Pygmalion.
+ """
+
+ bot_prefix_token_ids: List[int] = []
+
+ def __init__(self, prompter, tokenizer, *args, **kwargs):
+ super().__init__(prompter, tokenizer, *args, **kwargs)
+ res = self._tokenize("<|model|>", add_eos_token=False, strip_bos_token=True)
+ self.bot_prefix_token_ids = res["input_ids"]
+
+ def tokenize_prompt(self, prompt):
+ result, current_len = tokenize_prompt_default()
+ for _, part in enumerate(self.prompter.build_prompt(prompt["conversations"])):
+ role, message = part
+ if role == "system":
+ prefix = "<|system|>"
+ # this should include a bos token, no eos token, strip trailing "\n"
+ if message.endswith("\n"):
+ message = message[:-8]
+ res = self._tokenize(
+ prefix + "Persona: " + message.strip(),
+ add_eos_token=False,
+ strip_bos_token=False,
+ )
+ # everything from this is masked out from the labels
+ labels = [IGNORE_TOKEN_ID] * len(res["input_ids"])
+ elif role == "human":
+ prefix = "<|user|>"
+ res = self._tokenize(
+ prefix + " " + message.strip(),
+ add_eos_token=False,
+ strip_bos_token=True,
+ )
+ # everything from this is masked out from the labels
+ labels = [IGNORE_TOKEN_ID] * len(res["input_ids"])
+ elif role == "bot":
+ prefix = "<|model|>"
+ res = self._tokenize(
+ prefix + " " + message.strip(),
+ add_eos_token=True,
+ strip_bos_token=True,
+ )
+ # mask out the prefix token, rest is not masked out from labels
+ # make sure we create the labels first, otherwise we get incorrect lengths
+ labels = [IGNORE_TOKEN_ID] * len(self.bot_prefix_token_ids) + [
+ *copy.deepcopy(res["input_ids"])
+ ][len(self.bot_prefix_token_ids) :]
+ else:
+ LOG.warning(f"unknown role in conversation: {role}")
+ res = defaultdict(lambda: [])
+
+ # pylint: disable=duplicate-code
+ result, current_len = parse_tokenized_to_result(
+ result,
+ current_len,
+ res,
+ labels,
+ pad_token_id=self.tokenizer.pad_token_id,
+ )
+ return result
+
+
+class PygmalionPrompter:
+ """
+ Prompter for Pygmalion.
+ """
+
+ def __init__(self, *args, **kwargs):
+ pass
+
+ def build_prompt(
+ self, source, *args, **kwargs # pylint: disable=unused-argument
+ ) -> Generator[Tuple[str, str], None, None]:
+ for msg in source:
+ yield msg["role"], msg["value"]
+
+
+def load(tokenizer, cfg):
+ return PygmalionPromptTokenizingStrategy(
+ PygmalionPrompter(), tokenizer, cfg.train_on_inputs, cfg.sequence_len
+ )
diff --git a/src/axolotl/prompt_strategies/sharegpt.py b/src/axolotl/prompt_strategies/sharegpt.py
new file mode 100644
index 0000000000000000000000000000000000000000..b0e37b0c6ea48d33c963b8ef134ddb5416c91a24
--- /dev/null
+++ b/src/axolotl/prompt_strategies/sharegpt.py
@@ -0,0 +1,206 @@
+"""Module containing the SimpleShareGPTPromptTokenizingStrategy class"""
+
+import logging
+from typing import Any, Dict, Optional, Type
+
+from fastchat.conversation import Conversation, SeparatorStyle, register_conv_template
+
+from axolotl.prompt_tokenizers import ShareGPTPromptTokenizingStrategy
+from axolotl.prompters import ShareGPTPrompterV2
+from axolotl.utils.tokenization import (
+ chatml_to_conversation,
+ merge_consecutive_messages,
+)
+
+LOG = logging.getLogger("axolotl")
+
+
+def register_chatml_template(system_message=None):
+ system_message = system_message or "You are a helpful assistant."
+ register_conv_template(
+ Conversation(
+ name="chatml",
+ system_template="<|im_start|>system\n{system_message}",
+ system_message=system_message,
+ roles=("<|im_start|>user", "<|im_start|>assistant"),
+ sep_style=SeparatorStyle.CHATML,
+ sep="<|im_end|>",
+ )
+ )
+ register_conv_template(
+ Conversation(
+ name="chatml_glaive",
+ system_template="<|im_start|>system\n{system_message}",
+ system_message=system_message,
+ roles=("<|im_start|>user", "<|im_start|>assistant", "<|im_start|>tool"),
+ sep_style=SeparatorStyle.CHATML,
+ sep="<|im_end|>",
+ )
+ )
+
+
+def register_llama3_template(system_message=None):
+ system_message = system_message or "You are great at roleplay"
+ register_conv_template(
+ Conversation(
+ name="llama3",
+ system_template="<|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|>",
+ system_message=system_message,
+ roles=("user", "assistant"),
+ sep_style=SeparatorStyle.LLAMA3,
+ sep="",
+ stop_str="<|eot_id|>",
+ stop_token_ids=[128001, 128009],
+ )
+ )
+
+
+def build_loader(
+ tokenization_strategy_cls: Type["ShareGPTPromptTokenizingStrategy"],
+ prompter_cls: Type["ShareGPTPrompterV2"],
+ default_conversation: Optional[str] = None,
+):
+ def _load(tokenizer, cfg, ds_cfg: Optional[Dict[str, Any]] = None):
+ conversation = (
+ ds_cfg["conversation"]
+ if ds_cfg and "conversation" in ds_cfg
+ else default_conversation
+ )
+ field_human = (
+ ds_cfg["field_human"] if ds_cfg and "field_human" in ds_cfg else None
+ )
+ field_model = (
+ ds_cfg["field_model"] if ds_cfg and "field_model" in ds_cfg else None
+ )
+ roles = ds_cfg["roles"].to_dict() if ds_cfg and "roles" in ds_cfg else None
+ strategy = tokenization_strategy_cls(
+ prompter_cls(
+ conversation=conversation,
+ role_key_model=field_model,
+ role_key_human=field_human,
+ roles=roles,
+ ),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ if ds_cfg and "strict" in ds_cfg and hasattr(strategy, "strict"):
+ strategy.strict = ds_cfg["strict"]
+ return strategy
+
+ return _load
+
+
+class SimpleShareGPTPromptTokenizingStrategy(ShareGPTPromptTokenizingStrategy):
+ """
+ basic sharegpt strategy to grab conversations from the sample row
+ """
+
+ _strict = False
+
+ @property
+ def strict(self):
+ return self._strict
+
+ @strict.setter
+ def strict(self, strict):
+ self._strict = strict
+
+ def get_conversation_thread(self, prompt):
+ conversations = prompt["conversations"]
+ if self.strict:
+ return conversations
+ role_key = "from"
+ if "role" in conversations[0].keys():
+ role_key = "role"
+ value_key = "value"
+ if "text" in conversations[0].keys():
+ value_key = "text"
+ elif "content" in conversations[0].keys():
+ value_key = "content"
+ # remap roles - allow for assistant turn"
+ role_map = {
+ "user": "human",
+ "human": "human",
+ "assistant": "gpt",
+ "gpt": "gpt",
+ "system": "system",
+ }
+ turns = [
+ {
+ "from": (
+ role_map[t[role_key]] if t[role_key] in role_map else t[role_key]
+ ),
+ "value": t[value_key],
+ }
+ for t in conversations
+ ]
+ return turns
+
+
+class SimpleRoleShareGPTPromptTokenizingStrategy(
+ SimpleShareGPTPromptTokenizingStrategy
+):
+ """
+ basic sharegpt strategy to grab conversations from the sample row, but uses role instead of from
+ """
+
+ def get_conversation_thread(self, prompt):
+ conversations = prompt["conversations"]
+ # remap role: prompter/assistant, text: ... => from: human/gpt, value: ...
+ turns = [{"from": t["role"], "value": t["value"]} for t in conversations]
+ return turns
+
+
+class GuanacoShareGPTPromptTokenizingStrategy(ShareGPTPromptTokenizingStrategy):
+ """
+ sharegpt strategy that remaps oasst data to sharegpt format
+ """
+
+ def get_conversation_thread(self, prompt):
+ conversations = prompt["conversations"]
+ # remap role: prompter/assistant, text: ... => from: human/gpt, value: ...
+ role_map = {"prompter": "human", "assistant": "gpt"}
+ turns = [
+ {"from": role_map[t["role"]], "value": t["text"]} for t in conversations
+ ]
+ return turns
+
+
+class UltrachatShareGPTPromptTokenizingStrategy(SimpleShareGPTPromptTokenizingStrategy):
+ """
+ sharegpt strategy that remaps ultrachat data to sharegpt format
+ """
+
+ def get_conversation_thread(self, prompt):
+ conversations = prompt["messages"]
+ role_map = {"user": "human", "assistant": "gpt"}
+ turns = [
+ {"from": role_map[t["role"]], "value": t["content"]} for t in conversations
+ ]
+ return turns
+
+
+class GlaiveShareGPTPromptTokenizingStrategy(SimpleShareGPTPromptTokenizingStrategy):
+ """
+ sharegpt strategy that remaps glaive data to sharegpt format
+ """
+
+ def get_conversation_thread(self, prompt):
+ conversation = chatml_to_conversation(prompt)
+ conversation = merge_consecutive_messages(conversation)
+
+ return conversation
+
+
+load = build_loader(SimpleShareGPTPromptTokenizingStrategy, ShareGPTPrompterV2)
+load_role = build_loader(SimpleRoleShareGPTPromptTokenizingStrategy, ShareGPTPrompterV2)
+load_ultrachat = build_loader(
+ UltrachatShareGPTPromptTokenizingStrategy, ShareGPTPrompterV2
+)
+load_guanaco = build_loader(GuanacoShareGPTPromptTokenizingStrategy, ShareGPTPrompterV2)
+load_glaive = build_loader(
+ GlaiveShareGPTPromptTokenizingStrategy,
+ ShareGPTPrompterV2,
+ default_conversation="chatml_glaive",
+)
diff --git a/src/axolotl/prompt_strategies/sharegpt_jokes.py b/src/axolotl/prompt_strategies/sharegpt_jokes.py
new file mode 100644
index 0000000000000000000000000000000000000000..404302c81ea643aa841b6d1639ce6f95090ebfc9
--- /dev/null
+++ b/src/axolotl/prompt_strategies/sharegpt_jokes.py
@@ -0,0 +1,28 @@
+"""Module for Jokes prompts using sharegpt style """
+from axolotl.prompt_tokenizers import ShareGPTPromptTokenizingStrategy
+from axolotl.prompters import ShareGPTPrompterV2
+
+
+def load(tokenizer, cfg):
+ return SimpleJokesShareGPTPromptTokenizingStrategy(
+ ShareGPTPrompterV2(),
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+
+class SimpleJokesShareGPTPromptTokenizingStrategy(ShareGPTPromptTokenizingStrategy):
+ """
+ Tokenization strategy for asking bot to tell a joke and then explain why its funny
+ """
+
+ # title, text, explanation
+ def get_conversation_thread(self, prompt):
+ title = "" if not prompt["title"] else prompt["title"] + " "
+ return [
+ {"from": "human", "value": "Tell me a joke."},
+ {"from": "gpt", "value": title + prompt["text"]},
+ {"from": "human", "value": "Why is that joke funny?"},
+ {"from": "gpt", "value": prompt["explanation"]},
+ ]
diff --git a/src/axolotl/prompt_strategies/user_defined.py b/src/axolotl/prompt_strategies/user_defined.py
new file mode 100644
index 0000000000000000000000000000000000000000..e20e80c3a451fda218686128e3a32fc2a8ade369
--- /dev/null
+++ b/src/axolotl/prompt_strategies/user_defined.py
@@ -0,0 +1,98 @@
+"""
+User Defined prompts with configuration from the YML config
+"""
+
+from dataclasses import dataclass
+from functools import partial
+from typing import Optional, Tuple
+
+from axolotl.prompt_strategies.alpaca_w_system import (
+ InstructionWSystemPromptTokenizingStrategy,
+ SystemDataPrompter,
+)
+
+
+@dataclass
+class UserDefinedDatasetConfig:
+ """
+ dataclass configuration representing a userdefined dataset type
+ """
+
+ system_prompt: str = ""
+ field_system: str = "system"
+ field_instruction: str = "instruction"
+ field_input: str = "input"
+ field_output: str = "output"
+ format: str = "{instruction} {input} "
+ no_input_format: str = "{instruction} "
+ system_format: str = "{system}"
+
+ def __getitem__(self, item):
+ return getattr(self, item)
+
+
+class UserDefinedPromptTokenizationStrategy(InstructionWSystemPromptTokenizingStrategy):
+ """
+ Prompt Tokenization Strategy for user defined prompts
+ """
+
+
+def load(tokenizer, cfg, ds_cfg: Optional[UserDefinedDatasetConfig] = None):
+ if not ds_cfg:
+ raise ValueError("Missing dataset prompt configuration")
+
+ system_prompt = ""
+ if ds_cfg.system_prompt:
+ system_prompt = ds_cfg.system_prompt
+
+ def parse_instruction_fields(
+ field_instruction,
+ field_input,
+ field_output,
+ field_system,
+ system_prompt,
+ prompt,
+ ) -> Tuple[str, str, str, str]:
+ return (
+ prompt[field_instruction],
+ prompt[field_input] if field_input in prompt else "",
+ prompt[field_output] if field_output in prompt else "",
+ prompt[field_system] if field_system in prompt else system_prompt,
+ )
+
+ turn_format = ds_cfg.format
+ turn_no_input_format = ds_cfg.no_input_format
+ system_format = ds_cfg.system_format
+
+ class UserDefinedPrompter(SystemDataPrompter):
+ """
+ Prompter for user defined prompts
+ """
+
+ def match_prompt_style(self):
+ self.turn_format = turn_format
+ self.turn_no_input_format = turn_no_input_format
+ self.system_format = system_format
+
+ prompter = UserDefinedPrompter()
+
+ strat = UserDefinedPromptTokenizationStrategy(
+ prompter,
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+
+ setattr(
+ strat,
+ "parse_instruction_fields",
+ partial(
+ parse_instruction_fields,
+ ds_cfg.field_instruction,
+ ds_cfg.field_input,
+ ds_cfg.field_output,
+ ds_cfg.field_system,
+ system_prompt,
+ ),
+ )
+ return strat
diff --git a/src/axolotl/prompt_tokenizers.py b/src/axolotl/prompt_tokenizers.py
new file mode 100644
index 0000000000000000000000000000000000000000..bb13cf76dda7a10526ea5f1e57f5491c26087ff0
--- /dev/null
+++ b/src/axolotl/prompt_tokenizers.py
@@ -0,0 +1,501 @@
+"""Module containing PromptTokenizingStrategy and Prompter classes"""
+
+import abc
+import copy
+import logging
+from typing import Dict, List, Tuple, Union
+
+from fastchat.conversation import Conversation
+from transformers import BatchEncoding, PreTrainedTokenizer
+
+from axolotl.monkeypatch.fastchat_conversation_turns import (
+ add_get_turns_to_conversation,
+)
+from axolotl.prompters import IGNORE_TOKEN_ID, Prompter
+
+LOG = logging.getLogger("axolotl")
+
+IGNORE_INDEX = -100
+LLAMA_DEFAULT_PAD_TOKEN = "" # nosec
+LLAMA_DEFAULT_EOS_TOKEN = " " # nosec
+LLAMA_DEFAULT_BOS_TOKEN = "" # nosec
+LLAMA_DEFAULT_UNK_TOKEN = "" # nosec
+
+add_get_turns_to_conversation()
+
+
+class InvalidDataException(Exception):
+ """
+ Exception raised when the data is invalid
+ """
+
+
+class PromptTokenizingStrategy(abc.ABC):
+ """
+ Abstract class for tokenizing strategies
+ """
+
+ def __init__(
+ self,
+ prompter: Prompter,
+ tokenizer,
+ train_on_inputs: bool = False,
+ sequence_len: int = 2048,
+ ):
+ self.prompter = prompter
+ self.tokenizer: PreTrainedTokenizer = tokenizer
+ self.train_on_inputs = train_on_inputs
+ # sequence_len and max_length can be different for CompletionPromptTokenizingStrategy.
+ # TODO: Document how they are different.
+ self.sequence_len = sequence_len
+ self.max_length = sequence_len
+
+ @abc.abstractmethod
+ def tokenize_prompt(self, prompt):
+ pass
+
+ @property
+ def supports_batched(self):
+ return False
+
+ def _tokenize(
+ self, prompt: str, add_eos_token: bool = True, strip_bos_token: bool = False
+ ) -> BatchEncoding:
+ empty = BatchEncoding(data={"input_ids": [], "attention_mask": []})
+ if not prompt:
+ LOG.warning("Empty text requested for tokenization.")
+ return empty
+
+ result = self.tokenizer(
+ prompt,
+ truncation=True,
+ max_length=self.max_length,
+ padding=False,
+ return_tensors=None,
+ )
+ if len(result["input_ids"]) == 0:
+ LOG.warning("Tokenizer result is empty. You may want to audit your dataset")
+ return empty
+
+ if (
+ result["input_ids"][-1] != self.tokenizer.eos_token_id
+ and len(result["input_ids"]) < self.max_length
+ and add_eos_token
+ ):
+ result["input_ids"].append(self.tokenizer.eos_token_id)
+ result["attention_mask"].append(1)
+
+ if result["input_ids"][0] == self.tokenizer.bos_token_id and strip_bos_token:
+ result["input_ids"] = result["input_ids"][1:]
+ result["attention_mask"] = result["attention_mask"][1:]
+
+ result["labels"] = result["input_ids"].copy()
+ return result
+
+
+class InstructionPromptTokenizingStrategy(PromptTokenizingStrategy):
+ """
+ Tokenizing strategy for instruction-based prompts.
+ """
+
+ def parse_instruction_fields(
+ self, prompt
+ ) -> Union[Tuple[str, str, str], Tuple[str, str, str, str]]:
+ raise NotImplementedError
+
+ def tokenize_prompt(self, prompt):
+ (
+ instruction,
+ input, # pylint: disable=redefined-builtin
+ response,
+ ) = self.parse_instruction_fields(prompt)
+ user_prompt = next(
+ iter(
+ self.prompter.build_prompt(
+ instruction,
+ input,
+ )
+ )
+ )
+ tokenized_prompt = self._tokenize(user_prompt, add_eos_token=False)
+ if not self.train_on_inputs:
+ user_prompt_len = len(tokenized_prompt["input_ids"])
+ # TODO this could be sped up using numpy array slicing
+ tokenized_prompt["labels"] = [IGNORE_INDEX] * user_prompt_len
+ tokenized_res_prompt = self._tokenize(
+ response, strip_bos_token=True, add_eos_token=True
+ )
+ tokenized_prompt["input_ids"] += tokenized_res_prompt["input_ids"]
+ tokenized_prompt["attention_mask"] += tokenized_res_prompt["attention_mask"]
+ tokenized_prompt["labels"] += tokenized_res_prompt["input_ids"]
+
+ return tokenized_prompt
+
+ def _build_full_prompt(
+ self, instruction, input, response # pylint: disable=redefined-builtin
+ ):
+ return next(
+ iter(
+ self.prompter.build_prompt(
+ instruction,
+ input,
+ response,
+ )
+ )
+ )
+
+
+class AlpacaPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Alpaca prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["instruction"],
+ prompt["input"] if "input" in prompt else "",
+ prompt["output"],
+ )
+
+
+class AlpacaMultipleChoicePromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Alpaca Multiple Choice prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["question"],
+ "\n".join(f'- "{choice}"' for choice in prompt["choices"]),
+ prompt["solution"] if "solution" in prompt else prompt["explanation"],
+ )
+
+
+class JeopardyPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Jeopardy prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["question"],
+ prompt["category"],
+ "what is " + prompt["answer"],
+ )
+
+
+class OpenAssistantPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for OpenAssistant prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["INSTRUCTION"],
+ "",
+ prompt["RESPONSE"],
+ )
+
+
+class SummarizeTLDRPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for SummarizeTLDR prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["article"],
+ "",
+ prompt["summary"],
+ )
+
+
+class GPTeacherPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for GPTeacher prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["instruction"],
+ prompt["input"] if "input" in prompt else "",
+ prompt["response"],
+ )
+
+
+class NomicGPT4AllPromptTokenizingStrategy(InstructionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for NomicGPT4All prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str]:
+ return (
+ prompt["prompt"],
+ "",
+ prompt["response"],
+ )
+
+
+class ReflectionPromptTokenizingStrategy(PromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Reflection prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str, str, str]:
+ raise NotImplementedError
+
+ def tokenize_prompt(self, prompt):
+ # pylint: disable=duplicate-code
+ (
+ instruction,
+ input, # pylint: disable=redefined-builtin
+ output,
+ reflection,
+ corrected,
+ ) = self.parse_instruction_fields(prompt)
+ full_prompt = self._build_full_prompt(
+ instruction, input, output, reflection, corrected
+ )
+ tokenized_full_prompt = self._tokenize(full_prompt)
+ if not self.train_on_inputs:
+ user_prompt = next(
+ iter(
+ self.prompter.build_prompt(
+ instruction,
+ input,
+ )
+ )
+ )
+ tokenized_user_prompt = self._tokenize(user_prompt, add_eos_token=False)
+ user_prompt_len = len(tokenized_user_prompt["input_ids"])
+ # TODO this could be sped up using numpy array slicing
+ tokenized_full_prompt["labels"] = [
+ IGNORE_INDEX
+ ] * user_prompt_len + tokenized_full_prompt["labels"][user_prompt_len:]
+
+ return tokenized_full_prompt
+
+ def _build_full_prompt(
+ self, instruction, input, output, reflection, corrected
+ ): # pylint: disable=redefined-builtin
+ return next(
+ iter(
+ self.prompter.build_prompt(
+ instruction,
+ input,
+ output,
+ reflection,
+ corrected,
+ )
+ )
+ )
+
+ def _tokenize(self, prompt, add_eos_token=True, strip_bos_token=False):
+ result = self.tokenizer(
+ prompt,
+ truncation=True,
+ max_length=self.sequence_len,
+ padding=False,
+ return_tensors=None,
+ )
+ if (
+ result["input_ids"][-1] != self.tokenizer.eos_token_id
+ and len(result["input_ids"]) < self.sequence_len
+ and add_eos_token
+ ):
+ result["input_ids"].append(self.tokenizer.eos_token_id)
+ result["attention_mask"].append(1)
+
+ result["labels"] = result["input_ids"].copy()
+ return result
+
+
+class AlpacaReflectionPTStrategy(ReflectionPromptTokenizingStrategy):
+ """
+ Tokenizing strategy for Alpaca Reflection prompts.
+ """
+
+ def parse_instruction_fields(self, prompt) -> Tuple[str, str, str, str, str]:
+ return (
+ prompt["instruction"],
+ prompt["input"] if "input" in prompt else "",
+ prompt["output"],
+ prompt["reflection"],
+ prompt["corrected"],
+ )
+
+
+class ShareGPTPromptTokenizingStrategy(PromptTokenizingStrategy):
+ """
+ Tokenizing strategy for ShareGPT prompts.
+ """
+
+ def get_conversation_thread(self, prompt):
+ return prompt["conversations"]
+
+ def tokenize_prompt(self, prompt):
+ # Initial values. We will append to these as we go through the conversation.
+ result, current_len = tokenize_prompt_default()
+ conversation: Conversation = (
+ self.prompter._conversation.copy() # pylint: disable=protected-access
+ )
+
+ input_roles = {conversation.roles[0]}
+ output_roles = {conversation.roles[1]}
+
+ if len(conversation.roles) == 3:
+ tool_role_label = conversation.roles[2]
+ input_roles.add(tool_role_label)
+
+ # Add roles from the config
+ if self.prompter.roles:
+ if "input" in self.prompter.roles and self.prompter.roles["input"]:
+ for role in self.prompter.roles["input"]:
+ input_roles.add(role)
+
+ if "output" in self.prompter.roles and self.prompter.roles["output"]:
+ for role in self.prompter.roles["output"]:
+ output_roles.add(role)
+
+ # support for custom roles from the dataset, only useful for vicuna style prompts/roles
+ role_remap = []
+ if (
+ conversation.name == "vicuna_v1.1"
+ and "roles" in prompt
+ and len(prompt["roles"]) >= 2
+ ):
+ role_remap = [
+ {"from": conversation.roles[0], "to": prompt["roles"][0]},
+ {"from": conversation.roles[1], "to": prompt["roles"][1]},
+ ]
+
+ try:
+ for _, part in enumerate(
+ self.prompter.build_prompt(self.get_conversation_thread(prompt))
+ ):
+ if not isinstance(part, tuple):
+ LOG.warning(f"expected tuple, got {part}")
+ continue
+
+ role, content = part
+
+ # Uses "in" because role contains extra characters
+ input_turn = any(r.lower() in role.lower() for r in input_roles)
+ output_turn = any(r.lower() in role.lower() for r in output_roles)
+ empty_role = role.strip() == ""
+
+ if not any([input_turn, output_turn, empty_role]):
+ LOG.warning(f"unhandled role: {role}")
+ continue
+
+ if input_turn:
+ role = (
+ role.replace(role_remap[0]["from"], role_remap[0]["to"])
+ if role_remap
+ else role
+ )
+ turn = role + content
+ # this is still the user query, we should
+ if not content.strip():
+ LOG.warning(f"user turn has empty text: {prompt}")
+ res = self._tokenize(
+ turn,
+ add_eos_token=False,
+ strip_bos_token=True,
+ )
+ if self.train_on_inputs:
+ labels = copy.deepcopy(res["input_ids"])
+ else:
+ # everything from this is masked out from the labels
+ labels = [IGNORE_TOKEN_ID] * len(res["input_ids"])
+ elif output_turn:
+ role = (
+ role.replace(role_remap[1]["from"], role_remap[1]["to"])
+ if role_remap
+ else role
+ )
+ turn = role + content
+ # this should be the assistant response, should end with an eos token
+ if not content.strip():
+ LOG.warning(f"assistant turn has empty text: {prompt}")
+ add_eos_token = not (
+ conversation.name == "chatml"
+ and conversation.sep == self.tokenizer.eos_token
+ )
+ res = self._tokenize(
+ turn,
+ add_eos_token=add_eos_token,
+ strip_bos_token=True,
+ )
+ role_res = self._tokenize(
+ role.rstrip(),
+ add_eos_token=False,
+ strip_bos_token=True,
+ )
+ labels = copy.deepcopy(res["input_ids"])
+ if not self.train_on_inputs:
+ # mask out role tokens from the labels
+ len_role = len(role_res["input_ids"])
+ labels[:len_role] = [IGNORE_TOKEN_ID] * min(
+ len_role, len(labels)
+ )
+ elif empty_role:
+ turn = content
+ # this is only ever the first part, should include the bos token and the user query
+ res = self._tokenize(
+ turn, add_eos_token=False, strip_bos_token=False
+ )
+ if self.train_on_inputs:
+ labels = copy.deepcopy(res["input_ids"])
+ else:
+ # everything from this is masked out from the labels
+ labels = [IGNORE_TOKEN_ID] * len(res["input_ids"])
+
+ # pylint: disable=duplicate-code
+ result, current_len = parse_tokenized_to_result(
+ result,
+ current_len,
+ res,
+ labels,
+ pad_token_id=self.tokenizer.pad_token_id,
+ )
+ return result
+ except (KeyError, AssertionError, IndexError) as err:
+ raise InvalidDataException(str(err)) from err
+
+
+def tokenize_prompt_default() -> Tuple[Dict[str, List[int]], int]:
+ """
+ Returns the default values for the tokenize prompt function
+ """
+
+ result: Dict[str, List[int]] = {
+ "input_ids": [],
+ "attention_mask": [],
+ "labels": [],
+ }
+ current_len = 0
+ return result, current_len
+
+
+def parse_tokenized_to_result(
+ result: Dict[str, List[int]],
+ current_len: int,
+ res: Dict[str, List[int]],
+ labels: List[int],
+ pad_token_id: Union[int, None] = None,
+) -> Tuple[Dict[str, List[int]], int]:
+ """
+ Parses the tokenized prompt and append the tokenized input_ids, attention_mask and labels to the result
+ """
+
+ input_ids = res["input_ids"]
+ input_len = len(input_ids)
+ result["input_ids"][current_len : current_len + input_len] = input_ids
+ result["attention_mask"][current_len : current_len + input_len] = [
+ 1 if x != pad_token_id else 0 for x in input_ids
+ ]
+ result["labels"][current_len : current_len + input_len] = labels
+ current_len += input_len
+
+ return result, current_len
diff --git a/src/axolotl/prompters.py b/src/axolotl/prompters.py
new file mode 100644
index 0000000000000000000000000000000000000000..60ea5c99f9ee9e7e3b63a4968e7ccf91c4354f94
--- /dev/null
+++ b/src/axolotl/prompters.py
@@ -0,0 +1,403 @@
+"""Module containing prompters"""
+
+import logging
+from enum import Enum
+from typing import Generator, Optional, Union
+
+from colorama import Fore
+from fastchat.conversation import Conversation, get_conv_template
+
+LOG = logging.getLogger("axolotl")
+IGNORE_TOKEN_ID = -100
+REPR_TEMPLATE = "\n\n" + Fore.CYAN + "{full_prompt}" + Fore.RESET + "\n\n"
+
+
+class PromptStyle(Enum):
+ """
+ Enum for prompt styles
+ """
+
+ INSTRUCT = "instruct"
+ CHAT = "chat"
+ CHATML = "chatml"
+
+
+class Prompter:
+ """
+ Base prompter class for all prompters
+ """
+
+
+class AlpacaPrompter(Prompter):
+ """
+ Base class for alpaca prompters
+ """
+
+ system_prompt = "Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request."
+ system_no_input_prompt = "Below is an instruction that describes a task. Write a response that appropriately completes the request."
+ system_format: str = "{system}"
+ turn_format: str
+ turn_no_input_format: str
+ prompt_style: Optional[PromptStyle] = None
+
+ def __init__(self, prompt_style=PromptStyle.INSTRUCT.value):
+ self.prompt_style = prompt_style if prompt_style else PromptStyle.INSTRUCT.value
+ self.match_prompt_style()
+
+ def match_prompt_style(self):
+ # pylint: disable=duplicate-code
+ if self.prompt_style == PromptStyle.INSTRUCT.value:
+ self.turn_format = "### Instruction:\n{instruction}\n\n### Input:\n{input}\n\n### Response:\n"
+ self.turn_no_input_format = (
+ "### Instruction:\n{instruction}\n\n### Response:\n"
+ )
+ self.system_format = "{system}\n\n"
+ if self.prompt_style == PromptStyle.CHAT.value:
+ self.turn_format = "USER: {instruction}\n{input}\nASSISTANT:"
+ self.turn_no_input_format = "USER: {instruction}\nASSISTANT:"
+ self.system_format = "SYSTEM: {system}\n"
+ if self.prompt_style == PromptStyle.CHATML.value:
+ self.turn_format = "<|im_start|>user\n{instruction}\n{input}<|im_end|>\n<|im_start|>assistant\n"
+ self.turn_no_input_format = (
+ "<|im_start|>user\n{instruction}<|im_end|>\n<|im_start|>assistant\n"
+ )
+ self.system_format = "<|im_start|>system\n{system}<|im_end|>\n"
+
+ def _build_result(self, instruction, input_text, output):
+ # returns the full prompt from instruction and optional input
+ # if a label (=response, =output) is provided, it's also appended.
+ if input_text:
+ res = (
+ self.system_format.format(system=self.system_prompt)
+ if self.system_prompt
+ else ""
+ ) + self.turn_format.format(instruction=instruction, input=input_text)
+ else:
+ res = (
+ self.system_format.format(system=self.system_no_input_prompt)
+ if self.system_no_input_prompt
+ else ""
+ ) + self.turn_no_input_format.format(instruction=instruction)
+ if output:
+ res = f"{res}{output}"
+
+ return res
+
+ def build_prompt(
+ self,
+ instruction: str,
+ input: Union[None, str] = None, # pylint: disable=redefined-builtin
+ output: Union[None, str] = None,
+ ) -> Generator[str, None, None]:
+ yield self._build_result(instruction, input, output)
+
+ def __repr__(self) -> str:
+ return REPR_TEMPLATE.format(
+ full_prompt=self._build_result("{instruction}", "{input}", "{output}")
+ )
+
+
+class UnpromptedPrompter(AlpacaPrompter):
+ """
+ Prompter for alpaca no system prompt
+ """
+
+ system_prompt = ""
+ system_no_input_prompt = ""
+
+
+class JeopardyPrompter(AlpacaPrompter):
+ """
+ Prompter for Jeopardy
+ """
+
+ prompt_input = "Below is a Jeopardy clue paired with input providing the category of the clue. Write a concise response that best answers tbe clue given the category.\n\n### Instruction:\n{instruction}\n\n### Input:\n{input}\n\n### Response:\n"
+
+
+class MultipleChoiceExplainPrompter(AlpacaPrompter):
+ """
+ Prompter for multiple choice explain
+ """
+
+ system_prompt = (
+ "Choose the answer that best answers the question. Explain your reasoning.\n"
+ )
+ system_no_input_prompt = (
+ "Choose the answer that best answers the question. Explain your reasoning.\n"
+ )
+
+
+class MultipleChoiceConcisePrompter(AlpacaPrompter):
+ """
+ Prompter for multiple choice concise
+ """
+
+ system_prompt = "Choose the answer that best answers the question. Be concise in your response.\n\n"
+ system_no_input_prompt = "Choose the answer that best answers the question. Be concise in your response.\n\n"
+
+ def match_prompt_style(self):
+ self.turn_format = "USER: {instruction}\n{input}\nASSISTANT:"
+ self.turn_no_input_format = "USER: {instruction}\nASSISTANT:"
+
+
+class SummarizeTLDRPrompter(AlpacaPrompter):
+ """
+ Prompter for summarize TLDR
+ """
+
+ system_prompt = ""
+ system_no_input_prompt = ""
+
+ def match_prompt_style(self):
+ self.turn_format = "USER: Summarize the following article as a TL;DR.\n{instruction}\n{input}\nASSISTANT:"
+ self.turn_no_input_format = "USER: Summarize the following article as a TL;DR.\n{instruction}\nASSISTANT:"
+
+
+class GPTeacherPrompter(AlpacaPrompter):
+ """
+ Prompter for GPTeacher
+ """
+
+
+class NomicGPT4AllPrompter(AlpacaPrompter):
+ """
+ Prompter for NomicGPT4All
+ """
+
+
+class ReflectAlpacaPrompter(Prompter):
+ """
+ Prompter for ReflectAlpaca
+ """
+
+ system_prompt = "Below is an instruction that describes a task, paired with an input that provides further context. You, the Assistant, should generate a response as if it were an abstract for an academic or technical paper on the query along with a methodology. Then generate an Agent Reflection where you create a long form response as if from subject matter expert, be verbose, diligent, and creative in your application of knowledge, apply it through the lens of the response generated by the assistant. Look for flawed reasoning, faulty logic, or other mistakes in the method. Finally, generate a final response and method for the user with the Assistant abstract and Reflection analysis as augmentations to the generation\n\n"
+ system_no_input_prompt = "Below is an instruction that describes a task. You, the Assistant, should generate a response as if it were an abstract for an academic or technical paper on the query along with a methodology. Then generate an Agent Reflection where you create a long form response as if from subject matter expert, be verbose, diligent, and creative in your application of knowledge, apply it through the lens of the response generated by the assistant. Look for flawed reasoning, faulty logic, or other mistakes in the method. Finally, generate a final response and method for the user with the Assistant abstract and Reflection analysis as augmentations to the generation\n\n"
+
+ prompt_input = (
+ "### Instruction:\n{instruction}\n\n### Input:\n{input}\n\n### Response:\n"
+ )
+ prompt_no_input = "### Instruction:\n{instruction}\n\n### Response:\n"
+ agent_label = "### Thought:\n{output}\n\n### Agent Reflection:\n{reflection}\n\n### Final Response:\n{corrected}"
+ response_split = "### Response:"
+
+ def __init__(self, prompt_style="instruct"):
+ self.prompt_style = prompt_style
+ self.match_prompt_style()
+
+ def match_prompt_style(self):
+ if self.prompt_style == PromptStyle.INSTRUCT.value:
+ self.prompt_input = (
+ self.system_prompt
+ + "### Instruction:\n{instruction}\n\n### Input:\n{input}\n\n### Response:\n"
+ )
+ self.prompt_no_input = (
+ self.system_no_input_prompt
+ + "### Instruction:\n{instruction}\n\n### Response:\n"
+ )
+ self.agent_label = "### Thought:\n{output}\n\n### Agent Reflection:\n{reflection}\n\n### Final Response:\n{corrected}"
+ self.response_split = "### Final Response:"
+ if self.prompt_style == PromptStyle.CHAT.value:
+ self.prompt_input = (
+ self.system_prompt + "USER: {instruction}\n{input}\nASSISTANT:"
+ )
+ self.prompt_no_input = (
+ self.system_no_input_prompt + "USER: {instruction}\nASSISTANT:"
+ )
+ self.agent_label = (
+ "\nTHOUGHT: {output}\nASSISTANT REFLECTION: {reflection}\nASSISTANT:"
+ )
+ self.response_split = "ASSISTANT:"
+
+ def _build_result(
+ self,
+ instruction: str,
+ input: Union[None, str] = None, # pylint: disable=redefined-builtin
+ output: Union[None, str] = None,
+ reflection: Union[None, str] = None,
+ corrected: Union[None, str] = None,
+ ):
+ # returns the full prompt from instruction and optional input
+ # if a label (=response, =output) is provided, it's also appended.
+ if input:
+ res = self.prompt_input.format(instruction=instruction, input=input)
+ else:
+ res = self.prompt_no_input.format(instruction=instruction)
+ if output and reflection and corrected:
+ label = self.agent_label.format(
+ output=output,
+ reflection=reflection,
+ corrected=corrected,
+ )
+ res = f"{res}{label}"
+
+ return res
+
+ def build_prompt(
+ self,
+ instruction: str,
+ input: Union[None, str] = None, # pylint: disable=redefined-builtin
+ output: Union[None, str] = None,
+ reflection: Union[None, str] = None,
+ corrected: Union[None, str] = None,
+ ) -> Generator[str, None, None]:
+ # pylint: disable=duplicate-code
+ yield self._build_result(
+ instruction,
+ input,
+ output,
+ reflection,
+ corrected,
+ )
+
+ def __repr__(self) -> str:
+ return REPR_TEMPLATE.format(
+ full_prompt=self._build_result("{instruction}", "{input}", "{output}")
+ )
+
+
+SHAREGPT_ASSERTION_FAILED_ROLE = (
+ "Role did not alternate between turns (gpt and human). Please check your data."
+)
+
+CONVERSATION_ROLE_FORMAT = {
+ "chatml": "<|im_start|>{ROLE}",
+ "zephyr": "<|{ROLE}|>",
+ "vicuna_v1.1": "{ROLE}",
+ "llama3": "<|start_header_id|>{ROLE}<|end_header_id|>",
+}
+
+
+class ShareGPTPrompter(Prompter): # pylint: disable=too-few-public-methods
+ """
+ A prompter that generates prompts for the ShareGPT
+ """
+
+ role_key_human = "human"
+ role_key_model = "gpt"
+ # Optional, only used for tool usage datasets.
+ role_key_tool: Optional[str] = None
+ # Optional, role input/output mapping
+ roles: Optional[dict] = None
+
+ def __init__(
+ self,
+ prompt_style=None, # pylint: disable=unused-argument
+ conversation: Optional[Union[str, Conversation]] = None,
+ role_key_human: Optional[str] = None,
+ role_key_model: Optional[str] = None,
+ role_key_tool: Optional[str] = None,
+ roles: Optional[dict] = None,
+ ):
+ if conversation:
+ if isinstance(conversation, Conversation):
+ self._conversation = conversation
+ else:
+ self._conversation = get_conv_template(conversation)
+ else:
+ self._conversation = get_conv_template("vicuna_v1.1")
+ if role_key_human:
+ self.role_key_human = role_key_human
+ if role_key_model:
+ self.role_key_model = role_key_model
+ if role_key_tool:
+ self.role_key_tool = role_key_tool
+ if roles:
+ self.roles = roles
+
+ def _build_result(self, source):
+ if len(source) < 2:
+ # If there isn't a back and forth conversation, ignore it
+ # also happens on the data splitting leaving empty conversations
+ raise IndexError(
+ f"A conversation entry has less than 2 messages :\n{source}"
+ )
+
+ conv = self._conversation.copy()
+
+ # Add the conversation system prompt if provided, otherwise use the default one
+ if source[0]["from"] == "system":
+ conv.set_system_message(source[0]["value"])
+ source.pop(0)
+
+ roles = {self.role_key_human: conv.roles[0], self.role_key_model: conv.roles[1]}
+ if self.role_key_tool:
+ roles[self.role_key_tool] = conv.roles[2]
+
+ try:
+ # Apply prompt templates
+ if source[0]["from"] not in roles:
+ # Skip the first one if it is not from human
+ source = source[1:]
+ except IndexError as err:
+ # sometimes there is a bing or system chat
+ raise err
+
+ conv.messages = []
+ for _, sentence in enumerate(source):
+ from_role = sentence["from"]
+ if from_role in roles:
+ role = roles[from_role]
+ else:
+ if self._conversation.name not in CONVERSATION_ROLE_FORMAT:
+ raise NotImplementedError(
+ f"Role ({role}) not in default roles, and {self._conversation.name} does not support role remapping yet."
+ "Please help us by creating an Issue to add support for this conversation type."
+ )
+
+ role = CONVERSATION_ROLE_FORMAT[self._conversation.name].format(
+ ROLE=from_role
+ )
+
+ if len(conv.messages) > 0 and ((role == conv.messages[-1][0])):
+ if (
+ role != "assistant"
+ ): # back to back assistant calls may be okay for tool calls
+ LOG.warning(f"{SHAREGPT_ASSERTION_FAILED_ROLE}: {sentence}")
+
+ conv.append_message(role, sentence["value"])
+
+ return conv.get_turns()
+
+ def build_prompt(self, source) -> Generator[str, None, None]:
+ turns = self._build_result(source)
+
+ for part in turns:
+ if part[0] and not part[1]:
+ LOG.warning(f"role with empty message: {part[0]}")
+ yield part
+
+ def __repr__(self) -> str:
+ turns = self._build_result([{"from": "{from}", "value": "{value}"}])
+ return "\n".join([REPR_TEMPLATE.format(full_prompt=part) for part in turns])
+
+
+class ShareGPTPrompterV2(ShareGPTPrompter):
+ """
+ A V2 prompter that generates prompts for the ShareGPT
+ """
+
+ def __init__(
+ self,
+ conversation: Optional[Union[str, Conversation]] = None,
+ role_key_human: Optional[str] = None,
+ role_key_model: Optional[str] = None,
+ roles: Optional[dict] = None,
+ ):
+ super().__init__(
+ conversation=conversation,
+ role_key_human=role_key_human,
+ role_key_model=role_key_model,
+ roles=roles,
+ )
+
+
+class UnsupportedPrompter(Prompter):
+ """
+ A dummy class for custom prompters
+ """
+
+ def __init__(self) -> None:
+ pass
+
+ def __repr__(self):
+ return "Pre-tokenized or custom dataset types are unsupported for logging"
diff --git a/src/axolotl/train.py b/src/axolotl/train.py
new file mode 100644
index 0000000000000000000000000000000000000000..32bcbc1d0a13fedc158c9b2f39f57f4f77ac30b2
--- /dev/null
+++ b/src/axolotl/train.py
@@ -0,0 +1,248 @@
+"""Prepare and train a model on a dataset. Can also infer from a model or merge lora"""
+
+import os
+import signal
+import sys
+import weakref
+from dataclasses import dataclass
+from pathlib import Path
+from typing import Optional, Tuple, Union
+
+import torch
+import transformers.modelcard
+from accelerate import Accelerator
+from accelerate.logging import get_logger
+from datasets import Dataset
+from peft import PeftModel
+from pkg_resources import get_distribution # type: ignore
+from transformers import PreTrainedModel, PreTrainedTokenizer
+from transformers.integrations.deepspeed import is_deepspeed_zero3_enabled
+
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.logging_config import configure_logging
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.freeze import freeze_layers_except
+from axolotl.utils.models import load_model, load_tokenizer
+from axolotl.utils.trainer import setup_trainer
+
+try:
+ from optimum.bettertransformer import BetterTransformer
+except ImportError:
+ BetterTransformer = None
+
+project_root = os.path.abspath(os.path.join(os.path.dirname(__file__), ".."))
+src_dir = os.path.join(project_root, "src")
+sys.path.insert(0, src_dir)
+
+configure_logging()
+LOG = get_logger("axolotl.train")
+
+
+@dataclass
+class TrainDatasetMeta:
+ """
+ dataclass to capture the dataset specific options for training
+ """
+
+ train_dataset: Dataset
+ eval_dataset: Optional[Dataset] = None
+ total_num_steps: Optional[int] = None
+
+
+def train(
+ *, cfg: DictDefault, cli_args: TrainerCliArgs, dataset_meta: TrainDatasetMeta
+) -> Tuple[Union[PeftModel, PreTrainedModel], PreTrainedTokenizer]:
+ # load the tokenizer first
+ LOG.debug(
+ f"loading tokenizer... {cfg.tokenizer_config or cfg.base_model_config}",
+ main_process_only=True,
+ )
+ tokenizer = load_tokenizer(cfg)
+
+ train_dataset = dataset_meta.train_dataset
+ eval_dataset = dataset_meta.eval_dataset
+ total_num_steps = dataset_meta.total_num_steps
+
+ if cfg.resume_from_checkpoint is None and cfg.auto_resume_from_checkpoints:
+ possible_checkpoints = [
+ str(cp) for cp in Path(cfg.output_dir).glob("checkpoint-*")
+ ]
+ if len(possible_checkpoints) > 0:
+ sorted_paths = sorted(
+ possible_checkpoints,
+ key=lambda path: int(path.split("-")[-1]),
+ )
+ cfg.resume_from_checkpoint = sorted_paths[-1]
+ LOG.info(
+ f"Using Auto-resume functionality to start with checkpoint at {cfg.resume_from_checkpoint}"
+ )
+ resume_from_checkpoint = cfg.resume_from_checkpoint
+
+ # Load the model and tokenizer
+ msg = "loading model"
+ if cfg.adapter:
+ msg += " and peft_config..."
+ LOG.debug(msg)
+ # we wait unitl the last possible moment to setup Accelerator
+ Accelerator()
+ model, peft_config = load_model(cfg, tokenizer, inference=cli_args.inference)
+ model.generation_config.do_sample = True
+
+ model_ref = None
+ if cfg.rl and cfg.rl != "orpo":
+ if cfg.adapter and not cfg.rl_adapter_ref_model:
+ # use built-in trl autounwrap
+ LOG.debug("Passing model_ref: None to RL trainer")
+ model_ref = None # explicit setting to None
+ else:
+ # load the model again for model_ref/baseline
+ model_ref, _ = load_model(
+ cfg, tokenizer, inference=cli_args.inference, reference_model=True
+ )
+
+ safe_serialization = cfg.save_safetensors is True
+
+ if cfg.unfrozen_parameters:
+ freeze_layers_except(model, cfg.unfrozen_parameters)
+
+ trainer = setup_trainer(
+ cfg,
+ train_dataset,
+ eval_dataset,
+ (model, model_ref, peft_config),
+ tokenizer,
+ total_num_steps,
+ )
+
+ # go ahead and presave, so we have the adapter config available to inspect
+ if peft_config:
+ LOG.info(f"Pre-saving adapter config to {cfg.output_dir}")
+ peft_config.save_pretrained(cfg.output_dir)
+ # additionally presave the tokenizer and model configs
+ if not Path(cfg.output_dir).is_dir():
+ os.makedirs(cfg.output_dir, exist_ok=True)
+ tokenizer.save_pretrained(str(Path(cfg.output_dir)))
+ if hasattr(model, "config"):
+ model.config.save_pretrained(str(Path(cfg.output_dir)))
+
+ # In case we want to stop early with ctrl+c, this is a nice to have to save the pretrained model
+ if cfg.local_rank == 0:
+
+ def terminate_handler(_, __, model_weakref):
+ if model_weakref() is not None:
+ _model = model_weakref()
+ if cfg.flash_optimum and BetterTransformer:
+ _model = BetterTransformer.reverse(_model)
+ _model.save_pretrained(
+ cfg.output_dir, safe_serialization=safe_serialization
+ )
+ sys.exit(0)
+
+ _model_weakref = weakref.ref(model)
+ signal.signal(
+ signal.SIGINT,
+ lambda signum, frame: terminate_handler(signum, frame, _model_weakref),
+ )
+
+ badge_markdown = """[ ](https://github.com/OpenAccess-AI-Collective/axolotl)"""
+ transformers.modelcard.AUTOGENERATED_TRAINER_COMMENT += f"\n{badge_markdown}"
+
+ if getattr(cfg, "axolotl_config_path"):
+ raw_axolotl_cfg = Path(cfg.axolotl_config_path)
+ version = get_distribution("axolotl").version
+ if raw_axolotl_cfg.is_file():
+ transformers.modelcard.AUTOGENERATED_TRAINER_COMMENT += f"\nSee axolotl config \n\naxolotl version: `{version}`\n```yaml\n{raw_axolotl_cfg.read_text(encoding='utf-8')}\n```\n\n \n"
+
+ LOG.info("Starting trainer...")
+ if cfg.group_by_length:
+ LOG.info("hang tight... sorting dataset for group_by_length")
+
+ pretrain_hooks(cfg, trainer)
+ if cfg.flash_optimum:
+ with torch.backends.cuda.sdp_kernel(
+ # TODO configure these from the YAML w/ sdp_kernel_kwargs: ...
+ enable_flash=True,
+ enable_math=True,
+ enable_mem_efficient=True,
+ ):
+ trainer.train(resume_from_checkpoint=resume_from_checkpoint)
+ else:
+ trainer.train(resume_from_checkpoint=resume_from_checkpoint)
+ post_train_hooks(cfg, trainer)
+
+ LOG.info(f"Training Completed!!! Saving pre-trained model to {cfg.output_dir}")
+
+ # post training
+ for name, module in model.named_modules():
+ if hasattr(module, "_post_training"):
+ module._post_training(model, name) # pylint: disable=protected-access
+
+ if trainer.is_fsdp_enabled:
+ trainer.accelerator.state.fsdp_plugin.set_state_dict_type("FULL_STATE_DICT")
+ LOG.info("Set FSDP state dict type to FULL_STATE_DICT for saving.")
+
+ if cfg.relora_steps:
+ if cfg.adapter == "lora" and not (cfg.load_in_4bit or cfg.load_in_8bit):
+ model = model.merge_and_unload()
+ else:
+ # final model weights have already been saved by `ReLoRACallback.on_train_end`
+ return model, tokenizer
+
+ # TODO do we need this fix? https://huggingface.co/docs/accelerate/usage_guides/fsdp#saving-and-loading
+ # only save on rank 0, otherwise it corrupts output on multi-GPU when multiple processes attempt to write the same file
+ if cfg.fsdp:
+ trainer.save_model(cfg.output_dir)
+ elif cfg.deepspeed and is_deepspeed_zero3_enabled():
+ # Copied over from: https://github.com/huggingface/accelerate/blob/5ae611118057232f441055f7ef9ba0b0f2b8d533/docs/source/usage_guides/deepspeed.md#saving-and-loading
+ trainer.accelerator.wait_for_everyone()
+ unwrapped_model = trainer.accelerator.unwrap_model(trainer.model_wrapped)
+
+ # Saves the whole/unpartitioned fp16 model when in ZeRO Stage-3 to the output directory if
+ # `stage3_gather_16bit_weights_on_model_save` is True in DeepSpeed Config file or
+ # `zero3_save_16bit_model` is True in DeepSpeed Plugin.
+ # For Zero Stages 1 and 2, models are saved as usual in the output directory.
+ # The model name saved is `pytorch_model.bin`
+ unwrapped_model.save_pretrained(
+ cfg.output_dir,
+ is_main_process=trainer.accelerator.is_main_process,
+ save_function=trainer.accelerator.save,
+ state_dict=trainer.accelerator.get_state_dict(trainer.model_wrapped),
+ )
+ elif cfg.local_rank == 0:
+ if cfg.flash_optimum and BetterTransformer:
+ model = BetterTransformer.reverse(model)
+
+ if cfg.rl and cfg.adapter and not cfg.rl_adapter_ref_model:
+ trainer.model.save_pretrained(
+ cfg.output_dir, safe_serialization=safe_serialization
+ )
+ model.save_pretrained(cfg.output_dir, safe_serialization=safe_serialization)
+
+ if not cfg.hub_model_id:
+ try:
+ trainer.create_model_card(model_name=cfg.output_dir.lstrip("./"))
+ except AttributeError:
+ pass
+ elif cfg.hub_model_id:
+ # defensively push to the hub to ensure the model card is updated
+ trainer.push_to_hub()
+
+ return model, tokenizer
+
+
+def pretrain_hooks(_cfg, _trainer):
+ """
+ Run hooks right before kicking off the training
+ :param cfg:
+ :param trainer:
+ :return:
+ """
+
+
+def post_train_hooks(_cfg, _trainer):
+ """
+ Run hooks right after training completes
+ :param cfg:
+ :param trainer:
+ :return:
+ """
diff --git a/src/axolotl/utils/__init__.py b/src/axolotl/utils/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..99dec79f1bf928a386584f13cfdb7cd21ed523b3
--- /dev/null
+++ b/src/axolotl/utils/__init__.py
@@ -0,0 +1,8 @@
+"""
+Basic utils for Axolotl
+"""
+import importlib
+
+
+def is_mlflow_available():
+ return importlib.util.find_spec("mlflow") is not None
diff --git a/src/axolotl/utils/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..7564ffb7a0bd33ec5a8f582ba0ea4b5c1a805114
Binary files /dev/null and b/src/axolotl/utils/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/bench.cpython-310.pyc b/src/axolotl/utils/__pycache__/bench.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..0c186f0af6e46e7dc8f9bafcf9e8ed7153417f8c
Binary files /dev/null and b/src/axolotl/utils/__pycache__/bench.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/chat_templates.cpython-310.pyc b/src/axolotl/utils/__pycache__/chat_templates.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..678c328003a5a6da67e8f0d1a6c734b11785a222
Binary files /dev/null and b/src/axolotl/utils/__pycache__/chat_templates.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/collators.cpython-310.pyc b/src/axolotl/utils/__pycache__/collators.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..4ed1ea86794a0c16576013bae0f856326567d037
Binary files /dev/null and b/src/axolotl/utils/__pycache__/collators.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/dict.cpython-310.pyc b/src/axolotl/utils/__pycache__/dict.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..089c728fb91bd629ebaa01beaf8284a5e81c37d6
Binary files /dev/null and b/src/axolotl/utils/__pycache__/dict.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/distributed.cpython-310.pyc b/src/axolotl/utils/__pycache__/distributed.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..48e070fba7b9d37a63111656090101f37bd9c739
Binary files /dev/null and b/src/axolotl/utils/__pycache__/distributed.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/freeze.cpython-310.pyc b/src/axolotl/utils/__pycache__/freeze.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..7b622d53814db91792b20639df69c456a87639d2
Binary files /dev/null and b/src/axolotl/utils/__pycache__/freeze.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/lora_embeddings.cpython-310.pyc b/src/axolotl/utils/__pycache__/lora_embeddings.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..9dadcc928ad26329b12bf94e85f81632b2b4d4c1
Binary files /dev/null and b/src/axolotl/utils/__pycache__/lora_embeddings.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/mlflow_.cpython-310.pyc b/src/axolotl/utils/__pycache__/mlflow_.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..8f6b7605528a89ff1cf41f0306007f1d3ac771e6
Binary files /dev/null and b/src/axolotl/utils/__pycache__/mlflow_.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/model_shard_quant.cpython-310.pyc b/src/axolotl/utils/__pycache__/model_shard_quant.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..c1eb4c16351e39aeea9655d675c2d0a52616618e
Binary files /dev/null and b/src/axolotl/utils/__pycache__/model_shard_quant.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/models.cpython-310.pyc b/src/axolotl/utils/__pycache__/models.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..627381bde448377fcb93a003a6b5cd94438e2a11
Binary files /dev/null and b/src/axolotl/utils/__pycache__/models.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/schedulers.cpython-310.pyc b/src/axolotl/utils/__pycache__/schedulers.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..ab48eb6866ec180761b4c9ce27b492890a07c6d6
Binary files /dev/null and b/src/axolotl/utils/__pycache__/schedulers.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/tokenization.cpython-310.pyc b/src/axolotl/utils/__pycache__/tokenization.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..4e93cd7fbe72935557a6a3274e44c76d3ce10f6c
Binary files /dev/null and b/src/axolotl/utils/__pycache__/tokenization.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/trainer.cpython-310.pyc b/src/axolotl/utils/__pycache__/trainer.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..0996e656609e399975f9e8c3e4964d0e17c72ce6
Binary files /dev/null and b/src/axolotl/utils/__pycache__/trainer.cpython-310.pyc differ
diff --git a/src/axolotl/utils/__pycache__/wandb_.cpython-310.pyc b/src/axolotl/utils/__pycache__/wandb_.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..363eacb10646633103bed3aac4be226870cff7de
Binary files /dev/null and b/src/axolotl/utils/__pycache__/wandb_.cpython-310.pyc differ
diff --git a/src/axolotl/utils/bench.py b/src/axolotl/utils/bench.py
new file mode 100644
index 0000000000000000000000000000000000000000..11c25160da872eac524c860e649a66fc1215ae4c
--- /dev/null
+++ b/src/axolotl/utils/bench.py
@@ -0,0 +1,84 @@
+"""Benchmarking and measurement utilities"""
+import functools
+
+import pynvml
+import torch
+from pynvml.nvml import NVMLError
+
+
+def check_cuda_device(default_value):
+ """
+ wraps a function and returns the default value instead of running the
+ wrapped function if cuda isn't available or the device is auto
+ :param default_value:
+ :return:
+ """
+
+ def deco(func):
+ @functools.wraps(func)
+ def wrapper(*args, **kwargs):
+ device = kwargs.get("device", args[0] if args else None)
+
+ if (
+ device is None
+ or not torch.cuda.is_available()
+ or device == "auto"
+ or torch.device(device).type == "cpu"
+ or torch.device(device).type == "meta"
+ ):
+ return default_value
+ return func(*args, **kwargs)
+
+ return wrapper
+
+ return deco
+
+
+@check_cuda_device(0.0)
+def gpu_memory_usage(device=0):
+ return torch.cuda.memory_allocated(device) / 1024.0**3
+
+
+@check_cuda_device((0.0, 0.0, 0.0))
+def gpu_memory_usage_all(device=0):
+ usage = torch.cuda.memory_allocated(device) / 1024.0**3
+ reserved = torch.cuda.memory_reserved(device) / 1024.0**3
+ smi = gpu_memory_usage_smi(device)
+ return usage, reserved - usage, max(0, smi - reserved)
+
+
+def mps_memory_usage_all():
+ usage = torch.mps.current_allocated_memory() / 1024.0**3
+ reserved = torch.mps.driver_allocated_memory() / 1024.0**3
+ return usage, reserved - usage, 0
+
+
+@check_cuda_device(0.0)
+def gpu_memory_usage_smi(device=0):
+ if isinstance(device, torch.device):
+ device = device.index
+ if isinstance(device, str) and device.startswith("cuda:"):
+ device = int(device[5:])
+ try:
+ pynvml.nvmlInit()
+ handle = pynvml.nvmlDeviceGetHandleByIndex(device)
+ info = pynvml.nvmlDeviceGetMemoryInfo(handle)
+ return info.used / 1024.0**3
+ except NVMLError:
+ return 0.0
+
+
+def log_gpu_memory_usage(log, msg, device):
+ if torch.backends.mps.is_available():
+ usage, cache, misc = mps_memory_usage_all()
+ else:
+ usage, cache, misc = gpu_memory_usage_all(device)
+ extras = []
+ if cache > 0:
+ extras.append(f"+{cache:.03f}GB cache")
+ if misc > 0:
+ extras.append(f"+{misc:.03f}GB misc")
+ log.info(
+ f"GPU memory usage {msg}: {usage:.03f}GB ({', '.join(extras)})", stacklevel=2
+ )
+ return usage, cache, misc
diff --git a/src/axolotl/utils/callbacks/__init__.py b/src/axolotl/utils/callbacks/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..2965ac1e2934d50d95bf49b9a9e70d35eafe0697
--- /dev/null
+++ b/src/axolotl/utils/callbacks/__init__.py
@@ -0,0 +1,796 @@
+"""Callbacks for Trainer class"""
+
+from __future__ import annotations
+
+import logging
+import os
+from shutil import copyfile
+from tempfile import NamedTemporaryFile
+from typing import TYPE_CHECKING, Any, Dict, List
+
+import evaluate
+import numpy as np
+import pandas as pd
+import torch
+import torch.distributed as dist
+import wandb
+from datasets import load_dataset
+from optimum.bettertransformer import BetterTransformer
+from tqdm import tqdm
+from transformers import (
+ GenerationConfig,
+ Trainer,
+ TrainerCallback,
+ TrainerControl,
+ TrainerState,
+ TrainingArguments,
+)
+from transformers.trainer_utils import PREFIX_CHECKPOINT_DIR, IntervalStrategy
+
+from axolotl.utils import is_mlflow_available
+from axolotl.utils.bench import log_gpu_memory_usage
+from axolotl.utils.config.models.input.v0_4_1 import AxolotlInputConfig
+from axolotl.utils.distributed import (
+ barrier,
+ broadcast_dict,
+ gather_scalar_from_all_ranks,
+ get_world_size,
+ is_distributed,
+ is_main_process,
+ zero_first,
+)
+
+if TYPE_CHECKING:
+ from axolotl.core.trainer_builder import AxolotlTrainingArguments
+
+IGNORE_INDEX = -100
+LOG = logging.getLogger("axolotl.callbacks")
+
+
+class EvalFirstStepCallback(
+ TrainerCallback
+): # pylint: disable=too-few-public-methods disable=unused-argument
+ """
+ Callback to trigger evals on the first step
+ """
+
+ def on_step_end(
+ self,
+ args: TrainingArguments,
+ state: TrainerState,
+ control: TrainerControl,
+ **kwargs,
+ ):
+ if (
+ args.evaluation_strategy == IntervalStrategy.STEPS
+ and state.global_step == 1
+ ):
+ control.should_evaluate = True
+ return control
+
+
+class SaveBetterTransformerModelCallback(
+ TrainerCallback
+): # pylint: disable=too-few-public-methods
+ """Callback to save the BetterTransformer wrapped model"""
+
+ def on_step_end(
+ self,
+ args: TrainingArguments,
+ state: TrainerState,
+ control: TrainerControl,
+ **kwargs,
+ ):
+ # Save
+ if (
+ args.save_strategy == IntervalStrategy.STEPS
+ and args.save_steps > 0
+ and state.global_step % args.save_steps == 0
+ ):
+ control.should_save = True
+
+ if control.should_save:
+ checkpoint_folder = os.path.join(
+ args.output_dir,
+ f"{PREFIX_CHECKPOINT_DIR}-{state.global_step}",
+ )
+
+ model = BetterTransformer.reverse(kwargs["model"])
+ model.save_pretrained(checkpoint_folder)
+ # FIXME - need to cleanup old checkpoints
+
+ # since we're saving here, we don't need the trainer loop to attempt to save too b/c
+ # the trainer will raise an exception since it can't save a BetterTransformer wrapped model
+ control.should_save = False
+ return control
+
+
+class GPUStatsCallback(
+ TrainerCallback
+): # pylint: disable=too-few-public-methods disable=unused-argument
+ """Callback to track GPU utilization"""
+
+ def __init__(self, cfg):
+ self.cfg = cfg
+ self.logged = False
+
+ def on_step_end(
+ self,
+ args: TrainingArguments,
+ state: TrainerState,
+ control: TrainerControl,
+ **kwargs,
+ ):
+ if not self.logged and state.global_step > 1:
+ log_gpu_memory_usage(LOG, "while training", self.cfg.device)
+ self.logged = True
+ return control
+
+
+class LossWatchDogCallback(TrainerCallback):
+ """Callback to track loss and stop training if loss is too high"""
+
+ def __init__(self, cfg):
+ self.cfg = cfg
+ self.logged = False
+ self.violations = 0
+ self.threshold = cfg.loss_watchdog_threshold
+ self.patience = cfg.loss_watchdog_patience or 3
+
+ def on_step_end(
+ self,
+ _args: TrainingArguments,
+ state: TrainerState,
+ control: TrainerControl,
+ **_kwargs,
+ ):
+ if len(state.log_history) > 0 and "loss" in state.log_history[-1]:
+ if state.log_history[-1]["loss"] > self.threshold:
+ self.violations += 1
+ if self.violations >= self.patience:
+ LOG.warning(
+ "Loss is too high, stopping training (loss_watchdog_threshold)"
+ )
+ control.should_training_stop = True
+ else:
+ self.violations = 0
+ return control
+
+
+def bench_eval_callback_factory(trainer, tokenizer):
+ accuracy = evaluate.load("accuracy")
+ abcd_idx = [
+ tokenizer("A", add_special_tokens=False).input_ids[0],
+ tokenizer("B", add_special_tokens=False).input_ids[0],
+ tokenizer("C", add_special_tokens=False).input_ids[0],
+ tokenizer("D", add_special_tokens=False).input_ids[0],
+ tokenizer("E", add_special_tokens=False).input_ids[0],
+ tokenizer("F", add_special_tokens=False).input_ids[0],
+ tokenizer("G", add_special_tokens=False).input_ids[0],
+ ]
+ bench_split = "eval"
+
+ def transform_bench_subject(example):
+ # Split on ':' and trim whitespace
+ parts = example["subject"].split(":")
+ first_part = (
+ parts[0].strip().lower().replace("-", "_")
+ ) # Lowercase the first part
+ second_part = (
+ parts[1].strip().replace("-", "_") if len(parts) > 1 else "all"
+ ) # Replace hyphens with underscores
+
+ # Return the transformed values
+ return {"name": first_part, "subject": second_part}
+
+ if trainer.args.bench_dataset == "mmlu-zs":
+ bench_dataset = load_dataset(
+ "openaccess-ai-collective/mmlu-evals",
+ data_files={
+ "eval": "zero_shot_mmlu_val.json",
+ "test": "zero_shot_mmlu_test.json",
+ },
+ )
+ # bench_dataset = bench_dataset.remove_columns("subject")
+ # MMLU Five-shot (Eval/Test only)
+ elif trainer.args.bench_dataset in ["mmlu", "mmlu-fs"]:
+ bench_dataset = load_dataset(
+ "openaccess-ai-collective/mmlu-evals",
+ data_files={
+ "eval": "five_shot_mmlu_val.json",
+ "test": "five_shot_mmlu_test.json",
+ },
+ )
+ # bench_dataset = bench_dataset.remove_columns('subject')
+ elif "/" in trainer.args.bench_dataset:
+ bench_ds = trainer.args.bench_dataset
+ bench_ds_name = "/".join(bench_ds.split("/", 2)[:2])
+ bench_ds_data_file = "/".join(bench_ds.split("/", 2)[2:])
+ bench_dataset = load_dataset(
+ bench_ds_name,
+ data_files={
+ "eval": bench_ds_data_file,
+ },
+ )
+ bench_dataset["eval"] = bench_dataset["eval"].map(transform_bench_subject)
+ else:
+ raise ValueError(
+ f"unhandled value `{trainer.args.bench_dataset}` for bench_dataset training args"
+ )
+ bench_dataset = bench_dataset[trainer.args.bench_split]
+ if trainer.args.max_bench_samples is not None:
+ bench_dataset = bench_dataset.select(range(trainer.args.max_bench_samples))
+
+ def tokenize_evals(example):
+ source = f"{tokenizer.bos_token}{example['input']}"
+ target = f"{example['output']}{tokenizer.eos_token}"
+
+ tokenized_source = tokenizer(
+ source,
+ max_length=2048,
+ truncation=True,
+ add_special_tokens=False,
+ )
+ tokenized_target = tokenizer(
+ target,
+ max_length=2048,
+ truncation=True,
+ add_special_tokens=False,
+ )
+ input_ids = tokenized_source["input_ids"] + tokenized_target["input_ids"]
+ labels = [IGNORE_INDEX] * len(tokenized_source["input_ids"]) + tokenized_target[
+ "input_ids"
+ ]
+
+ return {
+ "input_ids": input_ids,
+ "labels": labels,
+ "subject": example["subject"],
+ }
+
+ with zero_first(is_main_process()):
+ bench_dataset = bench_dataset.map(tokenize_evals)
+ bench_dataset = bench_dataset.filter(lambda x: x["labels"][-2] in abcd_idx)
+
+ class BenchEvalCallback(TrainerCallback):
+ """
+ TrainerCallback that runs the MMLU evals
+ """
+
+ def on_evaluate(
+ self,
+ args: AxolotlTrainingArguments,
+ state: TrainerState, # pylint: disable=unused-argument
+ control: TrainerControl, # pylint: disable=unused-argument
+ metrics: Dict[str, float], # pylint: disable=unused-argument
+ **kwargs, # pylint: disable=unused-argument
+ ):
+ data_loader = trainer.get_bench_dataloader(
+ bench_dataset.remove_columns(["input", "subject", "output", "name"])
+ )
+ trainer.model.eval()
+ preds, refs = [], []
+ loss_bench = 0
+ for batch in tqdm(data_loader, total=len(data_loader)):
+ (loss, logits, labels) = trainer.prediction_step(
+ trainer.model,
+ batch,
+ prediction_loss_only=False,
+ )
+ # There are two tokens, the output, and eos token.
+ for i, logit in enumerate(logits):
+ label_non_zero_id = (batch["labels"][i] != IGNORE_INDEX).nonzero()[
+ 0
+ ][0]
+ logit_abcd = logit[label_non_zero_id - 1][abcd_idx]
+ preds.append(torch.argmax(logit_abcd).item())
+ labels = labels[labels != IGNORE_INDEX].view(-1, 2)[:, 0]
+ refs += [
+ abcd_idx.index(label) if label in abcd_idx else -1
+ for label in labels.tolist()
+ ]
+ loss_bench += loss.item()
+ # Extract results by subject.
+ bench_name = bench_dataset["name"]
+ bench_names: dict = {s: {"refs": [], "preds": []} for s in set(bench_name)}
+ for s, p, r in zip(bench_name, preds, refs): # pylint: disable=invalid-name
+ bench_names[s]["preds"].append(p)
+ bench_names[s]["refs"].append(r)
+ barrier()
+ local_bench_names = bench_names
+ gathered_bench_names: List[Dict] = [{} for _ in range(get_world_size())]
+ # Gather results from all GPUs to GPU 0
+
+ loss_bench_ranks = gather_scalar_from_all_ranks(
+ lambda: loss_bench, get_world_size()
+ )
+ len_data_loader_ranks = gather_scalar_from_all_ranks(
+ lambda: len(data_loader), get_world_size()
+ )
+
+ results = {}
+ if is_distributed() and not is_main_process():
+ dist.gather_object(local_bench_names, dst=0)
+ else:
+ if is_distributed():
+ dist.gather_object(local_bench_names, gathered_bench_names, dst=0)
+ else:
+ gathered_bench_names = [local_bench_names]
+ bench_loss = sum(loss_bench_ranks) / sum(len_data_loader_ranks)
+ results = {f"{bench_split}_bench_loss": bench_loss}
+
+ # Combine results from all GPUs
+ combined_bench_names: Dict[str, Dict[str, List]] = {}
+ for bench_name in gathered_bench_names:
+ for name, data in bench_name.items():
+ if name not in combined_bench_names:
+ combined_bench_names[name] = {"refs": [], "preds": []}
+ combined_bench_names[name]["refs"].extend(data["refs"])
+ combined_bench_names[name]["preds"].extend(data["preds"])
+
+ bench_scores = []
+ bench_refs = []
+ bench_preds = []
+ for (
+ bench_name
+ ) in combined_bench_names: # pylint: disable=consider-using-dict-items
+ bench_score = accuracy.compute(
+ references=combined_bench_names[bench_name]["refs"],
+ predictions=combined_bench_names[bench_name]["preds"],
+ )["accuracy"]
+ bench_refs.extend(combined_bench_names[bench_name]["refs"])
+ bench_preds.extend(combined_bench_names[bench_name]["preds"])
+ if not pd.isna(bench_score):
+ results[
+ f"{bench_split}_bench_accuracy_{bench_name}"
+ ] = bench_score
+ bench_scores.append(bench_score)
+ else:
+ results[f"{bench_split}_bench_accuracy_{bench_name}"] = 0.0
+ bench_scores.append(0.0)
+ results[f"{bench_split}_bench_average_accuracy"] = np.mean(bench_scores)
+ results[f"{bench_split}_bench_total_accuracy"] = accuracy.compute(
+ references=bench_refs, predictions=bench_preds
+ )["accuracy"]
+ trainer.log(results)
+
+ results = broadcast_dict(results)
+ for key, val in results.items():
+ metrics[key] = val
+
+ return BenchEvalCallback
+
+
+def causal_lm_bench_eval_callback_factory(trainer: Trainer, tokenizer):
+ class CausalLMBenchEvalCallback(TrainerCallback):
+ """Callback to log prediction values during each evaluation"""
+
+ def __init__(self, cfg):
+ self.cfg = cfg
+ self.logged = False
+ self.metrics = self.__maybe_load_metrics()
+
+ def __maybe_load_metrics(self):
+ metrics = {}
+ for metric in self.cfg.eval_causal_lm_metrics:
+ try:
+ metrics[metric] = evaluate.load(metric)
+ except Exception as exc: # pylint: disable=broad-exception-caught
+ LOG.warning(f"{metric}: {exc.args}")
+ return metrics
+
+ def on_evaluate(
+ self,
+ args: AxolotlTrainingArguments, # pylint: disable=unused-argument
+ state: TrainerState,
+ control: TrainerControl,
+ train_dataloader, # pylint: disable=unused-argument
+ eval_dataloader,
+ **kwargs, # pylint: disable=unused-argument
+ ):
+ trainer.model.eval()
+ device = torch.device(self.cfg.device)
+
+ # pylint: disable=duplicate-code
+ generation_config = GenerationConfig(
+ max_new_tokens=self.cfg.eval_max_new_tokens,
+ bos_token_id=tokenizer.bos_token_id,
+ eos_token_id=tokenizer.eos_token_id,
+ pad_token_id=tokenizer.pad_token_id,
+ do_sample=False,
+ use_cache=True,
+ return_dict_in_generate=True,
+ output_attentions=False,
+ output_hidden_states=False,
+ output_scores=False,
+ )
+
+ def find_ranges(lst):
+ ranges = []
+ start = 0
+ for i in range(1, len(lst)):
+ if lst[i] == 0:
+ ranges.append((start, i - 1))
+ start = i
+ end = len(lst) - 1
+ ranges.append((start, end))
+ return ranges
+
+ def compute(metric: evaluate.Metric, **kwargs):
+ # safely compute a metric and return the score if the format is correct
+ metric_score = None
+ try:
+ metric_score = metric.compute(**kwargs)
+ return (
+ metric_score["score"]
+ if "score" in metric_score
+ else metric_score["mean_score"]
+ )
+ except Exception: # pylint: disable=broad-exception-caught
+ LOG.debug(
+ f"Failed to compute metric {metric.name} with kwargs {kwargs.keys()}"
+ )
+ return metric_score
+
+ def evaluate_preds(sources, predictions, references):
+ scores = {}
+
+ for metric_name, metric in self.metrics.items():
+ score = compute(
+ metric,
+ references=references,
+ predictions=predictions,
+ sources=sources,
+ )
+ score = score or compute(
+ metric,
+ references=[[r] for r in references],
+ predictions=predictions,
+ )
+ scores[metric_name] = score
+ return scores
+
+ def predict_with_generate():
+ eval_src, eval_pred, eval_ref = [], [], []
+
+ for batch in tqdm(eval_dataloader):
+ batch_labels = batch["labels"].to(device)
+ batch_input_ids = batch["input_ids"].to(device)
+
+ if "position_ids" in batch:
+ batch_pos_ids = batch["position_ids"].tolist()
+ else:
+ batch_pos_ids = [None] * len(batch["input_ids"])
+
+ prompt_token_ids_list = []
+ completion_token_ids_list = []
+
+ for input_ids_all, labels_all, pos_ids in zip(
+ batch_input_ids,
+ batch_labels,
+ batch_pos_ids,
+ ):
+ if pos_ids is None:
+ pos_ranges = [(0, len(input_ids_all) - 1)]
+ else:
+ pos_ranges = find_ranges(pos_ids)
+
+ for pos_range in pos_ranges:
+ start, end = pos_range
+ if start == end:
+ continue
+
+ input_ids = input_ids_all[start : end + 1]
+ labels = labels_all[start : end + 1]
+
+ tokens_without_loss = labels == IGNORE_INDEX
+ tokens_with_loss = labels != IGNORE_INDEX
+ tokens_exclude_padding = input_ids != tokenizer.pad_token_id
+ prompt_token_includes = (
+ tokens_without_loss & tokens_exclude_padding
+ )
+
+ prompt_token_ids = input_ids[prompt_token_includes]
+ prompt_token_ids_list.append(prompt_token_ids)
+
+ completion_token_ids = input_ids[tokens_with_loss]
+ completion_token_ids_list.append(completion_token_ids)
+
+ prompt_texts = tokenizer.batch_decode(
+ prompt_token_ids_list, skip_special_tokens=True
+ )
+ completion_texts = tokenizer.batch_decode(
+ completion_token_ids_list, skip_special_tokens=True
+ )
+
+ with torch.no_grad():
+ prompt_encoding = tokenizer(
+ prompt_texts, padding=True, return_tensors="pt"
+ ).to(self.cfg.device)
+ predictions = trainer.model.generate(
+ **prompt_encoding, generation_config=generation_config
+ )
+
+ prediction_all_tokens = predictions["sequences"].cpu().tolist()
+ prediction_without_prompt_tokens_list = []
+ for prompt_token_ids, prediction_tokens in zip(
+ prompt_token_ids_list, prediction_all_tokens
+ ):
+ prediction_without_prompt_tokens = prediction_tokens[
+ len(prompt_token_ids) :
+ ]
+ prediction_without_prompt_tokens_list.append(
+ prediction_without_prompt_tokens
+ )
+
+ predicted_texts = tokenizer.batch_decode(
+ prediction_without_prompt_tokens_list, skip_special_tokens=True
+ )
+
+ eval_src.extend(prompt_texts)
+ eval_pred.extend(predicted_texts)
+ eval_ref.extend(completion_texts)
+
+ return eval_src, eval_pred, eval_ref
+
+ if is_main_process():
+ eval_preds = predict_with_generate()
+ trainer.log(evaluate_preds(*eval_preds))
+
+ return control
+
+ return CausalLMBenchEvalCallback
+
+
+def log_prediction_callback_factory(trainer: Trainer, tokenizer, logger: str):
+ class LogPredictionCallback(TrainerCallback):
+ """Callback to log prediction values during each evaluation"""
+
+ def __init__(self, cfg):
+ self.cfg = cfg
+ self.logged = False
+
+ def on_evaluate(
+ self,
+ args: AxolotlTrainingArguments, # pylint: disable=unused-argument
+ state: TrainerState,
+ control: TrainerControl,
+ train_dataloader, # pylint: disable=unused-argument
+ eval_dataloader,
+ **kwargs, # pylint: disable=unused-argument
+ ):
+ eval_table_size = self.cfg.eval_table_size
+
+ if eval_table_size <= 0:
+ return control
+
+ trainer.model.eval()
+ device = torch.device(self.cfg.device)
+
+ # pylint: disable=duplicate-code
+ generation_config = GenerationConfig(
+ max_new_tokens=self.cfg.eval_max_new_tokens,
+ bos_token_id=tokenizer.bos_token_id,
+ eos_token_id=tokenizer.eos_token_id,
+ pad_token_id=tokenizer.pad_token_id,
+ do_sample=False,
+ use_cache=True,
+ return_dict_in_generate=True,
+ output_attentions=False,
+ output_hidden_states=False,
+ output_scores=False,
+ )
+
+ def logits_to_tokens(logits) -> torch.Tensor:
+ probabilities = torch.softmax(logits, dim=-1)
+ # Get the predicted token ids (the ones with the highest probability)
+ predicted_token_ids = torch.argmax(probabilities, dim=-1)
+ return predicted_token_ids
+
+ def find_ranges(lst):
+ ranges = []
+ start = 0
+ for i in range(1, len(lst)):
+ if lst[i] == 0:
+ ranges.append((start, i - 1))
+ start = i
+ end = len(lst) - 1
+ ranges.append((start, end))
+ return ranges
+
+ def log_table_from_dataloader(name: str, table_dataloader):
+ table_data: Dict[str, List[Any]] = {
+ "id": [],
+ "Prompt": [],
+ "Correct Completion": [],
+ "Predicted Completion (model.generate)": [],
+ "Predicted Completion (trainer.prediction_step)": [],
+ }
+ row_index = 0
+
+ for batch in tqdm(table_dataloader):
+ if row_index > eval_table_size:
+ break
+
+ batch_labels = batch["labels"].to(device)
+ batch_input_ids = batch["input_ids"].to(device)
+
+ if "position_ids" in batch:
+ batch_pos_ids = batch["position_ids"].tolist()
+ else:
+ batch_pos_ids = [None] * len(batch["input_ids"])
+
+ (_, batch_logits, _) = trainer.prediction_step(
+ trainer.model,
+ batch,
+ prediction_loss_only=False,
+ )
+
+ prompt_token_ids_list = []
+ pred_step_token_ids_list = []
+ completion_token_ids_list = []
+
+ for input_ids_all, labels_all, pos_ids, logits in zip(
+ batch_input_ids,
+ batch_labels,
+ batch_pos_ids,
+ batch_logits,
+ ):
+ if pos_ids is None:
+ pos_ranges = [(0, len(input_ids_all) - 1)]
+ else:
+ pos_ranges = find_ranges(pos_ids)
+
+ for pos_range in pos_ranges:
+ start, end = pos_range
+ if start == end:
+ continue
+
+ input_ids = input_ids_all[start : end + 1]
+ labels = labels_all[start : end + 1]
+
+ tokens_without_loss = labels == IGNORE_INDEX
+ tokens_with_loss = labels != IGNORE_INDEX
+ tokens_exclude_padding = input_ids != tokenizer.pad_token_id
+ prompt_token_includes = (
+ tokens_without_loss & tokens_exclude_padding
+ )
+
+ prompt_token_ids = input_ids[prompt_token_includes]
+ prompt_token_ids_list.append(prompt_token_ids)
+
+ completion_token_ids = input_ids[tokens_with_loss]
+ completion_token_ids_list.append(completion_token_ids)
+
+ pred_step_token_ids = logits_to_tokens(
+ logits[start : end + 1]
+ )[tokens_with_loss]
+ pred_step_token_ids_list.append(pred_step_token_ids)
+
+ prompt_texts = tokenizer.batch_decode(
+ prompt_token_ids_list, skip_special_tokens=True
+ )
+ completion_texts = tokenizer.batch_decode(
+ completion_token_ids_list, skip_special_tokens=True
+ )
+ pred_step_texts = tokenizer.batch_decode(
+ pred_step_token_ids_list, skip_special_tokens=True
+ )
+
+ with torch.no_grad():
+ prompt_encoding = tokenizer(
+ prompt_texts, padding=True, return_tensors="pt"
+ ).to(self.cfg.device)
+ predictions = trainer.model.generate(
+ **prompt_encoding, generation_config=generation_config
+ )
+
+ prediction_all_tokens = predictions["sequences"].cpu().tolist()
+ prediction_without_prompt_tokens_list = []
+ for prompt_token_ids, prediction_tokens in zip(
+ prompt_token_ids_list, prediction_all_tokens
+ ):
+ prediction_without_prompt_tokens = prediction_tokens[
+ len(prompt_token_ids) :
+ ]
+ prediction_without_prompt_tokens_list.append(
+ prediction_without_prompt_tokens
+ )
+
+ predicted_texts = tokenizer.batch_decode(
+ prediction_without_prompt_tokens_list, skip_special_tokens=True
+ )
+
+ for (
+ prompt_text,
+ completion_text,
+ prediction_text,
+ pred_step_text,
+ ) in zip(
+ prompt_texts, completion_texts, predicted_texts, pred_step_texts
+ ):
+ table_data["id"].append(row_index)
+ table_data["Prompt"].append(prompt_text)
+ table_data["Correct Completion"].append(completion_text)
+ table_data["Predicted Completion (model.generate)"].append(
+ prediction_text
+ )
+ table_data[
+ "Predicted Completion (trainer.prediction_step)"
+ ].append(pred_step_text)
+ row_index += 1
+ if logger == "wandb":
+ wandb.run.log({f"{name} - Predictions vs Ground Truth": pd.DataFrame(table_data)}) # type: ignore[attr-defined]
+ elif logger == "mlflow" and is_mlflow_available():
+ import mlflow
+
+ tracking_uri = AxolotlInputConfig(
+ **self.cfg.to_dict()
+ ).mlflow_tracking_uri
+ mlflow.log_table(
+ data=table_data,
+ artifact_file="PredictionsVsGroundTruth.json",
+ tracking_uri=tracking_uri,
+ )
+
+ if is_main_process():
+ log_table_from_dataloader("Eval", eval_dataloader)
+
+ return control
+
+ return LogPredictionCallback
+
+
+class SaveAxolotlConfigtoWandBCallback(TrainerCallback):
+ """Callback to save axolotl config to wandb"""
+
+ def __init__(self, axolotl_config_path):
+ self.axolotl_config_path = axolotl_config_path
+
+ def on_train_begin(
+ self,
+ args: AxolotlTrainingArguments, # pylint: disable=unused-argument
+ state: TrainerState, # pylint: disable=unused-argument
+ control: TrainerControl,
+ **kwargs, # pylint: disable=unused-argument
+ ):
+ if is_main_process():
+ try:
+ # sync config to top level in run, cannot delete file right away because wandb schedules it to be synced even w/policy = 'now', so let OS delete it later.
+ with NamedTemporaryFile(
+ mode="w", delete=False, suffix=".yml", prefix="axolotl_config_"
+ ) as temp_file:
+ copyfile(self.axolotl_config_path, temp_file.name)
+ artifact = wandb.Artifact(
+ f"config-{wandb.run.id}", type="axolotl-config"
+ )
+ artifact.add_file(temp_file.name)
+ wandb.log_artifact(artifact)
+ wandb.save(temp_file.name)
+ LOG.info(
+ "The Axolotl config has been saved to the WandB run under files."
+ )
+ except (FileNotFoundError, ConnectionError) as err:
+ LOG.warning(f"Error while saving Axolotl config to WandB: {err}")
+ return control
+
+
+class SaveModelOnTrainEndCallback(TrainerCallback):
+ """Callback to save model on train end"""
+
+ def on_step_end( # pylint: disable=unused-argument
+ self,
+ args: TrainingArguments,
+ state: TrainerState,
+ control: TrainerControl,
+ **kwargs,
+ ):
+ # Save
+ if state.global_step >= state.max_steps:
+ control.should_save = True
+
+ def on_train_end( # pylint: disable=unused-argument
+ self, args, state, control, **kwargs
+ ):
+ control.should_save = True
+ return control
diff --git a/src/axolotl/utils/callbacks/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/callbacks/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..adafdbf38172f0e53dce6674128445784f8c985b
Binary files /dev/null and b/src/axolotl/utils/callbacks/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/callbacks/__pycache__/lisa.cpython-310.pyc b/src/axolotl/utils/callbacks/__pycache__/lisa.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..1dc8486cfdf0e8d7cebd615037a0809615646254
Binary files /dev/null and b/src/axolotl/utils/callbacks/__pycache__/lisa.cpython-310.pyc differ
diff --git a/src/axolotl/utils/callbacks/lisa.py b/src/axolotl/utils/callbacks/lisa.py
new file mode 100644
index 0000000000000000000000000000000000000000..ff20959a59365f41e0f16a407a0bd60926204166
--- /dev/null
+++ b/src/axolotl/utils/callbacks/lisa.py
@@ -0,0 +1,91 @@
+"""
+module for LISA
+
+Adapted from https://github.com/OptimalScale/LMFlow/pull/701 for HF transformers & Axolotl
+Arxiv: https://arxiv.org/abs/2403.17919
+License: Apache 2.0
+"""
+
+import logging
+from functools import reduce
+from typing import TYPE_CHECKING
+
+import numpy as np
+from transformers import TrainerCallback
+
+if TYPE_CHECKING:
+ from axolotl.core.trainer_builder import AxolotlTrainer
+
+LOG = logging.getLogger("axolotl.callbacks.lisa")
+
+
+def lisa_callback_factory(trainer: "AxolotlTrainer"):
+ class LISACallback(TrainerCallback):
+ """trainer callback for lisa layer switching"""
+
+ def __init__(
+ self, n_layers, step_interval, trainer, layers_attribute="model.layers"
+ ):
+ super().__init__()
+ self.n_layers = n_layers
+ self.step_interval = step_interval
+ self.layers_attribute = layers_attribute
+ self.trainer = trainer
+
+ reduce(getattr, self.layers_attribute.split("."), self.trainer.model)
+
+ self.total_layers = len(
+ reduce(getattr, self.layers_attribute.split("."), self.trainer.model)
+ )
+ self.active_layers_indices = []
+
+ layers = reduce(
+ getattr, self.layers_attribute.split("."), self.trainer.model
+ )
+ LOG.info(
+ f"LISA will activate {self.n_layers}/{len(layers)} layers ({self.n_layers*100/len(layers)}%) every {self.step_interval} steps"
+ )
+
+ def freeze_all_layers(self):
+ layers = reduce(
+ getattr, self.layers_attribute.split("."), self.trainer.model
+ )
+ for layer in layers:
+ for param in layer.parameters():
+ param.requires_grad = False
+
+ def on_step_begin(
+ self, args, state, control, **kwargs
+ ): # pylint: disable=unused-argument
+ # Check if it's time to switch active layers, including at step 0
+ if state.global_step % self.step_interval == 0 or state.global_step == 1:
+ self.switch_active_layers()
+
+ def switch_active_layers(self):
+ # First, disable gradients for all layers
+ self.freeze_all_layers()
+
+ # Randomly select n_layers to activate
+ layers = reduce(
+ getattr, self.layers_attribute.split("."), self.trainer.model
+ )
+ self.active_layers_indices = np.random.choice(
+ range(self.total_layers), self.n_layers, replace=False
+ )
+ LOG.info(
+ f"Activating layers at indices: {self.active_layers_indices} for the next steps."
+ )
+
+ # Enable gradients only for the selected layers
+ for idx in self.active_layers_indices:
+ for param in layers[idx].parameters():
+ param.requires_grad = True
+
+ lisa_callback = LISACallback(
+ n_layers=trainer.args.lisa_n_layers,
+ step_interval=trainer.args.lisa_step_interval,
+ trainer=trainer,
+ layers_attribute=trainer.args.lisa_layers_attribute,
+ )
+
+ return lisa_callback
diff --git a/src/axolotl/utils/callbacks/mlflow_.py b/src/axolotl/utils/callbacks/mlflow_.py
new file mode 100644
index 0000000000000000000000000000000000000000..fcbb88edcd4eae61fdeae19678a1edba9f0faa41
--- /dev/null
+++ b/src/axolotl/utils/callbacks/mlflow_.py
@@ -0,0 +1,44 @@
+"""MLFlow module for trainer callbacks"""
+import logging
+from shutil import copyfile
+from tempfile import NamedTemporaryFile
+from typing import TYPE_CHECKING
+
+import mlflow
+from transformers import TrainerCallback, TrainerControl, TrainerState
+
+from axolotl.utils.distributed import is_main_process
+
+if TYPE_CHECKING:
+ from axolotl.core.trainer_builder import AxolotlTrainingArguments
+
+LOG = logging.getLogger("axolotl.callbacks")
+
+
+class SaveAxolotlConfigtoMlflowCallback(TrainerCallback):
+ # pylint: disable=duplicate-code
+ """Callback to save axolotl config to mlflow"""
+
+ def __init__(self, axolotl_config_path):
+ self.axolotl_config_path = axolotl_config_path
+
+ def on_train_begin(
+ self,
+ args: "AxolotlTrainingArguments", # pylint: disable=unused-argument
+ state: TrainerState, # pylint: disable=unused-argument
+ control: TrainerControl,
+ **kwargs, # pylint: disable=unused-argument
+ ):
+ if is_main_process():
+ try:
+ with NamedTemporaryFile(
+ mode="w", delete=False, suffix=".yml", prefix="axolotl_config_"
+ ) as temp_file:
+ copyfile(self.axolotl_config_path, temp_file.name)
+ mlflow.log_artifact(temp_file.name, artifact_path="")
+ LOG.info(
+ "The Axolotl config has been saved to the MLflow artifacts."
+ )
+ except (FileNotFoundError, ConnectionError) as err:
+ LOG.warning(f"Error while saving Axolotl config to MLflow: {err}")
+ return control
diff --git a/src/axolotl/utils/chat_templates.py b/src/axolotl/utils/chat_templates.py
new file mode 100644
index 0000000000000000000000000000000000000000..01b147356858a6d800dbaf47ae99876bb742b361
--- /dev/null
+++ b/src/axolotl/utils/chat_templates.py
@@ -0,0 +1,33 @@
+"""
+This module provides functionality for selecting chat templates based on user choices.
+These templates are used for formatting messages in a conversation.
+"""
+
+
+def chat_templates(user_choice: str):
+ """
+ Finds the correct chat_template for the tokenizer_config.
+
+ Args:
+ user_choice (str): The user's choice of template.
+
+ Returns:
+ str: The chosen template string.
+
+ Raises:
+ ValueError: If the user_choice is not found in the templates.
+ """
+
+ templates = {
+ "alpaca": "{% for message in messages %}{% if message['role'] == 'user' %}{{ '### Instruction: ' + message['content'] + '\n\n' }}{% elif message['role'] == 'assistant' %}{{ '### Response: ' + message['content'] + eos_token}}{% endif %}{% endfor %}",
+ "inst": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", # I don't know what this one is called. Used by Mistral/Mixtral.
+ "chatml": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
+ "gemma": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '' + role + '\n' + message['content'] | trim + '\n' }}{% endfor %}{% if add_generation_prompt %}{{'model\n'}}{% endif %}",
+ "cohere": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% elif false == true %}{% set loop_messages = messages %}{% set system_message = 'You are Command-R, a brilliant, sophisticated, AI-assistant trained to assist human users by providing thorough responses. You are trained by Cohere.' %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% if system_message != false %}{{ '<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>' + system_message + '<|END_OF_TURN_TOKEN|>' }}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|START_OF_TURN_TOKEN|><|USER_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% elif message['role'] == 'assistant' %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' + content.strip() + '<|END_OF_TURN_TOKEN|>' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>' }}{% endif %}",
+ "llama3": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% else %}{{ eos_token }}{% endif %}",
+ }
+
+ if user_choice in templates:
+ return templates[user_choice]
+
+ raise ValueError(f"Template '{user_choice}' not found.")
diff --git a/src/axolotl/utils/collators.py b/src/axolotl/utils/collators.py
new file mode 100644
index 0000000000000000000000000000000000000000..26c7fa9f3c95d84eee5690264a7214297362e516
--- /dev/null
+++ b/src/axolotl/utils/collators.py
@@ -0,0 +1,242 @@
+"""
+DataCollator for axolotl to pad labels and position_ids for packed sequences
+"""
+from dataclasses import dataclass
+from typing import Any, Dict, Optional, Sequence, Union
+
+import numpy as np
+import torch
+import transformers
+from transformers import PreTrainedTokenizerBase
+from transformers.utils import PaddingStrategy
+
+IGNORE_INDEX = -100
+
+
+@dataclass
+class DataCollatorForSeq2Seq:
+ """
+ Data collator that will dynamically pad the inputs received, as well as the labels and position_ids
+
+ Args:
+ tokenizer ([`PreTrainedTokenizer`] or [`PreTrainedTokenizerFast`]):
+ The tokenizer used for encoding the data.
+ model ([`PreTrainedModel`]):
+ The model that is being trained. If set and has the *prepare_decoder_input_ids_from_labels*, use it to
+ prepare the *decoder_input_ids*
+
+ This is useful when using *label_smoothing* to avoid calculating loss twice.
+ padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `True`):
+ Select a strategy to pad the returned sequences (according to the model's padding side and padding index)
+ among:
+
+ - `True` or `'longest'` (default): Pad to the longest sequence in the batch (or no padding if only a single
+ sequence is provided).
+ - `'max_length'`: Pad to a maximum length specified with the argument `max_length` or to the maximum
+ acceptable input length for the model if that argument is not provided.
+ - `False` or `'do_not_pad'`: No padding (i.e., can output a batch with sequences of different lengths).
+ max_length (`int`, *optional*):
+ Maximum length of the returned list and optionally padding length (see above).
+ pad_to_multiple_of (`int`, *optional*):
+ If set will pad the sequence to a multiple of the provided value.
+
+ This is especially useful to enable the use of Tensor Cores on NVIDIA hardware with compute capability >=
+ 7.5 (Volta).
+ label_pad_token_id (`int`, *optional*, defaults to -100):
+ The id to use when padding the labels (-100 will be automatically ignored by PyTorch loss functions).
+ return_tensors (`str`):
+ The type of Tensor to return. Allowable values are "np", "pt" and "tf".
+ """
+
+ tokenizer: PreTrainedTokenizerBase
+ model: Optional[Any] = None
+ padding: Union[bool, str, PaddingStrategy] = True
+ max_length: Optional[int] = None
+ pad_to_multiple_of: Optional[int] = None
+ label_pad_token_id: int = -100
+ position_pad_token_id: int = 0
+ return_tensors: str = "pt"
+
+ def __call__(self, features, return_tensors=None):
+ labels = None
+ if return_tensors is None:
+ return_tensors = self.return_tensors
+
+ for feature_name, pad_token_id in [
+ ("labels", self.label_pad_token_id),
+ ("position_ids", self.position_pad_token_id),
+ ]:
+ feat = (
+ [feature[feature_name] for feature in features]
+ if feature_name in features[0].keys()
+ else None
+ )
+ labels = feat if feat and feature_name == "labels" else labels
+ # We have to pad the labels before calling `tokenizer.pad` as this method won't pad them and needs them of the
+ # same length to return tensors.
+ if feat is not None:
+ max_feature_length = max(len(l) for l in feat) # noqa: E741
+ if self.pad_to_multiple_of is not None:
+ max_feature_length = (
+ (max_feature_length + self.pad_to_multiple_of - 1)
+ // self.pad_to_multiple_of
+ * self.pad_to_multiple_of
+ )
+
+ padding_side = self.tokenizer.padding_side
+ for feature in features:
+ remainder = [pad_token_id] * (
+ max_feature_length - len(feature[feature_name])
+ )
+ if isinstance(feature[feature_name], list):
+ feature[feature_name] = (
+ feature[feature_name] + remainder
+ if padding_side == "right"
+ else remainder + feature[feature_name]
+ )
+ elif padding_side == "right":
+ feature[feature_name] = np.concatenate(
+ [feature[feature_name], remainder]
+ ).astype(np.int64)
+ else:
+ feature[feature_name] = np.concatenate(
+ [remainder, feature[feature_name]]
+ ).astype(np.int64)
+
+ features = self.tokenizer.pad(
+ features,
+ padding=self.padding,
+ max_length=self.max_length,
+ pad_to_multiple_of=self.pad_to_multiple_of,
+ return_tensors=return_tensors,
+ )
+
+ # prepare decoder_input_ids
+ if (
+ labels is not None
+ and self.model is not None
+ and hasattr(self.model, "prepare_decoder_input_ids_from_labels")
+ ):
+ decoder_input_ids = self.model.prepare_decoder_input_ids_from_labels(
+ labels=features["labels"]
+ )
+ features["decoder_input_ids"] = decoder_input_ids
+
+ return features
+
+
+@dataclass
+class BatchSamplerDataCollatorForSeq2Seq(DataCollatorForSeq2Seq):
+ """
+ Collator for multipack specific to the using the BatchSampler
+ """
+
+ def __call__(self, features, return_tensors=None):
+ if not isinstance(features[0], list):
+ features = [features]
+ out_features = [{} for _ in features]
+ for i, features_ in enumerate(features):
+ for feature in features_[0].keys():
+ if feature == "length":
+ continue
+ if feature == "attention_mask":
+ arrays = [
+ (1) * np.array(item[feature])
+ for i, item in enumerate(features_)
+ if feature in item
+ ]
+ out_features[i][feature] = np.concatenate(arrays)
+ else:
+ arrays = [
+ np.array(item[feature]) for item in features_ if feature in item
+ ]
+ out_features[i][feature] = np.concatenate(arrays)
+ return super().__call__(out_features, return_tensors=return_tensors)
+
+
+@dataclass
+class V2BatchSamplerDataCollatorForSeq2Seq(DataCollatorForSeq2Seq):
+ """
+ Collator for multipack specific to the using the BatchSampler
+ """
+
+ def __call__(self, features, return_tensors=None):
+ if not isinstance(features[0], list):
+ features = [features]
+ out_features = [{} for _ in features]
+ for i, features_ in enumerate(features):
+ for feature in features_[0].keys():
+ if feature == "length":
+ continue
+ if feature == "attention_mask":
+ arrays = [
+ (i + 1) * np.array(item[feature])
+ for i, item in enumerate(features_)
+ if feature in item
+ ]
+ out_features[i][feature] = np.concatenate(arrays)
+ else:
+ arrays = [
+ np.array(item[feature]) for item in features_ if feature in item
+ ]
+ out_features[i][feature] = np.concatenate(arrays)
+ return super().__call__(out_features, return_tensors=return_tensors)
+
+
+@dataclass
+class MambaDataCollator:
+ """
+ Collator for State Space Models (Mamba)
+ """
+
+ tokenizer: transformers.PreTrainedTokenizer
+
+ def __call__(self, instances: Sequence[Dict]) -> Dict[str, torch.Tensor]:
+ input_ids, labels = tuple(
+ [torch.LongTensor(instance[key]) for instance in instances]
+ for key in ("input_ids", "labels")
+ )
+ input_ids = torch.nn.utils.rnn.pad_sequence(
+ input_ids,
+ batch_first=True,
+ padding_value=self.tokenizer.pad_token_id,
+ )
+ labels = torch.nn.utils.rnn.pad_sequence(
+ labels, batch_first=True, padding_value=IGNORE_INDEX
+ )
+
+ return {
+ "input_ids": input_ids,
+ "labels": labels,
+ }
+
+
+@dataclass
+class PretrainingBatchSamplerDataCollatorForSeq2Seq(DataCollatorForSeq2Seq):
+ """
+ Collator for multipack specific to the using the BatchSampler
+ """
+
+ def __init__(self, *args, multipack_attn=True, **kwargs):
+ super().__init__(*args, **kwargs)
+ self.multipack_attn = multipack_attn
+
+ def __call__(self, features, return_tensors=None):
+ chunked_data = {}
+ for feature in features.keys():
+ if feature == "length":
+ continue
+ if feature == "attention_mask":
+ if self.multipack_attn:
+ arrays = [
+ (i + 1) * np.array(item)
+ for i, item in enumerate(features[feature])
+ ]
+ else:
+ arrays = [(1) * np.array(item) for item in features[feature]]
+ chunked_data[feature] = np.concatenate(arrays)
+ else:
+ arrays = [np.array(item) for item in features[feature]]
+ chunked_data[feature] = np.concatenate(arrays)
+ features = [chunked_data]
+ return super().__call__(features, return_tensors=return_tensors)
diff --git a/src/axolotl/utils/config/__init__.py b/src/axolotl/utils/config/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..a054f24a7fa8bbc65ec14bb183e8e90dc1e29e6f
--- /dev/null
+++ b/src/axolotl/utils/config/__init__.py
@@ -0,0 +1,605 @@
+"""Module for working with config dicts"""
+import json
+import logging
+import os
+from pathlib import Path
+from typing import Optional
+
+import torch
+from transformers.utils import is_torch_bf16_gpu_available
+
+from axolotl.utils.bench import log_gpu_memory_usage
+from axolotl.utils.config.models.input.v0_4_1 import (
+ AxolotlConfigWCapabilities,
+ AxolotlInputConfig,
+)
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.models import load_model_config
+
+LOG = logging.getLogger("axolotl")
+
+
+def choose_device(cfg):
+ def get_device():
+ try:
+ if torch.cuda.is_available():
+ return f"cuda:{cfg.local_rank}"
+
+ if torch.backends.mps.is_available():
+ return "mps"
+
+ raise SystemError("No CUDA/mps device found")
+ except Exception: # pylint: disable=broad-exception-caught
+ return "cpu"
+
+ cfg.device = get_device()
+ if cfg.world_size == 1:
+ cfg.device_map = cfg.device_map or "auto"
+ else:
+ if cfg.device.startswith("cuda"):
+ cfg.device_map = {"": torch.cuda.current_device()}
+ else:
+ cfg.device_map = {"": cfg.device}
+
+ # in `accelerate launch`, we need to not pass through any device map and let
+ # accelerate figure out which parts of the model to put on which gpu
+ accelerate_vars = [var for var in os.environ if var.startswith("ACCELERATE_USE_")]
+ if accelerate_vars:
+ cfg.device_map = None
+
+
+def normalize_config(cfg):
+ # setup some derived config / hyperparams
+ cfg.gradient_accumulation_steps = cfg.gradient_accumulation_steps or (
+ cfg.batch_size // cfg.micro_batch_size
+ )
+ cfg.batch_size = (
+ cfg.batch_size or cfg.micro_batch_size * cfg.gradient_accumulation_steps
+ )
+ if cfg.eval_batch_size is None:
+ cfg.eval_batch_size = cfg.micro_batch_size
+ cfg.world_size = int(os.environ.get("WORLD_SIZE", 1))
+ cfg.local_rank = int(os.environ.get("LOCAL_RANK", 0))
+ cfg.eval_table_size = cfg.eval_table_size or 0
+ cfg.eval_max_new_tokens = cfg.eval_max_new_tokens or 128
+ cfg.eval_causal_lm_metrics = cfg.eval_causal_lm_metrics or [
+ "sacrebleu",
+ "comet",
+ "ter",
+ "chrf",
+ ]
+ choose_device(cfg)
+ cfg.ddp = cfg.ddp if cfg.ddp is not None else cfg.world_size != 1
+ if cfg.ddp:
+ cfg.device_map = {"": int(os.environ.get("LOCAL_RANK", 0))}
+ cfg.batch_size = cfg.batch_size * cfg.world_size
+
+ if cfg.bf16 == "auto":
+ if is_torch_bf16_gpu_available():
+ LOG.debug("bf16 support detected, enabling for this configuration.")
+ cfg.bf16 = True
+ else:
+ LOG.debug("bf16 support not detected, disabling for this configuration.")
+ cfg.bf16 = False
+ if cfg.fp16 is None:
+ cfg.fp16 = True
+
+ if cfg.device == "mps":
+ cfg.load_in_8bit = False
+ cfg.tf32 = False
+ if cfg.bf16:
+ cfg.fp16 = True
+ cfg.bf16 = False
+ else:
+ torch.backends.cuda.matmul.allow_tf32 = cfg.tf32 or False
+ if cfg.bf16:
+ cfg.fp16 = False
+
+ if cfg.bf16 or cfg.bfloat16:
+ cfg.torch_dtype = torch.bfloat16
+ elif cfg.load_in_8bit or cfg.fp16 or cfg.float16:
+ cfg.torch_dtype = torch.float16
+ else:
+ cfg.torch_dtype = torch.float32
+
+ if cfg.saves_per_epoch:
+ save_steps = 1.0 / (cfg.saves_per_epoch * cfg.num_epochs)
+ if save_steps < 1.0: # prevent saves on every step
+ cfg.save_steps = save_steps
+ if (cfg.val_set_size or cfg.test_datasets) and cfg.evals_per_epoch:
+ eval_steps = 1.0 / (cfg.evals_per_epoch * cfg.num_epochs)
+ if eval_steps < 1.0: # prevent evals on every step
+ cfg.eval_steps = eval_steps
+
+ cfg.dataset_processes = cfg.dataset_processes or os.cpu_count()
+
+ if not cfg.base_model_config:
+ cfg.base_model_config = cfg.base_model
+
+ model_config = load_model_config(cfg)
+ cfg.model_config_type = model_config.model_type
+
+ cfg.tokenizer_config = (
+ cfg.tokenizer_config or cfg.base_model_config or cfg.base_model
+ )
+
+ # figure out if the model is llama
+ cfg.is_llama_derived_model = (
+ (hasattr(model_config, "model_type") and model_config.model_type == "llama")
+ or cfg.is_llama_derived_model
+ or "llama" in cfg.base_model.lower()
+ or (cfg.type_of_model and "llama" in cfg.type_of_model.lower())
+ )
+
+ # figure out if the model is falcon
+ cfg.is_falcon_derived_model = (
+ (
+ hasattr(model_config, "model_type")
+ and model_config.model_type
+ in [
+ "falcon",
+ "RefinedWebModel",
+ "RefinedWeb",
+ ]
+ )
+ or cfg.is_falcon_derived_model
+ or "falcon" in cfg.base_model.lower()
+ or (cfg.type_of_model and "rwforcausallm" in cfg.type_of_model.lower())
+ )
+
+ cfg.is_mistral_derived_model = (
+ (
+ hasattr(model_config, "model_type")
+ and model_config.model_type
+ in [
+ "mistral",
+ ]
+ )
+ or cfg.is_mistral_derived_model
+ or "mistral" in cfg.base_model.lower().split("/")[-1]
+ or (cfg.type_of_model and "mistral" in cfg.type_of_model.lower())
+ )
+
+ cfg.is_qwen_derived_model = (
+ hasattr(model_config, "model_type")
+ and model_config.model_type
+ in [
+ "qwen",
+ ]
+ ) or cfg.is_qwen_derived_model
+
+ if isinstance(cfg.pretraining_dataset, dict):
+ cfg.pretraining_dataset = [cfg.pretraining_dataset]
+
+ if (
+ cfg.gradient_checkpointing
+ and cfg.unfrozen_parameters is None
+ and cfg.gradient_checkpointing_kwargs is None
+ and cfg.rl is None
+ ):
+ cfg.gradient_checkpointing_kwargs = {"use_reentrant": True}
+
+ log_gpu_memory_usage(LOG, "baseline", cfg.device)
+
+
+def normalize_cfg_datasets(cfg):
+ """
+ helpers for mapping chat_template to various dataset configurations as necessary
+ """
+
+ if cfg.chat_template and cfg.chat_template == "chatml":
+ if cfg.datasets:
+ for idx, ds_cfg in enumerate(cfg.datasets):
+ if ds_cfg.type == "sharegpt" and not ds_cfg.conversation:
+ LOG.info(
+ f"updating dataset {ds_cfg.path} with `conversation: chatml` to match your chat_template"
+ )
+ cfg.datasets[idx].conversation = "chatml"
+ if ds_cfg.type == "orpo.chat_template" and not ds_cfg.chat_template:
+ LOG.info(
+ f"updating dataset {ds_cfg.path} with `chat_template: chatml` to match your chat_template"
+ )
+ cfg.datasets[idx].chat_template = "chatml"
+
+
+def validate_config(cfg: DictDefault, capabilities: Optional[dict] = None):
+ if capabilities:
+ return DictDefault(
+ dict(
+ AxolotlConfigWCapabilities(
+ **cfg.to_dict(), capabilities=capabilities
+ ).model_dump(exclude_none=True)
+ )
+ )
+ return DictDefault(
+ dict(AxolotlInputConfig(**cfg.to_dict()).model_dump(exclude_none=True))
+ )
+
+
+def legacy_validate_config(cfg):
+ """
+ This is a "pre-validation" step that handles the yaml configuration before we have any
+ information about the model architecture
+ """
+ if is_torch_bf16_gpu_available():
+ if not cfg.bf16 and not cfg.bfloat16:
+ LOG.info("bf16 support detected, but not enabled for this configuration.")
+ else:
+ if (
+ not cfg.merge_lora
+ and not cfg.is_preprocess
+ and (cfg.bf16 is True or cfg.bfloat16 is True)
+ ):
+ raise ValueError(
+ "bf16 requested, but AMP is not supported on this GPU. Requires Ampere series or above."
+ )
+ if (
+ # pylint: disable=too-many-boolean-expressions
+ not (cfg.bf16 or cfg.bfloat16)
+ and (cfg.fp16 or cfg.float16)
+ and not cfg.adapter
+ and not cfg.flash_attention
+ and cfg.sample_packing
+ ):
+ LOG.warning(
+ "Full fine tune w/o FA2 w/ sample packing and fp16/float16 is likely to raise errors. Try LoRA."
+ )
+ # ValueError: Attempting to unscale FP16 gradients.
+ # OR
+ # RuntimeError: expected mat1 and mat2 to have the same dtype, but got: float != c10::Half
+ if cfg.max_packed_sequence_len:
+ raise DeprecationWarning("`max_packed_sequence_len` is no longer supported")
+
+ if cfg.sample_packing and cfg.rl:
+ raise ValueError("`sample_packing: true` does not work with RLHF training")
+
+ if cfg.sample_packing and not cfg.pad_to_sequence_len:
+ LOG.warning(
+ "`pad_to_sequence_len: true` is recommended when using sample_packing"
+ )
+
+ if cfg.gradient_accumulation_steps and cfg.batch_size:
+ raise ValueError(
+ "please set only one of gradient_accumulation_steps or batch_size"
+ )
+ if cfg.batch_size:
+ LOG.warning(
+ "%s\n%s",
+ "batch_size is not recommended. Please use gradient_accumulation_steps instead.",
+ "To calculate the equivalent gradient_accumulation_steps, divide batch_size / micro_batch_size / number of gpus.",
+ )
+ if (
+ cfg.eval_batch_size
+ and cfg.micro_batch_size
+ and cfg.eval_batch_size != cfg.micro_batch_size
+ ):
+ LOG.warning(
+ "eval_batch_size != micro_batch_size. This can lead to VRAM instability."
+ )
+
+ if cfg.adapter == "qlora":
+ if cfg.merge_lora:
+ # can't merge qlora if loaded in 8bit or 4bit
+ if cfg.load_in_8bit:
+ raise ValueError("Can't merge qlora if loaded in 8bit")
+
+ if cfg.gptq:
+ raise ValueError("Can't merge qlora if gptq")
+
+ if cfg.load_in_4bit:
+ raise ValueError("Can't merge qlora if loaded in 4bit")
+
+ else:
+ if cfg.load_in_8bit:
+ raise ValueError("Can't load qlora in 8bit")
+
+ if cfg.gptq:
+ raise ValueError("Can't load qlora if gptq")
+
+ if not cfg.load_in_4bit:
+ raise ValueError("Require cfg.load_in_4bit to be True for qlora")
+
+ if cfg.flash_attn_fuse_qkv or cfg.flash_attn_fuse_mlp:
+ raise ValueError("Fused modules are not supported with QLoRA")
+
+ loftq = cfg.peft and cfg.peft.loftq_config and cfg.peft.loftq_config.loftq_bits
+ if not cfg.load_in_8bit and cfg.adapter == "lora" and not loftq:
+ LOG.warning("We recommend setting `load_in_8bit: true` for LORA finetuning")
+
+ if cfg.adapter == "lora" and (cfg.flash_attn_fuse_qkv or cfg.flash_attn_fuse_mlp):
+ raise ValueError("Fused modules are not supported with LoRA")
+
+ if cfg.adapter and cfg.peft_layers_to_transform and cfg.unfrozen_parameters:
+ raise ValueError(
+ "`unfrozen_parameters` used with `peft_layers_to_transform` can have unexpected behavior."
+ )
+
+ if cfg.relora_steps:
+ if cfg.adapter not in ("lora", "qlora"):
+ raise ValueError("cfg.adapter must be lora or qlora to use ReLoRA")
+
+ if cfg.fsdp:
+ raise ValueError("fsdp not supported with ReLoRA")
+
+ if cfg.deepspeed:
+ raise ValueError("deepspeed not supported with ReLoRA")
+
+ if cfg.lr_scheduler == "one_cycle":
+ raise ValueError("ReLoRA is not compatible with the one_cycle scheduler")
+
+ if cfg.flash_attn_fuse_qkv or cfg.flash_attn_fuse_mlp:
+ raise ValueError("Fused modules are not supported with ReLoRA")
+
+ if cfg.trust_remote_code:
+ LOG.warning(
+ "`trust_remote_code` is set to true. Please make sure that you reviewed the remote code/model."
+ )
+
+ if cfg.push_dataset_to_hub and cfg.hf_use_auth_token is not True:
+ raise ValueError(
+ "Require cfg.hf_use_auth_token to be True for push_dataset_to_hub"
+ )
+
+ if (cfg.base_model and "falcon" in cfg.base_model.lower()) and cfg.fsdp:
+ raise ValueError("FSDP is not supported for falcon models")
+
+ if (
+ cfg.base_model and "mpt" in cfg.base_model.lower()
+ ) and cfg.gradient_checkpointing:
+ raise ValueError("gradient_checkpointing is not supported for MPT models")
+
+ if cfg.flash_optimum is True:
+ if cfg.adapter:
+ LOG.warning("BetterTransformers probably doesn't work with PEFT adapters")
+ if cfg.fp16 or cfg.bf16:
+ raise ValueError("AMP is not supported with BetterTransformer")
+ if cfg.float16 is not True and cfg.bfloat16 is not True:
+ LOG.warning(
+ "You should probably set bfloat16 or float16 to true to "
+ "load the model in float16 for BetterTransformers"
+ )
+ if int(torch.__version__.split(".", maxsplit=1)[0]) < 2:
+ LOG.warning("torch>=2.0.0 required")
+ raise ValueError(
+ f"flash_optimum for BetterTransformers may not be used with {torch.__version__}"
+ )
+
+ if cfg.pretraining_dataset and cfg.group_by_length:
+ LOG.warning(
+ "You probably want to disable group_by_length as it will force a streamed dataset to download completely."
+ )
+ if cfg.pretraining_dataset and not cfg.max_steps:
+ raise ValueError(
+ "max_steps must be set when using iterable pretraining_dataset, Trainer can't infer length and schedule optimizer/learning rate without it!"
+ )
+
+ if any([cfg.adam_beta1, cfg.adam_beta2, cfg.adam_epsilon]) and (
+ not cfg.optimizer or "adamw" not in cfg.optimizer
+ ):
+ LOG.warning("adamw hyperparameters found, but no adamw optimizer set")
+
+ if cfg.push_to_hub_model_id:
+ raise ValueError(
+ "push_to_hub_model_id is deprecated. Please use hub_model_id instead."
+ )
+
+ if cfg.hub_model_id and cfg.save_strategy not in ["steps", "epoch", None]:
+ LOG.warning(
+ "hub_model_id is set without any models being saved. To save a model, set save_strategy to steps, epochs or leave empty."
+ )
+
+ if cfg.gptq and cfg.revision_of_model:
+ raise ValueError(
+ "revision_of_model is not supported for GPTQ models. "
+ + "Please download the model from HuggingFace Hub manually for correct branch, "
+ + "point to its path, and remove revision_of_model from the config."
+ )
+
+ # if cfg.sample_packing and cfg.sdp_attention:
+ # # incompatible due to bug w/ accelerate causing 0.0 loss when using llama2
+ # raise ValueError(
+ # "sample_packing not compatible with sdp_attention. Use flash_attention"
+ # )
+
+ if cfg.sample_packing and cfg.xformers_attention:
+ raise ValueError(
+ "sample_packing not compatible with xformers_attention. Use flash_attention"
+ )
+
+ if cfg.sample_packing and cfg.sdp_attention and (cfg.bfloat16 or cfg.bf16):
+ # https://github.com/pytorch/pytorch/blob/1b03423526536b5f3d35bdfa95ccc6197556cf9b/test/test_transformers.py#L2440-L2450
+ LOG.warning(
+ "sample_packing & torch sdpa with bf16 is unsupported may results in 0.0 loss. "
+ "This may work on H100s."
+ )
+
+ if cfg.early_stopping_patience:
+ if not cfg.save_steps or not cfg.eval_steps:
+ raise ValueError(
+ "`early_stopping_patience` requires save_steps and eval_steps to be set. eval_steps should evenly divide save_steps."
+ )
+ if cfg.save_steps % cfg.eval_steps != 0:
+ raise ValueError(
+ "`early_stopping_patience` requires that eval_steps should evenly divide save_steps."
+ )
+
+ if cfg.datasets:
+ for idx, ds_cfg in enumerate(cfg.datasets):
+ if not ds_cfg.type:
+ continue
+ if ds_cfg.type == "sharegpt:chat":
+ LOG.warning(
+ PendingDeprecationWarning(
+ "`type: sharegpt:chat` will soon be deprecated. simply use `type: sharegpt` instead."
+ )
+ )
+ cfg.datasets[idx].type = "sharegpt"
+ if "sharegpt_simple" in ds_cfg.type:
+ LOG.warning(
+ PendingDeprecationWarning(
+ "`type: sharegpt_simple` will soon be deprecated. simply use `type: sharegpt` instead."
+ )
+ )
+ cfg.datasets[idx].type = cfg.datasets[idx].type.replace(
+ "sharegpt_simple", "sharegpt"
+ )
+
+ if cfg.saves_per_epoch and cfg.save_steps:
+ raise ValueError(
+ "save_steps and saves_per_epoch are mutually exclusive and cannot be used together."
+ )
+ if cfg.save_strategy and cfg.saves_per_epoch and cfg.save_strategy != "steps":
+ raise ValueError(
+ "save_strategy must be empty or set to `steps` when used with saves_per_epoch."
+ )
+ if cfg.save_strategy and cfg.save_steps and cfg.save_strategy != "steps":
+ raise ValueError(
+ "save_strategy and save_steps mismatch. Please set save_strategy to 'steps' or remove save_steps."
+ )
+ if cfg.evals_per_epoch and cfg.eval_steps:
+ raise ValueError(
+ "eval_steps and evals_per_epoch are mutually exclusive and cannot be used together."
+ )
+ if (
+ cfg.evals_per_epoch
+ and cfg.evaluation_strategy
+ and cfg.evaluation_strategy != "steps"
+ ):
+ raise ValueError(
+ "evaluation_strategy must be empty or set to `steps` when used with evals_per_epoch."
+ )
+ if (
+ cfg.evaluation_strategy
+ and cfg.eval_steps
+ and cfg.evaluation_strategy != "steps"
+ ):
+ raise ValueError(
+ "evaluation_strategy and eval_steps mismatch. Please set evaluation_strategy to 'steps' or remove eval_steps."
+ )
+
+ if (
+ cfg.val_set_size == 0
+ and (cfg.eval_steps or cfg.evaluation_strategy)
+ and not cfg.test_datasets
+ ):
+ raise ValueError(
+ "eval_steps and evaluation_strategy are not supported with val_set_size == 0"
+ )
+
+ if (
+ cfg.sample_packing
+ and cfg.eval_table_size
+ and cfg.eval_sample_packing is not False
+ ):
+ raise ValueError(
+ "eval_table_size and eval_sample_packing are not supported together with sample_packing. Please set 'eval_sample_packing' to false."
+ )
+
+ if not cfg.adapter and (cfg.load_in_8bit or cfg.load_in_4bit):
+ raise ValueError(
+ "load_in_8bit and load_in_4bit are not supported without setting an adapter."
+ "If you want to full finetune, please turn off load_in_8bit and load_in_4bit."
+ )
+
+ if cfg.rope_scaling:
+ LOG.warning("`rope_scaling` should now be be a key under `model_config`")
+
+ if cfg.wandb_run_id and not cfg.wandb_name:
+ cfg.wandb_name = cfg.wandb_run_id
+
+ LOG.warning(
+ "wandb_run_id sets the ID of the run. If you would like to set the name, please use wandb_name instead."
+ )
+
+ if cfg.noisy_embedding_alpha is not None:
+ # Deprecated, use neftune_noise_alpha
+ LOG.warning("noisy_embedding_alpha is deprecated, use neftune_noise_alpha")
+ if cfg.neftune_noise_alpha is None:
+ cfg.neftune_noise_alpha = cfg.noisy_embedding_alpha
+ else:
+ # User is providing both; bail and have them sort out their settings
+ raise ValueError(
+ "noisy_embedding_alpha is deprecated, use neftune_noise_alpha; both are set, please remove the deprecated noisy_embedding_alpha setting"
+ )
+
+ if cfg.neftune_noise_alpha is not None and cfg.neftune_noise_alpha <= 0.0:
+ raise ValueError("neftune_noise_alpha must be > 0.0")
+
+ if cfg.max_memory is not None and cfg.gpu_memory_limit is not None:
+ raise ValueError(
+ "max_memory and gpu_memory_limit are mutually exclusive and cannot be used together."
+ )
+
+ if (
+ cfg.unfrozen_parameters
+ and cfg.gradient_checkpointing_kwargs
+ and cfg.gradient_checkpointing_kwargs.use_reentrant is True
+ ):
+ # https://github.com/huggingface/transformers/issues/21381
+ raise ValueError(
+ "`use_reentrant` must be false when used with partially frozen model."
+ )
+
+ if cfg.deepspeed and Path(cfg.deepspeed).is_file():
+ with open(cfg.deepspeed, encoding="utf-8") as file:
+ contents = file.read()
+ deepspeed_cfg: DictDefault = DictDefault(json.loads(contents))
+ if cfg.flash_attention:
+ if (
+ deepspeed_cfg.zero_optimization
+ and deepspeed_cfg.zero_optimization.stage == 3
+ ):
+ if not (
+ (
+ deepspeed_cfg.bf16
+ and deepspeed_cfg.bf16.enabled # pylint: disable=no-member
+ is True
+ )
+ or (
+ deepspeed_cfg.fp16
+ and deepspeed_cfg.fp16.enabled # pylint: disable=no-member
+ is True
+ )
+ ):
+ raise ValueError(
+ "bf16.enabled or fp16.enabled must be set to true when using ZeRO-3 with flash-attention"
+ )
+ if "8bit" in cfg.optimizer and deepspeed_cfg.optimizer:
+ LOG.warning(
+ f"conflicting optimizer: {cfg.optimizer} used alongside deepspeed optimizer."
+ )
+
+ if cfg.test_datasets and cfg.val_set_size:
+ raise ValueError(
+ "non-zero val_set_size should not be used with test_datasets configuration"
+ )
+
+ if cfg.fsdp and "bnb" in cfg.optimizer:
+ raise ValueError(f"FSDP not compatible with {cfg.optimizer}")
+
+ if cfg.do_causal_lm_eval and cfg.eval_sample_packing:
+ raise ValueError(
+ "do_causal_lm_eval is enabled, eval_sample_packing must be set to False"
+ )
+
+ if cfg.eval_causal_lm_metrics:
+ supported_metrics = ["sacrebleu", "comet", "ter", "chrf"]
+ if not isinstance(cfg.eval_causal_lm_metrics, list):
+ raise ValueError("eval_causal_lm_metrics must be a list")
+ # only ["sacrebleu", "comet", "ter", "chrf"] supported
+ if set(cfg.eval_causal_lm_metrics) - set(supported_metrics):
+ raise ValueError(
+ f"eval_causal_lm_metrics must be one of {supported_metrics}"
+ )
+
+ # TODO
+ # MPT 7b
+ # https://github.com/facebookresearch/bitsandbytes/issues/25
+ # no 8bit adaAmw w bf16
+
+ # GPT-NeoX
+ # evals broken when extending context len
+ # File "/root/miniconda3/envs/py3.9/lib/python3.9/site-packages/transformers/models/gpt_neox/modeling_gpt_neox.py", line 162, in forward attn_output, attn_weights = self._attn(query, key, value, attention_mask, head_mask)
+ # File "/root/miniconda3/envs/py3.9/lib/python3.9/site-packages/optimum/bettertransformer/models/attention.py", line 74, in gpt2_wrapped_scaled_dot_product
+ # attention_mask = causal_mask + attention_mask
+ # RuntimeError: The size of tensor a (2048) must match the size of tensor b (8132) at non-singleton dimension 3
diff --git a/src/axolotl/utils/config/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/config/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..cc514966806c93fcdf39f49729efbe82de6fcf44
Binary files /dev/null and b/src/axolotl/utils/config/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/config/models/__init__.py b/src/axolotl/utils/config/models/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/axolotl/utils/config/models/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/config/models/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..dd415b6e5c22eb567719ef6fc16795c4efc4b858
Binary files /dev/null and b/src/axolotl/utils/config/models/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/config/models/input/__init__.py b/src/axolotl/utils/config/models/input/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/axolotl/utils/config/models/input/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/config/models/input/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..f6fc736dd642264f4fac4ce7907925e94bd48c0d
Binary files /dev/null and b/src/axolotl/utils/config/models/input/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/config/models/input/next/__init__.py b/src/axolotl/utils/config/models/input/next/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/src/axolotl/utils/config/models/input/v0_4_1/__init__.py b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..c0a3ff4c2aaff79104c86d5ec9aa91ffd21abe38
--- /dev/null
+++ b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py
@@ -0,0 +1,1091 @@
+"""
+Module for pydantic models for configuration
+"""
+
+# pylint: disable=too-many-lines
+
+import logging
+import os
+from enum import Enum
+from typing import Any, Dict, List, Literal, Optional, Tuple, Union
+
+from pydantic import BaseModel, Field, conlist, field_validator, model_validator
+from transformers import SchedulerType
+from transformers.training_args import OptimizerNames
+
+from axolotl.utils.config.models.internals import GPUCapabilities
+
+LOG = logging.getLogger("axolotl.utils.config.models.input")
+
+
+class DeprecatedParameters(BaseModel):
+ """configurations that are deprecated"""
+
+ max_packed_sequence_len: Optional[int] = None
+ rope_scaling: Optional[Any] = None
+ noisy_embedding_alpha: Optional[float] = None
+
+ @field_validator("max_packed_sequence_len")
+ @classmethod
+ def validate_max_packed_sequence_len(cls, max_packed_sequence_len):
+ if max_packed_sequence_len:
+ raise DeprecationWarning("`max_packed_sequence_len` is no longer supported")
+ return max_packed_sequence_len
+
+ @field_validator("rope_scaling")
+ @classmethod
+ def validate_rope_scaling(cls, rope_scaling):
+ if rope_scaling:
+ raise DeprecationWarning(
+ "`rope_scaling` is no longer supported, it should now be be a key under `model_config`"
+ )
+ return rope_scaling
+
+ @field_validator("noisy_embedding_alpha")
+ @classmethod
+ def validate_noisy_embedding_alpha(cls, noisy_embedding_alpha):
+ if noisy_embedding_alpha:
+ LOG.warning("noisy_embedding_alpha is deprecated, use neftune_noise_alpha")
+ return noisy_embedding_alpha
+
+
+class RemappedParameters(BaseModel):
+ """parameters that have been remapped to other names"""
+
+ overrides_of_model_config: Optional[Dict[str, Any]] = Field(
+ default=None, alias="model_config"
+ )
+ type_of_model: Optional[str] = Field(default=None, alias="model_type")
+ revision_of_model: Optional[str] = Field(default=None, alias="model_revision")
+
+
+class PretrainingDataset(BaseModel):
+ """pretraining dataset configuration subset"""
+
+ name: Optional[str] = None
+ path: Optional[str] = None
+ split: Optional[str] = "train"
+ text_column: Optional[str] = "text"
+ type: Optional[str] = "pretrain"
+
+
+class UserDefinedPrompterType(BaseModel):
+ """structure for user defined prompt types"""
+
+ system_prompt: Optional[str] = None
+ system_format: Optional[str] = None
+ field_system: Optional[str] = None
+ field_instruction: Optional[str] = None
+ field_input: Optional[str] = None
+ field_output: Optional[str] = None
+
+ format: Optional[str] = None
+ no_input_format: Optional[str] = None
+ field: Optional[str] = None
+
+
+class SFTDataset(BaseModel):
+ """SFT configuration subset"""
+
+ path: Optional[str] = None
+ split: Optional[str] = None
+ type: Optional[Union[str, UserDefinedPrompterType]] = None
+ shards: Optional[int] = None
+ conversation: Optional[str] = None
+ chat_template: Optional[str] = None
+ data_files: Optional[Union[str, List[str]]] = None
+ name: Optional[str] = None
+ ds_type: Optional[str] = None
+ train_on_split: Optional[str] = None
+
+ field: Optional[str] = None
+ field_human: Optional[str] = None
+ field_model: Optional[str] = None
+
+ roles: Optional[Dict[str, List[str]]] = None
+
+
+class UserDefinedDPOType(BaseModel):
+ """User defined typing for DPO"""
+
+ field_system: Optional[str] = None
+ field_prompt: Optional[str] = None
+ field_chosen: Optional[str] = None
+ field_rejected: Optional[str] = None
+ prompt_format: Optional[str] = None
+ chosen_format: Optional[str] = None
+ rejected_format: Optional[str] = None
+
+
+class DPODataset(BaseModel):
+ """DPO configuration subset"""
+
+ path: Optional[str] = None
+ split: Optional[str] = None
+ type: Optional[Union[UserDefinedDPOType, str]] = None
+ data_files: Optional[List[str]] = None
+
+
+class RLType(str, Enum):
+ """RL trainer type configuration subset"""
+
+ dpo = "dpo" # pylint: disable=invalid-name
+ ipo = "ipo" # pylint: disable=invalid-name
+ kto_pair = "kto_pair" # pylint: disable=invalid-name
+ orpo = "orpo" # pylint: disable=invalid-name
+
+
+class ChatTemplate(str, Enum):
+ """Chat templates configuration subset"""
+
+ alpaca = "alpaca" # pylint: disable=invalid-name
+ chatml = "chatml" # pylint: disable=invalid-name
+ inst = "inst" # pylint: disable=invalid-name
+ gemma = "gemma" # pylint: disable=invalid-name
+ cohere = "cohere" # pylint: disable=invalid-name
+ llama3 = "llama3" # pylint: disable=invalid-name
+
+class LoftQConfig(BaseModel):
+ """LoftQ configuration subset"""
+
+ loftq_bits: int = Field(default=4, metadata={"help": "Quantization bits for LoftQ"})
+ # loftq_iter: int = Field(default=1, metadata={"help": "Alternating iterations for LoftQ"})
+
+
+class PeftConfig(BaseModel):
+ """peftq configuration subset"""
+
+ loftq_config: Optional[LoftQConfig] = None
+
+
+class SpecialTokensConfig(BaseModel):
+ """Special tokens configuration subset"""
+
+ bos_token: Optional[str] = None
+ eos_token: Optional[str] = None
+ pad_token: Optional[str] = None
+ unk_token: Optional[str] = None
+ additional_special_tokens: Optional[List[str]] = None
+
+
+class LoraConfig(BaseModel):
+ """Peft / LoRA configuration subset"""
+
+ load_in_8bit: Optional[bool] = Field(default=False)
+ load_in_4bit: Optional[bool] = Field(default=False)
+
+ adapter: Optional[str] = None
+ lora_model_dir: Optional[str] = None
+ lora_r: Optional[int] = None
+ lora_alpha: Optional[int] = None
+ lora_fan_in_fan_out: Optional[bool] = None
+ lora_target_modules: Optional[List[str]] = None
+ lora_target_linear: Optional[bool] = None
+ lora_modules_to_save: Optional[List[str]] = None
+ lora_dropout: Optional[float] = None
+ peft_layers_to_transform: Optional[List[int]] = None
+ peft: Optional[PeftConfig] = None
+ peft_use_dora: Optional[bool] = None
+ peft_use_rslora: Optional[bool] = None
+ peft_layer_replication: Optional[List[Tuple[int, int]]] = None
+
+ lora_on_cpu: Optional[bool] = None
+ gptq: Optional[bool] = None
+ bnb_config_kwargs: Optional[Dict[str, Any]] = None
+
+ loraplus_lr_ratio: Optional[float] = Field(
+ default=None,
+ metadata={
+ "help": "loraplus learning rate ratio lr_B / lr_A. Recommended value is 2^4."
+ },
+ )
+ loraplus_lr_embedding: Optional[float] = Field(
+ default=1e-6,
+ metadata={"help": "loraplus learning rate for lora embedding layers."},
+ )
+
+ merge_lora: Optional[bool] = None
+
+ @model_validator(mode="before")
+ @classmethod
+ def validate_adapter(cls, data):
+ if not data.get("adapter") and (
+ data.get("load_in_8bit") or data.get("load_in_4bit")
+ ):
+ raise ValueError(
+ "load_in_8bit and load_in_4bit are not supported without setting an adapter."
+ "If you want to full finetune, please turn off load_in_8bit and load_in_4bit."
+ )
+ return data
+
+ @model_validator(mode="after")
+ def validate_qlora(self):
+ if self.adapter == "qlora":
+ if self.merge_lora:
+ # can't merge qlora if loaded in 8bit or 4bit
+ if self.load_in_8bit:
+ raise ValueError("Can't merge qlora if loaded in 8bit")
+
+ if self.gptq:
+ raise ValueError("Can't merge qlora if gptq")
+
+ if self.load_in_4bit:
+ raise ValueError("Can't merge qlora if loaded in 4bit")
+
+ else:
+ if self.load_in_8bit:
+ raise ValueError("Can't load qlora in 8bit")
+
+ if self.gptq:
+ raise ValueError("Can't load qlora if gptq")
+
+ if not self.load_in_4bit:
+ raise ValueError("Require cfg.load_in_4bit to be True for qlora")
+ return self
+
+
+class ReLoRAConfig(BaseModel):
+ """ReLoRA configuration subset"""
+
+ relora_steps: Optional[int] = None
+ relora_warmup_steps: Optional[int] = None
+ relora_anneal_steps: Optional[int] = None
+ relora_prune_ratio: Optional[float] = None
+ relora_cpu_offload: Optional[bool] = None
+
+
+class ModelInputConfig(BaseModel):
+ """model to train on configuration subset"""
+
+ base_model: str
+ base_model_config: Optional[str] = None
+ cls_model_config: Optional[str] = None
+ tokenizer_config: Optional[str] = None
+ tokenizer_use_fast: Optional[bool] = None
+ tokenizer_legacy: Optional[bool] = None
+ tokenizer_type: Optional[str] = Field(
+ default=None, metadata={"help": "transformers tokenizer class"}
+ )
+ trust_remote_code: Optional[bool] = None
+
+ @field_validator("trust_remote_code")
+ @classmethod
+ def hint_trust_remote_code(cls, trust_remote_code):
+ if trust_remote_code:
+ LOG.warning(
+ "`trust_remote_code` is set to true. Please make sure that you reviewed the remote code/model."
+ )
+ return trust_remote_code
+
+
+class HyperparametersConfig(BaseModel):
+ """training hyperparams configuration subset"""
+
+ gradient_accumulation_steps: Optional[int] = Field(default=1)
+ micro_batch_size: Optional[int] = Field(
+ default=1,
+ metadata={"help": "per gpu micro batch size for training"},
+ )
+ batch_size: Optional[int] = Field(
+ default=None,
+ metadata={
+ "help": "Total batch size, we do not recommended setting this manually"
+ },
+ )
+ eval_batch_size: Optional[int] = Field(
+ default=None,
+ metadata={
+ "help": "per gpu micro batch size for evals, defaults to value of micro_batch_size"
+ },
+ )
+
+ train_on_inputs: Optional[bool] = False
+ group_by_length: Optional[bool] = None
+
+ learning_rate: Union[str, float]
+ weight_decay: Optional[float] = 0.0
+ optimizer: Optional[
+ Union[OptimizerNames, Literal["lion_pytorch"]]
+ ] = OptimizerNames.ADAMW_HF.value
+ optim_args: Optional[Union[str, Dict[str, Any]]] = Field(
+ default=None, metadata={"help": "Optional arguments to supply to optimizer."}
+ )
+ optim_target_modules: Optional[Union[List[str], Literal["all_linear"]]] = Field(
+ default=None,
+ metadata={
+ "help": "The target modules to optimize, i.e. the module names that you would like to train."
+ },
+ )
+ torchdistx_path: Optional[str] = None
+ lr_scheduler: Optional[SchedulerType] = "cosine"
+ lr_scheduler_kwargs: Optional[Dict[str, Any]] = None
+ lr_quadratic_warmup: Optional[bool] = None
+ cosine_min_lr_ratio: Optional[float] = None
+ cosine_constant_lr_ratio: Optional[float] = None
+ lr_div_factor: Optional[float] = None
+
+ adam_epsilon: Optional[float] = None
+ adam_beta1: Optional[float] = None
+ adam_beta2: Optional[float] = None
+ max_grad_norm: Optional[float] = None
+ num_epochs: int = Field(default=1)
+
+ @field_validator("batch_size")
+ @classmethod
+ def hint_batch_size_set(cls, batch_size):
+ if batch_size:
+ LOG.warning(
+ "%s\n%s",
+ "batch_size is not recommended. Please use gradient_accumulation_steps instead.",
+ "To calculate the equivalent gradient_accumulation_steps, divide batch_size / micro_batch_size / number of gpus.",
+ )
+ return batch_size
+
+ @field_validator("learning_rate")
+ @classmethod
+ def convert_learning_rate(cls, learning_rate):
+ if learning_rate and isinstance(learning_rate, str):
+ learning_rate = float(learning_rate)
+ return learning_rate
+
+
+class ModelOutputConfig(BaseModel):
+ """model save configuration subset"""
+
+ output_dir: str = Field(default="./model-out")
+ hub_model_id: Optional[str] = None
+ hub_strategy: Optional[str] = None
+ save_safetensors: Optional[bool] = None
+
+
+class MLFlowConfig(BaseModel):
+ """mlflow configuration subset"""
+
+ use_mlflow: Optional[bool] = None
+ mlflow_tracking_uri: Optional[str] = None
+ mlflow_experiment_name: Optional[str] = None
+ hf_mlflow_log_artifacts: Optional[bool] = None
+
+
+class LISAConfig(BaseModel):
+ """LISA options"""
+
+ lisa_n_layers: Optional[int] = Field(
+ default=None,
+ metadata={"help": "the number of activate layers in LISA"},
+ )
+ lisa_step_interval: Optional[int] = Field(
+ default=None,
+ metadata={"help": "how often to switch layers in LISA"},
+ )
+ lisa_layers_attribute: Optional[str] = Field(
+ default="model.layers",
+ metadata={"help": "path under the model to access the layers"},
+ )
+
+
+class WandbConfig(BaseModel):
+ """wandb configuration subset"""
+
+ use_wandb: Optional[bool] = None
+ wandb_name: Optional[str] = None
+ wandb_run_id: Optional[str] = None
+ wandb_mode: Optional[str] = None
+ wandb_project: Optional[str] = None
+ wandb_entity: Optional[str] = None
+ wandb_watch: Optional[str] = None
+ wandb_log_model: Optional[str] = None
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_wandb_run(cls, data):
+ if data.get("wandb_run_id") and not data.get("wandb_name"):
+ data["wandb_name"] = data.get("wandb_run_id")
+
+ LOG.warning(
+ "wandb_run_id sets the ID of the run. If you would like to set the name, please use wandb_name instead."
+ )
+
+ return data
+
+
+class GradioConfig(BaseModel):
+ """Gradio configuration subset"""
+
+ gradio_title: Optional[str] = None
+ gradio_share: Optional[bool] = None
+ gradio_server_name: Optional[str] = None
+ gradio_server_port: Optional[int] = None
+ gradio_max_new_tokens: Optional[int] = None
+ gradio_temperature: Optional[float] = None
+
+
+# pylint: disable=too-many-public-methods,too-many-ancestors
+class AxolotlInputConfig(
+ ModelInputConfig,
+ ModelOutputConfig,
+ LoraConfig,
+ ReLoRAConfig,
+ HyperparametersConfig,
+ WandbConfig,
+ MLFlowConfig,
+ LISAConfig,
+ GradioConfig,
+ RemappedParameters,
+ DeprecatedParameters,
+ BaseModel,
+):
+ """wrapper of all config options"""
+
+ class Config:
+ """Config for alias"""
+
+ populate_by_name = True
+
+ strict: Optional[bool] = Field(default=False)
+ resume_from_checkpoint: Optional[str] = None
+ auto_resume_from_checkpoints: Optional[bool] = None
+ resize_token_embeddings_to_32x: Optional[bool] = None
+
+ rl: Optional[RLType] = None
+
+ datasets: Optional[conlist(Union[SFTDataset, DPODataset], min_length=1)] = None # type: ignore
+ test_datasets: Optional[conlist(Union[SFTDataset, DPODataset], min_length=1)] = None # type: ignore
+ shuffle_merged_datasets: Optional[bool] = True
+ dataset_prepared_path: Optional[str] = None
+ dataset_shard_num: Optional[int] = None
+ dataset_shard_idx: Optional[int] = None
+
+ pretraining_dataset: Optional[ # type: ignore
+ conlist(Union[PretrainingDataset, SFTDataset], min_length=1)
+ ] = Field(
+ default=None, metadata={"help": {"streaming dataset to use for pretraining"}}
+ )
+ dataset_processes: Optional[int] = Field(default=os.cpu_count())
+ dataset_keep_in_memory: Optional[bool] = None
+ dataloader_pin_memory: Optional[bool] = None
+ dataloader_num_workers: Optional[int] = None
+ dataloader_prefetch_factor: Optional[int] = None
+ dataloader_drop_last: Optional[bool] = None
+
+ remove_unused_columns: Optional[bool] = None
+
+ push_dataset_to_hub: Optional[str] = None
+ hf_use_auth_token: Optional[bool] = None
+
+ device: Optional[Any] = None
+ device_map: Optional[Any] = None
+ world_size: Optional[int] = None
+ local_rank: Optional[int] = None
+ ddp: Optional[bool] = None
+
+ seed: Optional[int] = None
+ ddp_timeout: Optional[int] = None
+ ddp_bucket_cap_mb: Optional[int] = None
+ ddp_broadcast_buffers: Optional[bool] = None
+ ddp_find_unused_parameters: Optional[bool] = None
+
+ eval_table_size: Optional[int] = None
+ eval_max_new_tokens: Optional[int] = None
+ do_causal_lm_eval: Optional[bool] = None
+ eval_causal_lm_metrics: Optional[List[str]] = None
+ do_bench_eval: Optional[bool] = None
+ bench_dataset: Optional[str] = None
+ bench_split: Optional[str] = None
+ metric_for_best_model: Optional[str] = None
+ greater_is_better: Optional[bool] = None
+
+ loss_watchdog_threshold: Optional[float] = None
+ loss_watchdog_patience: Optional[int] = None
+
+ bf16: Optional[Union[Literal["auto"], bool]] = "auto"
+ fp16: Optional[bool] = None
+ bfloat16: Optional[bool] = None # for non-AMP cases
+ float16: Optional[bool] = None # for non-AMP cases
+ tf32: Optional[bool] = None
+ float32: Optional[bool] = None
+
+ # torch_dtype: Optional[torch.dtype]
+
+ gradient_checkpointing: Optional[Union[Literal["unsloth"], bool]] = Field(
+ default=False
+ )
+ gradient_checkpointing_kwargs: Optional[Dict[str, Any]] = None
+
+ unfrozen_parameters: Optional[List[str]] = None
+
+ sequence_len: int = Field(default=512)
+ min_sample_len: Optional[int] = None
+ sample_packing: Optional[bool] = None
+ eval_sample_packing: Optional[bool] = None
+ pad_to_sequence_len: Optional[bool] = None
+ curriculum_sampling: Optional[bool] = None
+
+ # for PoSE context length extension
+ use_pose: Optional[bool] = None
+ pose_split_on_token_ids: Optional[List[int]] = None
+ pose_max_context_len: Optional[int] = None
+ pose_num_chunks: Optional[int] = None
+
+ pretrain_multipack_buffer_size: Optional[int] = 10_000
+ pretrain_multipack_attn: Optional[bool] = Field(
+ default=True,
+ metadata={
+ "help": "whether to prevent cross attention for packed sequences during pretraining",
+ },
+ )
+
+ xformers_attention: Optional[bool] = None
+ sdp_attention: Optional[bool] = None
+ s2_attention: Optional[bool] = None
+ flash_attention: Optional[bool] = None
+ flash_attn_cross_entropy: Optional[bool] = None
+ flash_attn_rms_norm: Optional[bool] = None
+ flash_attn_fuse_qkv: Optional[bool] = None
+ flash_attn_fuse_mlp: Optional[bool] = None
+ flash_optimum: Optional[bool] = None
+
+ unsloth_cross_entropy_loss: Optional[bool] = None
+ unsloth_lora_mlp: Optional[bool] = None
+ unsloth_lora_qkv: Optional[bool] = None
+ unsloth_lora_o: Optional[bool] = None
+
+ deepspeed: Optional[Union[str, Dict[str, Any]]] = None
+ fsdp: Optional[List[str]] = None
+ fsdp_config: Optional[Dict[str, Any]] = None
+
+ val_set_size: Optional[float] = Field(default=0.0)
+
+ special_tokens: Optional[SpecialTokensConfig] = None
+ tokens: Optional[List[str]] = None
+
+ torch_compile: Optional[bool] = None
+ torch_compile_backend: Optional[str] = None
+
+ max_steps: Optional[int] = None
+ warmup_steps: Optional[int] = None
+ warmup_ratio: Optional[float] = None
+ eval_steps: Optional[Union[int, float]] = None
+ evals_per_epoch: Optional[Union[int]] = None
+ evaluation_strategy: Optional[str] = None
+ save_steps: Optional[Union[int, float]] = None
+ saves_per_epoch: Optional[int] = None
+ save_strategy: Optional[str] = None
+ save_total_limit: Optional[int] = None
+ logging_steps: Optional[int] = None
+ early_stopping_patience: Optional[int] = None
+ load_best_model_at_end: Optional[bool] = False
+
+ neftune_noise_alpha: Optional[float] = None
+
+ orpo_alpha: Optional[float] = None
+
+ max_memory: Optional[
+ Dict[Union[int, Literal["cpu", "disk"]], Union[int, str]]
+ ] = None
+ gpu_memory_limit: Optional[Union[int, str]] = None
+ low_cpu_mem_usage: Optional[bool] = None
+
+ chat_template: Optional[ChatTemplate] = None
+ default_system_message: Optional[str] = None
+
+ # INTERNALS - document for now, generally not set externally
+ is_preprocess: Optional[bool] = None
+
+ total_num_tokens: Optional[int] = None
+ total_supervised_tokens: Optional[int] = None
+ sample_packing_eff_est: Optional[float] = None
+ axolotl_config_path: Optional[str] = None
+
+ is_falcon_derived_model: Optional[bool] = Field(default=None)
+ is_llama_derived_model: Optional[bool] = Field(default=None)
+ is_mistral_derived_model: Optional[bool] = Field(default=None)
+ is_qwen_derived_model: Optional[bool] = Field(default=None)
+
+ @field_validator("datasets", mode="before")
+ @classmethod
+ def fix_sharegpt_datasets(cls, datasets):
+ for idx, ds_cfg in enumerate(datasets):
+ if not ds_cfg["type"]:
+ continue
+ if ds_cfg["type"] == "sharegpt:chat":
+ LOG.warning(
+ PendingDeprecationWarning(
+ "`type: sharegpt:chat` will soon be deprecated. simply use `type: sharegpt` instead."
+ )
+ )
+ datasets[idx]["type"] = "sharegpt"
+ if "sharegpt_simple" in ds_cfg["type"]:
+ LOG.warning(
+ PendingDeprecationWarning(
+ "`type: sharegpt_simple` will soon be deprecated. simply use `type: sharegpt` instead."
+ )
+ )
+ datasets[idx]["type"] = datasets[idx]["type"].replace(
+ "sharegpt_simple", "sharegpt"
+ )
+ return datasets
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_batch_size_fields(cls, data):
+ fields = ("micro_batch_size", "gradient_accumulation_steps", "batch_size")
+ non_empty_count = sum(1 for field in fields if data.get(field))
+
+ if non_empty_count < 2:
+ raise ValueError(f"At least two of {', '.join(fields)} must be set")
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_pretraining_w_max_steps(cls, data):
+ if data.get("pretraining_dataset") and not data.get("max_steps"):
+ raise ValueError(
+ "max_steps must be set when using iterable pretraining_dataset, Trainer can't infer length and schedule optimizer/learning rate without it!"
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_pretraining_w_group_by_length(cls, data):
+ if data.get("pretraining_dataset") and data.get("group_by_length"):
+ LOG.warning(
+ "You probably want to disable group_by_length as it will force a streamed dataset to download completely."
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_gptq_w_revision(cls, data):
+ if data.get("gptq") and data.get("revision_of_model"):
+ raise ValueError(
+ "revision_of_model is not supported for GPTQ models. "
+ + "Please download the model from HuggingFace Hub manually for correct branch, "
+ + "point to its path, and remove revision_of_model from the config."
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_sample_packing_w_xformers(cls, data):
+ if data.get("sample_packing") and data.get("xformers_attention"):
+ raise ValueError(
+ "sample_packing not compatible with xformers_attention. Use flash_attention"
+ )
+
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_sample_packing_wo_flash(cls, data):
+ if (
+ data.get("sample_packing")
+ and not data.get("flash_attention")
+ and not data.get("sdp_attention")
+ ):
+ LOG.warning(
+ "sample_packing without flash_attention or sdp_attention does not handle cross-attention."
+ )
+
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_sample_packing_w_rl(cls, data):
+ if data.get("sample_packing") and data.get("rl"):
+ raise ValueError("`sample_packing: true` does not work with RLHF training")
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def hint_sample_packing_padding(cls, data):
+ if data.get("sample_packing") and not data.get("pad_to_sequence_len"):
+ LOG.warning(
+ "`pad_to_sequence_len: true` is recommended when using sample_packing"
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_gas_bsz(cls, data):
+ if data.get("gradient_accumulation_steps") and data.get("batch_size"):
+ raise ValueError(
+ "please set only one of gradient_accumulation_steps or batch_size"
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def hint_eval_train_mbsz(cls, data):
+ if (
+ data.get("eval_batch_size")
+ and data.get("micro_batch_size")
+ and data.get("eval_batch_size") != data.get("micro_batch_size")
+ ):
+ LOG.warning(
+ "eval_batch_size != micro_batch_size. This can lead to VRAM instability."
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_push_ds_auth(cls, data):
+ if (
+ data.get("push_dataset_to_hub")
+ and data.get("hf_use_auth_token") is not True
+ ):
+ raise ValueError(
+ "Require cfg.hf_use_auth_token to be True for push_dataset_to_hub"
+ )
+ return data
+
+ @model_validator(mode="after")
+ def check_falcon_fsdp(self):
+ if (self.base_model and "falcon" in self.base_model.lower()) and self.fsdp:
+ raise ValueError("FSDP is not supported for falcon models")
+ return self
+
+ @model_validator(mode="after")
+ def check_mpt_checkpointing(self):
+ if (
+ self.base_model and "mpt" in self.base_model.lower()
+ ) and self.gradient_checkpointing:
+ raise ValueError("gradient_checkpointing is not supported for MPT models")
+ return self
+
+ @model_validator(mode="after")
+ def check_better_transformers(self):
+ if self.flash_optimum is True:
+ if self.adapter:
+ LOG.warning(
+ "BetterTransformers probably doesn't work with PEFT adapters"
+ )
+ if self.fp16 or self.bf16:
+ raise ValueError("AMP is not supported with BetterTransformer")
+ if self.float16 is not True and self.bfloat16 is not True:
+ LOG.warning(
+ "You should probably set bfloat16 or float16 to true to "
+ "load the model in float16 for BetterTransformers"
+ )
+ return self
+
+ @model_validator(mode="after")
+ def check_adamw_optimizer_params(self):
+ if any([self.adam_beta1, self.adam_beta2, self.adam_epsilon]) and (
+ not self.optimizer or "adamw" not in self.optimizer.value
+ ):
+ LOG.warning("adamw hyperparameters found, but no adamw optimizer set")
+ return self
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_saves(cls, data):
+ if (
+ data.get("save_strategy")
+ and data.get("save_steps")
+ and data.get("save_strategy") != "steps"
+ ):
+ raise ValueError(
+ "save_strategy and save_steps mismatch. Please set save_strategy to 'steps' or remove save_steps."
+ )
+ if data.get("saves_per_epoch") and data.get("save_steps"):
+ raise ValueError(
+ "save_steps and saves_per_epoch are mutually exclusive and cannot be used together."
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_push_save(cls, data):
+ if data.get("hub_model_id") and (
+ data.get("save_strategy") not in ["steps", "epoch", None]
+ ):
+ LOG.warning(
+ "hub_model_id is set without any models being saved. To save a model, set save_strategy."
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_evals(cls, data):
+ if (
+ data.get("evaluation_strategy")
+ and data.get("eval_steps")
+ and data.get("evaluation_strategy") != "steps"
+ ):
+ raise ValueError(
+ "evaluation_strategy and eval_steps mismatch. Please set evaluation_strategy to 'steps' or remove eval_steps."
+ )
+
+ if (
+ data.get("val_set_size") == 0
+ and (data.get("eval_steps") or data.get("evaluation_strategy"))
+ and not data.get("test_datasets")
+ ):
+ raise ValueError(
+ "eval_steps and evaluation_strategy are not supported with val_set_size == 0"
+ )
+ if data.get("evals_per_epoch") and data.get("eval_steps"):
+ raise ValueError(
+ "eval_steps and evals_per_epoch are mutually exclusive and cannot be used together."
+ )
+ if (
+ data.get("evals_per_epoch")
+ and data.get("evaluation_strategy")
+ and data.get("evaluation_strategy") != "steps"
+ ):
+ raise ValueError(
+ "evaluation_strategy must be empty or set to `steps` when used with evals_per_epoch."
+ )
+
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_eval_packing(cls, data):
+ if (
+ data.get("sample_packing")
+ and data.get("eval_table_size")
+ and data.get("eval_sample_packing") is not False
+ ):
+ raise ValueError(
+ "eval_table_size and eval_sample_packing are not supported together with sample_packing. Please set 'eval_sample_packing' to false."
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_warmup(cls, data):
+ if data.get("warmup_steps") and data.get("warmup_ratio"):
+ raise ValueError("warmup_steps and warmup_ratio are mutually exclusive")
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_neftune(cls, data):
+ if data.get("noisy_embedding_alpha") and not data.get("neftune_noise_alpha"):
+ data["neftune_noise_alpha"] = data["noisy_embedding_alpha"]
+ del data["noisy_embedding_alpha"]
+ elif data.get("noisy_embedding_alpha") and not data.get("neftune_noise_alpha"):
+ raise ValueError(
+ "noisy_embedding_alpha is deprecated, use neftune_noise_alpha; both are set, please remove the deprecated noisy_embedding_alpha setting"
+ )
+ return data
+
+ @field_validator("neftune_noise_alpha")
+ @classmethod
+ def validate_neftune_noise_alpha(cls, neftune_noise_alpha):
+ if neftune_noise_alpha is not None and neftune_noise_alpha <= 0.0:
+ raise ValueError("neftune_noise_alpha must be > 0.0")
+ return neftune_noise_alpha
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_frozen(cls, data):
+ if (
+ data.get("adapter")
+ and data.get("peft_layers_to_transform")
+ and data.get("unfrozen_parameters")
+ ):
+ raise ValueError(
+ "`unfrozen_parameters` used with `peft_layers_to_transform` can have unexpected behavior."
+ )
+
+ return data
+
+ @model_validator(mode="after")
+ def check_fft_possible_bad_config(self):
+ if (
+ # pylint: disable=too-many-boolean-expressions
+ not (self.bf16 or self.bfloat16)
+ and (self.fp16 or self.float16)
+ and not self.adapter
+ and not self.flash_attention
+ and self.sample_packing
+ ):
+ LOG.warning(
+ "Full fine tune w/o FA2 w/ sample packing and fp16/float16 is likely to raise errors. Try LoRA."
+ )
+ # ValueError: Attempting to unscale FP16 gradients.
+ # OR
+ # RuntimeError: expected mat1 and mat2 to have the same dtype, but got: float != c10::Half
+ return self
+
+ @model_validator(mode="after")
+ def check_fused_lora(self):
+ if self.adapter in ["lora", "qlora"] and (
+ self.flash_attn_fuse_qkv or self.flash_attn_fuse_mlp
+ ):
+ raise ValueError("Fused modules are not supported with LoRA/QLoRA")
+ return self
+
+ @model_validator(mode="after")
+ def hint_lora_8bit(self):
+ loftq = (
+ self.peft and self.peft.loftq_config and self.peft.loftq_config.loftq_bits
+ )
+ if not self.load_in_8bit and self.adapter == "lora" and not loftq:
+ LOG.warning("We recommend setting `load_in_8bit: true` for LORA finetuning")
+ return self
+
+ @model_validator(mode="after")
+ def check_early_stopping(self):
+ if self.early_stopping_patience:
+ if not self.save_steps or not self.eval_steps:
+ raise ValueError(
+ "`early_stopping_patience` requires save_steps and eval_steps to be set. eval_steps should evenly divide save_steps."
+ )
+ if self.save_steps % self.eval_steps != 0:
+ raise ValueError(
+ "`early_stopping_patience` requires that eval_steps should evenly divide save_steps."
+ )
+ return self
+
+ @model_validator(mode="after")
+ def check_relora(self):
+ if self.relora_steps:
+ if self.adapter not in ("lora", "qlora"):
+ raise ValueError("cfg.adapter must be lora or qlora to use ReLoRA")
+
+ if self.fsdp:
+ raise ValueError("fsdp not supported with ReLoRA")
+
+ if self.deepspeed:
+ raise ValueError("deepspeed not supported with ReLoRA")
+
+ if self.lr_scheduler == "one_cycle":
+ raise ValueError(
+ "ReLoRA is not compatible with the one_cycle scheduler"
+ )
+
+ if self.flash_attn_fuse_qkv or self.flash_attn_fuse_mlp:
+ raise ValueError("Fused modules are not supported with ReLoRA")
+ return self
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_mem_mismatch(cls, data):
+ if (
+ data.get("max_memory") is not None
+ and data.get("gpu_memory_limit") is not None
+ ):
+ raise ValueError(
+ "max_memory and gpu_memory_limit are mutually exclusive and cannot be used together."
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_use_reentrant_mismatch(cls, data):
+ if (
+ data.get("unfrozen_parameters")
+ and data.get("gradient_checkpointing_kwargs")
+ and data.get("gradient_checkpointing_kwargs", {}).get("use_reentrant")
+ is True
+ ):
+ # https://github.com/huggingface/transformers/issues/21381
+ raise ValueError(
+ "`use_reentrant` must be false when used with partially frozen model."
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_val_w_test_datasets(cls, data):
+ if data.get("test_datasets") and data.get("val_set_size"):
+ raise ValueError(
+ "non-zero val_set_size should not be used with test_datasets configuration"
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_fsdp_offload_w_8bit_optimizer(cls, data):
+ if (
+ data.get("fsdp")
+ and "8bit" in data.get("optimizer", "")
+ and data.get("fsdp_config")
+ and data["fsdp_config"].get("fsdp_offload_params")
+ ):
+ raise ValueError(
+ f"FSDP Offload not compatible with {data.get('optimizer')}"
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_causal_lm_evals(cls, data):
+ if data.get("do_causal_lm_eval") and data.get("eval_sample_packing"):
+ raise ValueError(
+ "do_causal_lm_eval is enabled, eval_sample_packing must be set to False"
+ )
+
+ if data.get("eval_causal_lm_metrics"):
+ supported_metrics = ["sacrebleu", "comet", "ter", "chrf"]
+ if not isinstance(data.get("eval_causal_lm_metrics"), list):
+ raise ValueError("eval_causal_lm_metrics must be a list")
+ # only ["sacrebleu", "comet", "ter", "chrf"] supported
+ if set(data.get("eval_causal_lm_metrics")) - set(supported_metrics):
+ raise ValueError(
+ f"eval_causal_lm_metrics must be one of {supported_metrics}"
+ )
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_dataset_or_pretraining_dataset(cls, data):
+ if data.get("datasets") is None and data.get("pretraining_dataset") is None:
+ raise ValueError("either datasets or pretraining_dataset is required")
+ return data
+
+
+class AxolotlConfigWCapabilities(AxolotlInputConfig):
+ """wrapper to valdiate gpu capabilities with the configured options"""
+
+ capabilities: GPUCapabilities
+
+ @model_validator(mode="after")
+ def check_bf16(self):
+ if self.capabilities.bf16:
+ if not self.bf16 and not self.bfloat16:
+ LOG.info(
+ "bf16 support detected, but not enabled for this configuration."
+ )
+ else:
+ if (
+ not self.merge_lora
+ and not self.is_preprocess
+ and (self.bf16 is True or self.bfloat16 is True)
+ ):
+ raise ValueError(
+ "bf16 requested, but AMP is not supported on this GPU. Requires Ampere series or above."
+ )
+ return self
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_sample_packing_w_sdpa_bf16(cls, data):
+ is_sm_90: bool = (
+ data["capabilities"]
+ and data["capabilities"].get("compute_capability") == "sm_90"
+ )
+ if (
+ data.get("sample_packing")
+ and data.get("sdp_attention")
+ and (data.get("bfloat16") or data.get("bf16"))
+ and not is_sm_90
+ ):
+ # https://github.com/pytorch/pytorch/blob/1b03423526536b5f3d35bdfa95ccc6197556cf9b/test/test_transformers.py#L2440-L2450
+ LOG.warning(
+ "sample_packing & torch sdpa with bf16 is unsupported may results in 0.0 loss. "
+ "This may work on H100s."
+ )
+
+ return data
+
+ @model_validator(mode="before")
+ @classmethod
+ def check_fsdp_deepspeed(cls, data):
+ if data.get("deepspeed") and data.get("fsdp"):
+ raise ValueError("deepspeed and fsdp cannot be used together.")
+ return data
diff --git a/src/axolotl/utils/config/models/input/v0_4_1/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/config/models/input/v0_4_1/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..c3d4f794a063b0a0d690efdf7fde030249c8ab13
Binary files /dev/null and b/src/axolotl/utils/config/models/input/v0_4_1/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/config/models/internals/__init__.py b/src/axolotl/utils/config/models/internals/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..dd742caf45c1a257fd52908e6e6e25aeaaea9c6a
--- /dev/null
+++ b/src/axolotl/utils/config/models/internals/__init__.py
@@ -0,0 +1,14 @@
+"""module for gpu capabilities"""
+from typing import Optional
+
+from pydantic import BaseModel, Field
+
+
+class GPUCapabilities(BaseModel):
+ """model to manage the gpu capabilities statically"""
+
+ bf16: bool = Field(default=False)
+ fp8: bool = Field(default=False)
+ n_gpu: int = Field(default=1)
+ n_node: int = Field(default=1)
+ compute_capability: Optional[str] = Field(default=None)
diff --git a/src/axolotl/utils/config/models/internals/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/config/models/internals/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..be666810346f11e1252a2f496d472c6974fc83e9
Binary files /dev/null and b/src/axolotl/utils/config/models/internals/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/data/__init__.py b/src/axolotl/utils/data/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..140d02106dac1c41d8d8490dad2dd709e700ca43
--- /dev/null
+++ b/src/axolotl/utils/data/__init__.py
@@ -0,0 +1,15 @@
+"""
+Data processing modules
+"""
+from axolotl.utils.data.pretraining import ( # noqa: F401
+ encode_pretraining,
+ wrap_pretraining_dataset,
+)
+from axolotl.utils.data.rl import load_prepare_dpo_datasets # noqa: F401
+from axolotl.utils.data.sft import ( # noqa: F401
+ get_dataset_wrapper,
+ load_prepare_datasets,
+ load_tokenized_prepared_datasets,
+ prepare_dataset,
+)
+from axolotl.utils.data.utils import md5 # noqa: F401
diff --git a/src/axolotl/utils/data/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/data/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..6eedb70ca28007df42725f8e2ec798b8a51a987a
Binary files /dev/null and b/src/axolotl/utils/data/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/data/__pycache__/pretraining.cpython-310.pyc b/src/axolotl/utils/data/__pycache__/pretraining.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..a48661418cc81fbab835b7d89c319ef427de862b
Binary files /dev/null and b/src/axolotl/utils/data/__pycache__/pretraining.cpython-310.pyc differ
diff --git a/src/axolotl/utils/data/__pycache__/rl.cpython-310.pyc b/src/axolotl/utils/data/__pycache__/rl.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..c5772e4e46830043d19bfb156bcf87fe1de2cac9
Binary files /dev/null and b/src/axolotl/utils/data/__pycache__/rl.cpython-310.pyc differ
diff --git a/src/axolotl/utils/data/__pycache__/sft.cpython-310.pyc b/src/axolotl/utils/data/__pycache__/sft.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..55314b0bc3fbfa5ff1a5521ea39ba4c3e0f9d5e9
Binary files /dev/null and b/src/axolotl/utils/data/__pycache__/sft.cpython-310.pyc differ
diff --git a/src/axolotl/utils/data/__pycache__/utils.cpython-310.pyc b/src/axolotl/utils/data/__pycache__/utils.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..9829242ab39c1361670fd74a1ed9d58ea1ae7bc1
Binary files /dev/null and b/src/axolotl/utils/data/__pycache__/utils.cpython-310.pyc differ
diff --git a/src/axolotl/utils/data/pretraining.py b/src/axolotl/utils/data/pretraining.py
new file mode 100644
index 0000000000000000000000000000000000000000..544ed131623f37c3cfd90bdfaa9058b357febd34
--- /dev/null
+++ b/src/axolotl/utils/data/pretraining.py
@@ -0,0 +1,232 @@
+"""data handling specific to pretraining"""
+
+import functools
+import logging
+from collections import defaultdict
+from typing import Callable, Dict, List, Optional
+
+import torch
+from datasets import Dataset
+from torch.utils.data import RandomSampler
+from transformers import PreTrainedTokenizerBase
+
+from axolotl.utils.collators import PretrainingBatchSamplerDataCollatorForSeq2Seq
+from axolotl.utils.samplers import MultipackBatchSampler, get_dataset_lengths
+from axolotl.utils.trainer import process_pretraining_datasets_for_packing
+
+LOG = logging.getLogger("axolotl")
+
+
+def encode_pretraining(
+ tokenizer: PreTrainedTokenizerBase, max_tokens: int, examples: List[str]
+) -> Dict[str, List]:
+ res = tokenizer(
+ examples,
+ truncation=True,
+ max_length=max_tokens - 2,
+ add_special_tokens=True,
+ )
+ # Convert to PyTorch tensors
+ input_ids = [torch.tensor(seq) for seq in res["input_ids"]]
+ attention_mask = [torch.tensor(seq) for seq in res["attention_mask"]]
+ new_input_ids = []
+ new_attention_mask = []
+ # Append EOS and PAD tokens to input_ids, and correct attention_mask
+ for i, _ in enumerate(input_ids):
+ input_ids[i] = torch.cat(
+ (
+ input_ids[i],
+ torch.tensor([tokenizer.eos_token_id, tokenizer.pad_token_id]),
+ ),
+ dim=0,
+ )
+ attention_mask[i] = torch.cat((attention_mask[i], torch.tensor([1, 0])), dim=0)
+
+ # Concatenate tokens so that their lengths are less than max_tokens
+ buffer_input_ids = torch.tensor([], dtype=torch.long)
+ buffer_attention_mask = torch.tensor([], dtype=torch.long)
+
+ for ids, mask in zip(input_ids, attention_mask):
+ if buffer_input_ids.numel() == max_tokens:
+ new_input_ids.append(buffer_input_ids)
+ new_attention_mask.append(buffer_attention_mask)
+ buffer_input_ids = torch.tensor([], dtype=torch.long)
+ buffer_attention_mask = torch.tensor([], dtype=torch.long)
+ buffer_input_ids = torch.cat((buffer_input_ids, ids), dim=0)
+ buffer_attention_mask = torch.cat((buffer_attention_mask, mask), dim=0)
+ elif buffer_input_ids.numel() + ids.numel() <= max_tokens:
+ buffer_input_ids = torch.cat((buffer_input_ids, ids), dim=0)
+ buffer_attention_mask = torch.cat((buffer_attention_mask, mask), dim=0)
+ else:
+ buffer_input_ids = torch.cat(
+ (
+ buffer_input_ids,
+ torch.full(
+ (max_tokens - buffer_input_ids.numel(),),
+ tokenizer.pad_token_id,
+ dtype=torch.long,
+ ),
+ ),
+ dim=0,
+ )
+ buffer_attention_mask = torch.cat(
+ (
+ buffer_attention_mask,
+ torch.full(
+ (max_tokens - buffer_attention_mask.numel(),),
+ 0,
+ dtype=torch.long,
+ ),
+ ),
+ dim=0,
+ )
+ new_input_ids.append(buffer_input_ids)
+ new_attention_mask.append(buffer_attention_mask)
+ buffer_input_ids = torch.tensor([], dtype=torch.long)
+ buffer_attention_mask = torch.tensor([], dtype=torch.long)
+
+ buffer_input_ids = torch.cat((buffer_input_ids, ids), dim=0)
+ buffer_attention_mask = torch.cat((buffer_attention_mask, mask), dim=0)
+
+ if buffer_input_ids.numel() > 0: # for any leftover tokens
+ while buffer_input_ids.numel() < max_tokens: # make all sequences equal in size
+ buffer_input_ids = torch.cat(
+ (
+ buffer_input_ids,
+ torch.full(
+ (max_tokens - buffer_input_ids.numel(),),
+ tokenizer.pad_token_id,
+ dtype=torch.long,
+ ),
+ ),
+ dim=0,
+ )
+ buffer_attention_mask = torch.cat(
+ (
+ buffer_attention_mask,
+ torch.full(
+ (max_tokens - buffer_attention_mask.numel(),),
+ 0,
+ dtype=torch.long,
+ ),
+ ),
+ dim=0,
+ )
+ new_input_ids.append(buffer_input_ids)
+ new_attention_mask.append(buffer_attention_mask)
+
+ ret = {
+ "input_ids": [seq.tolist() for seq in new_input_ids],
+ "labels": [seq.tolist() for seq in new_input_ids],
+ "attention_mask": [seq.tolist() for seq in new_attention_mask],
+ }
+
+ LOG.debug(len(ret["input_ids"]))
+ return ret
+
+
+def wrap_pretraining_dataset(
+ dataset,
+ tokenizer,
+ cfg,
+ ds_wrapper_fn,
+ max_tokens=2048,
+ batch_size=1,
+ seed=42,
+ buffer_size=10_000,
+):
+ if cfg.sample_packing:
+ collate_fn = PretrainingBatchSamplerDataCollatorForSeq2Seq(
+ tokenizer,
+ return_tensors="pt",
+ padding=True,
+ pad_to_multiple_of=max_tokens * batch_size,
+ multipack_attn=cfg.pretrain_multipack_attn,
+ )
+ encode = functools.partial(
+ encode_packed_pretraining,
+ collate_fn,
+ ds_wrapper_fn,
+ max_seq_length=max_tokens,
+ batch_size=batch_size,
+ multipack_attn=cfg.pretrain_multipack_attn,
+ )
+ # set this to 1 so downstream data_loader doesn't try to increase the batch again
+ cfg.micro_batch_size = 1
+ else:
+ encode = functools.partial(encode_pretraining, tokenizer, max_tokens)
+
+ if cfg.shuffle_merged_datasets:
+ dataset = dataset.shuffle(seed=seed, buffer_size=buffer_size)
+ else:
+ LOG.debug("NOT shuffling merged pretraining datasets")
+
+ # remove all the existing columns after mapping since they end up having
+ # a different length than the encoded/tokenized column
+ # this is empty during streaming/pretraining
+ remove_columns = []
+ if dataset.features is None:
+ for first_row in dataset:
+ remove_columns = first_row.keys()
+ break
+ else:
+ remove_columns = dataset.features.keys()
+
+ dataset = dataset.map(
+ encode,
+ batched=True,
+ batch_size=buffer_size,
+ # input_columns="text",
+ remove_columns=remove_columns,
+ )
+ return dataset
+
+
+def encode_packed_pretraining(
+ collate_fn,
+ ds_wrapper: Callable,
+ examples: Dict[str, List],
+ max_seq_length: int = 2048,
+ batch_size: int = 4,
+ multipack_attn: Optional[bool] = False,
+) -> Dict[str, List]:
+ # pylint: disable=duplicate-code
+ # tokenize all the examples
+ # rows get split with stride (overlap)
+ train_dataset = ds_wrapper(Dataset.from_dict(examples))[0]
+
+ train_dataset = process_pretraining_datasets_for_packing(
+ train_dataset,
+ max_seq_length,
+ skip_position_ids=not multipack_attn,
+ )
+
+ sampler = MultipackBatchSampler(
+ RandomSampler(train_dataset),
+ batch_size=1,
+ drop_last=True,
+ batch_max_len=batch_size * max_seq_length,
+ lengths=get_dataset_lengths(train_dataset),
+ )
+
+ chunked_data = defaultdict(list)
+
+ for batch in sampler:
+ for data in batch:
+ features = train_dataset[data]
+ if "num_truncated_tokens" in features:
+ del features["num_truncated_tokens"]
+ if "num_truncated_tokens" in features:
+ del features["num_truncated_tokens"]
+ if "overflow_to_sample_mapping" in features:
+ del features["overflow_to_sample_mapping"]
+ if "labels" not in features:
+ features["labels"] = features["input_ids"].copy()
+ collated_features = collate_fn(features)
+
+ for feature in features.keys():
+ if feature == "length":
+ continue
+ chunked_data[feature].append(collated_features[feature].squeeze(0))
+
+ return chunked_data
diff --git a/src/axolotl/utils/data/rl.py b/src/axolotl/utils/data/rl.py
new file mode 100644
index 0000000000000000000000000000000000000000..ff5ca87ddf4508769014531f4db92a27f1c4f896
--- /dev/null
+++ b/src/axolotl/utils/data/rl.py
@@ -0,0 +1,130 @@
+"""data handling specific to DPO"""
+import inspect
+import logging
+from functools import partial
+from pathlib import Path
+from typing import Any, List
+
+import yaml
+from datasets import DatasetDict, concatenate_datasets, load_dataset, load_from_disk
+
+from axolotl.common.const import DEFAULT_DATASET_PREPARED_PATH
+from axolotl.prompt_strategies.dpo import load as load_dpo
+from axolotl.prompt_strategies.orpo import load as load_orpo
+from axolotl.utils.data.utils import md5
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.distributed import is_main_process, zero_first
+from axolotl.utils.models import load_tokenizer
+
+LOG = logging.getLogger("axolotl")
+
+
+def _get_path(ds_hash, cfg):
+ prepared_ds_path = (
+ Path(cfg.dataset_prepared_path) / ds_hash
+ if cfg.dataset_prepared_path
+ else Path(DEFAULT_DATASET_PREPARED_PATH) / ds_hash
+ )
+
+ return prepared_ds_path
+
+
+def _load_preprocessed_ds(cfg, sub_cfg):
+ ds_hash = md5(yaml.dump(sub_cfg, Dumper=yaml.Dumper))
+ prepared_ds_path = _get_path(ds_hash, cfg)
+ dataset = None
+
+ # pylint: disable=duplicate-code
+ if (
+ cfg.dataset_prepared_path
+ and any(prepared_ds_path.glob("*"))
+ and not cfg.is_preprocess
+ ):
+ LOG.info(f"Loading prepared dataset from disk at {prepared_ds_path}...")
+ dataset = load_from_disk(str(prepared_ds_path))
+
+ return dataset
+
+
+def _save_preprocessed_ds(cfg, sub_cfg, dataset):
+ ds_hash = md5(yaml.dump(sub_cfg, Dumper=yaml.Dumper))
+ prepared_ds_path = _get_path(ds_hash, cfg)
+
+ if cfg.is_preprocess and is_main_process():
+ LOG.info(f"Loading prepared dataset from disk at {prepared_ds_path}...")
+ dataset.save_to_disk(str(prepared_ds_path))
+
+
+def load_prepare_dpo_datasets(cfg):
+ def load_split(dataset_cfgs, _cfg):
+ split_datasets: List[Any] = []
+ for i, ds_cfg in enumerate(dataset_cfgs):
+ if ds_cfg["ds_type"] == "json":
+ for data_file in ds_cfg["data_files"]:
+ data_files = {ds_cfg["split"]: data_file}
+ ds = load_dataset( # pylint: disable=invalid-name
+ "json",
+ data_files=data_files,
+ split=ds_cfg["split"],
+ )
+ split_datasets.insert(i, ds)
+ else:
+ ds = load_dataset( # pylint: disable=invalid-name
+ ds_cfg["path"],
+ split=ds_cfg["split"],
+ )
+ split_datasets.insert(i, ds)
+
+ tokenizer = None
+ for i, data_set in enumerate(split_datasets):
+ _type = dataset_cfgs[i]["type"]
+ if _type:
+ if isinstance(_type, DictDefault):
+ _type = "user_defined.default"
+ if _cfg.rl == "orpo":
+ ds_transform_fn = load_orpo(_type, _cfg, dataset_idx=i)
+ else:
+ ds_transform_fn = load_dpo(_type, _cfg, dataset_idx=i)
+ sig = inspect.signature(ds_transform_fn)
+ if "tokenizer" in sig.parameters:
+ if not tokenizer:
+ tokenizer = load_tokenizer(_cfg)
+ ds_transform_fn = partial(ds_transform_fn, tokenizer=tokenizer)
+
+ data_set = data_set.map(
+ ds_transform_fn,
+ desc="Mapping RL Dataset",
+ )
+ if isinstance(data_set, DatasetDict):
+ data_set = data_set["train"]
+ split_datasets[i] = data_set
+ else:
+ # If no `type` is provided, assume the dataset is already in the expected format with
+ # "prompt", "chosen" and "rejected" already preprocessed
+ split_datasets[i] = data_set
+
+ return concatenate_datasets(split_datasets)
+
+ with zero_first(is_main_process()):
+ train_is_preprocessed = False
+ eval_is_preprocessed = False
+ if train_dataset := _load_preprocessed_ds(cfg, cfg.datasets):
+ train_is_preprocessed = True
+ else:
+ train_dataset = load_split(cfg.datasets, cfg)
+
+ eval_dataset = None
+ if cfg.test_datasets:
+ if eval_dataset := _load_preprocessed_ds(cfg, cfg.test_datasets):
+ eval_is_preprocessed = True
+ else:
+ eval_dataset = load_split(cfg.test_datasets, cfg)
+ if not eval_dataset:
+ eval_dataset = None
+
+ if not train_is_preprocessed:
+ _save_preprocessed_ds(cfg, cfg.datasets, train_dataset)
+ if eval_dataset and not eval_is_preprocessed:
+ _save_preprocessed_ds(cfg, cfg.test_datasets, eval_dataset)
+
+ return train_dataset, eval_dataset
diff --git a/src/axolotl/utils/data/sft.py b/src/axolotl/utils/data/sft.py
new file mode 100644
index 0000000000000000000000000000000000000000..dbc4172b4b7a2d7dc49cd89712a61a3105246bf9
--- /dev/null
+++ b/src/axolotl/utils/data/sft.py
@@ -0,0 +1,681 @@
+"""data handling specific to SFT"""
+
+import functools
+import logging
+from pathlib import Path
+from typing import List, Optional, Tuple, Union
+
+from datasets import (
+ Dataset,
+ DatasetDict,
+ concatenate_datasets,
+ load_dataset,
+ load_from_disk,
+)
+from huggingface_hub import hf_hub_download
+from huggingface_hub.utils import HFValidationError
+from transformers import PreTrainedTokenizerBase
+
+from axolotl.common.const import DEFAULT_DATASET_PREPARED_PATH
+from axolotl.datasets import TokenizedPromptDataset
+from axolotl.prompt_strategies import load
+from axolotl.prompt_tokenizers import (
+ AlpacaMultipleChoicePromptTokenizingStrategy,
+ AlpacaPromptTokenizingStrategy,
+ AlpacaReflectionPTStrategy,
+ GPTeacherPromptTokenizingStrategy,
+ JeopardyPromptTokenizingStrategy,
+ OpenAssistantPromptTokenizingStrategy,
+ SummarizeTLDRPromptTokenizingStrategy,
+)
+from axolotl.prompters import (
+ AlpacaPrompter,
+ GPTeacherPrompter,
+ JeopardyPrompter,
+ MultipleChoiceConcisePrompter,
+ MultipleChoiceExplainPrompter,
+ Prompter,
+ ReflectAlpacaPrompter,
+ SummarizeTLDRPrompter,
+ UnsupportedPrompter,
+)
+from axolotl.utils.data.pretraining import wrap_pretraining_dataset
+from axolotl.utils.data.utils import md5
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.distributed import is_main_process, zero_first
+from axolotl.utils.trainer import (
+ calculate_total_num_steps,
+ process_datasets_for_packing,
+)
+
+LOG = logging.getLogger("axolotl")
+
+
+def prepare_dataset(cfg, tokenizer):
+ prompters = []
+ if not cfg.pretraining_dataset:
+ with zero_first(is_main_process()):
+ if cfg.test_datasets:
+ train_dataset, _, prompters = load_prepare_datasets(
+ tokenizer, cfg, DEFAULT_DATASET_PREPARED_PATH, split="train"
+ )
+ _, eval_dataset, _ = load_prepare_datasets(
+ tokenizer, cfg, DEFAULT_DATASET_PREPARED_PATH, split="test"
+ )
+ else:
+ train_dataset, eval_dataset, prompters = load_prepare_datasets(
+ tokenizer, cfg, DEFAULT_DATASET_PREPARED_PATH
+ )
+ else:
+ path = cfg.pretraining_dataset
+ split = "train"
+ name = None
+ if isinstance(cfg.pretraining_dataset, list) and isinstance(
+ cfg.pretraining_dataset[0], dict
+ ):
+ path = cfg.pretraining_dataset[0]["path"]
+ name = cfg.pretraining_dataset[0]["name"]
+ if "split" in cfg.pretraining_dataset[0]:
+ split = cfg.pretraining_dataset[0]["split"]
+
+ ds_wrapper_partial = functools.partial(
+ get_dataset_wrapper,
+ cfg.pretraining_dataset[0],
+ tokenizer,
+ cfg,
+ cfg.pretraining_dataset[0]["type"] or "pretrain",
+ )
+
+ train_dataset = wrap_pretraining_dataset(
+ load_dataset(path, streaming=True, split=split, name=name),
+ tokenizer,
+ cfg,
+ ds_wrapper_partial,
+ max_tokens=cfg.sequence_len,
+ batch_size=cfg.micro_batch_size,
+ seed=cfg.seed or 42,
+ buffer_size=cfg.pretrain_multipack_buffer_size or 10_000,
+ )
+ # https://discuss.huggingface.co/t/how-to-use-huggingface-trainer-streaming-datasets-without-wrapping-it-with-torchdatas-iterablewrapper/25230
+ train_dataset = train_dataset.with_format("torch")
+ eval_dataset = None
+ return train_dataset, eval_dataset, cfg.max_steps, prompters
+
+ if eval_dataset and cfg.sample_packing and cfg.eval_sample_packing is not False:
+ total_eval_steps = calculate_total_num_steps(cfg, eval_dataset, update=False)
+ if total_eval_steps == 0:
+ raise ValueError(
+ "eval dataset split is too small for sample_packing. You should set `eval_sample_packing: False`. "
+ )
+
+ if cfg.max_steps:
+ total_num_steps = min(
+ calculate_total_num_steps(cfg, train_dataset), cfg.max_steps
+ )
+ LOG.info(f"Maximum number of steps set at {total_num_steps}")
+ else:
+ total_num_steps = calculate_total_num_steps(cfg, train_dataset)
+ return train_dataset, eval_dataset, total_num_steps, prompters
+
+
+def load_tokenized_prepared_datasets(
+ tokenizer,
+ cfg,
+ default_dataset_prepared_path,
+ split="train",
+) -> Tuple[DatasetDict, List[Prompter]]:
+ cfg_datasets = cfg.test_datasets if split == "test" else cfg.datasets
+ tokenizer_name = cfg.tokenizer_config
+ ds_hash = str(
+ md5(
+ (
+ str(cfg.sequence_len)
+ + "@"
+ + str(cfg.sample_packing)
+ + "@"
+ + str(cfg.eval_sample_packing)
+ + "@"
+ + str(cfg.group_by_length)
+ + "@"
+ + "|".join(
+ sorted(
+ [
+ f"{d.path}:{d.type}:{d.shards}:{d.conversation}{d.split}"
+ for d in cfg_datasets
+ ]
+ )
+ )
+ + "|"
+ + tokenizer_name
+ )
+ )
+ )
+ prepared_ds_path = (
+ Path(cfg.dataset_prepared_path) / ds_hash
+ if cfg.dataset_prepared_path
+ else Path(default_dataset_prepared_path) / ds_hash
+ )
+ dataset = None
+ prompters = []
+ use_auth_token = cfg.hf_use_auth_token
+ try:
+ if cfg.push_dataset_to_hub:
+ dataset = load_dataset(
+ f"{cfg.push_dataset_to_hub}/{ds_hash}",
+ token=use_auth_token,
+ )
+ dataset = dataset[split]
+ except Exception: # pylint: disable=broad-except # nosec
+ pass
+
+ # pylint: disable=duplicate-code
+ if dataset:
+ ...
+ elif (
+ cfg.dataset_prepared_path
+ and any(prepared_ds_path.glob("*"))
+ and not cfg.is_preprocess
+ ):
+ LOG.info(f"Loading prepared dataset from disk at {prepared_ds_path}...")
+ dataset = load_from_disk(str(prepared_ds_path))
+ LOG.info("Prepared dataset loaded from disk...")
+ else:
+ LOG.info(f"Unable to find prepared dataset in {prepared_ds_path}")
+ LOG.info("Loading raw datasets...")
+ if not cfg.is_preprocess:
+ LOG.warning(
+ "Processing datasets during training can lead to VRAM instability. Please pre-process your dataset."
+ )
+
+ if cfg.seed:
+ seed = cfg.seed
+ else:
+ LOG.info("No seed provided, using default seed of 42")
+ seed = 42
+
+ datasets = []
+
+ def for_d_in_datasets(dataset_configs):
+ for dataset in dataset_configs:
+ if dataset.name and isinstance(dataset.name, list):
+ for name in dataset.name:
+ yield DictDefault({**dataset, "name": name})
+ else:
+ yield dataset
+
+ # pylint: disable=invalid-name
+ for config_dataset in for_d_in_datasets(cfg_datasets):
+ ds: Optional[Union[Dataset, DatasetDict]] = None
+ ds_from_hub = False
+ try:
+ load_dataset(
+ config_dataset.path,
+ name=config_dataset.name,
+ streaming=True,
+ token=use_auth_token,
+ )
+ ds_from_hub = True
+ except (FileNotFoundError, ConnectionError, HFValidationError, ValueError):
+ pass
+
+ ds_from_cloud = False
+ storage_options = {}
+ remote_file_system = None
+ if config_dataset.path.startswith("s3://"):
+ try:
+ import aiobotocore.session # type: ignore
+ import s3fs # type: ignore
+ except ImportError as exc:
+ raise ImportError(
+ "s3:// paths require aiobotocore and s3fs to be installed"
+ ) from exc
+
+ # Takes credentials from ~/.aws/credentials for default profile
+ s3_session = aiobotocore.session.AioSession(profile="default")
+ storage_options = {"session": s3_session}
+ remote_file_system = s3fs.S3FileSystem(**storage_options)
+ elif config_dataset.path.startswith(
+ "gs://"
+ ) or config_dataset.path.startswith("gcs://"):
+ try:
+ import gcsfs # type: ignore
+ except ImportError as exc:
+ raise ImportError(
+ "gs:// or gcs:// paths require gcsfs to be installed"
+ ) from exc
+
+ # gcsfs will use default credentials from the environment else anon
+ # https://gcsfs.readthedocs.io/en/latest/#credentials
+ storage_options = {"token": None}
+ remote_file_system = gcsfs.GCSFileSystem(**storage_options)
+ # TODO: Figure out how to get auth creds passed
+ # elif config_dataset.path.startswith("adl://") or config_dataset.path.startswith("abfs://"):
+ # try:
+ # import adlfs
+ # except ImportError as exc:
+ # raise ImportError(
+ # "adl:// or abfs:// paths require adlfs to be installed"
+ # ) from exc
+
+ # # Gen 1
+ # storage_options = {
+ # "tenant_id": TENANT_ID,
+ # "client_id": CLIENT_ID,
+ # "client_secret": CLIENT_SECRET,
+ # }
+ # # Gen 2
+ # storage_options = {
+ # "account_name": ACCOUNT_NAME,
+ # "account_key": ACCOUNT_KEY,
+ # }
+
+ # remote_file_system = adlfs.AzureBlobFileSystem(**storage_options)
+ try:
+ if remote_file_system and remote_file_system.exists(
+ config_dataset.path
+ ):
+ ds_from_cloud = True
+ except (FileNotFoundError, ConnectionError):
+ pass
+
+ # prefer local dataset, even if hub exists
+ local_path = Path(config_dataset.path)
+ if local_path.exists():
+ if local_path.is_dir():
+ if config_dataset.data_files:
+ ds_type = get_ds_type(config_dataset)
+ ds = load_dataset(
+ ds_type,
+ name=config_dataset.name,
+ data_files=config_dataset.data_files,
+ streaming=False,
+ split=None,
+ )
+ else:
+ ds = load_from_disk(config_dataset.path)
+ elif local_path.is_file():
+ ds_type = get_ds_type(config_dataset)
+
+ ds = load_dataset(
+ ds_type,
+ name=config_dataset.name,
+ data_files=config_dataset.path,
+ streaming=False,
+ split=None,
+ )
+ else:
+ raise ValueError(
+ "unhandled dataset load: local path exists, but is neither a directory or a file"
+ )
+ elif ds_from_hub:
+ ds = load_dataset(
+ config_dataset.path,
+ name=config_dataset.name,
+ streaming=False,
+ data_files=config_dataset.data_files,
+ token=use_auth_token,
+ )
+ elif ds_from_cloud and remote_file_system:
+ if remote_file_system.isdir(config_dataset.path):
+ ds = load_from_disk(
+ config_dataset.path,
+ storage_options=storage_options,
+ )
+ elif remote_file_system.isfile(config_dataset.path):
+ ds_type = get_ds_type(config_dataset)
+ ds = load_dataset(
+ ds_type,
+ name=config_dataset.name,
+ data_files=config_dataset.path,
+ streaming=False,
+ split=None,
+ storage_options=storage_options,
+ )
+ elif config_dataset.path.startswith("https://"):
+ ds_type = get_ds_type(config_dataset)
+ ds = load_dataset(
+ ds_type,
+ name=config_dataset.name,
+ data_files=config_dataset.path,
+ streaming=False,
+ split=None,
+ storage_options=storage_options,
+ )
+ else:
+ if isinstance(config_dataset.data_files, str):
+ fp = hf_hub_download(
+ repo_id=config_dataset.path,
+ repo_type="dataset",
+ filename=config_dataset.data_files,
+ )
+ elif isinstance(config_dataset.data_files, list):
+ fp = []
+ for file in config_dataset.data_files:
+ fp.append(
+ hf_hub_download(
+ repo_id=config_dataset.path,
+ repo_type="dataset",
+ filename=file,
+ )
+ )
+ else:
+ raise ValueError(
+ "data_files must be either a string or list of strings"
+ )
+ ds = load_dataset(
+ "json",
+ name=config_dataset.name,
+ data_files=fp,
+ streaming=False,
+ split=None,
+ )
+ if not ds:
+ raise ValueError("unhandled dataset load")
+
+ d_base_type = d_prompt_style = None
+ d_type = config_dataset.type
+ if isinstance(d_type, str):
+ d_type_split = d_type.split(":")
+ d_base_type = d_type_split[0]
+ d_prompt_style = d_type_split[1] if len(d_type_split) > 1 else None
+
+ if isinstance(ds, DatasetDict):
+ if config_dataset.split and config_dataset.split in ds:
+ ds = ds[config_dataset.split]
+ elif split in ds:
+ ds = ds[split]
+ else:
+ raise ValueError(
+ f"no {split} split found for dataset {config_dataset.path}, you may specify a split with 'split: `"
+ )
+
+ # support for using a subset of the data
+ if config_dataset.shards:
+ shards_idx = config_dataset.get("shards_idx", 0)
+ ds = ds.shuffle(seed=seed).shard(
+ num_shards=config_dataset.shards, index=shards_idx
+ )
+
+ dataset_wrapper, dataset_prompter = get_dataset_wrapper(
+ config_dataset=config_dataset,
+ tokenizer=tokenizer,
+ cfg=cfg,
+ dataset=ds,
+ d_base_type=d_base_type,
+ d_prompt_style=d_prompt_style,
+ )
+ datasets.append(dataset_wrapper)
+ prompters.append(dataset_prompter)
+
+ LOG.info("merging datasets")
+ dataset = concatenate_datasets(datasets)
+
+ if len(datasets) > 1:
+ if cfg.shuffle_merged_datasets:
+ LOG.debug("shuffle merged datasets")
+ dataset = dataset.shuffle(seed=seed)
+ else:
+ LOG.debug("NOT shuffling merged datasets")
+
+ dataset, _ = process_datasets_for_packing(cfg, dataset, None)
+
+ if cfg.local_rank == 0:
+ LOG.info(f"Saving merged prepared dataset to disk... {prepared_ds_path}")
+ dataset.save_to_disk(str(prepared_ds_path))
+ if cfg.push_dataset_to_hub:
+ LOG.info(
+ f"Saving merged prepared dataset with push_to_hub... {cfg.push_dataset_to_hub}/{ds_hash}"
+ )
+ dataset.push_to_hub(
+ f"{cfg.push_dataset_to_hub}/{ds_hash}", private=True
+ )
+
+ return dataset, prompters
+
+
+def get_ds_type(config_dataset: DictDefault):
+ """
+ Get the dataset type from the path if it's not specified
+ """
+ ds_type = "json"
+ if config_dataset.ds_type:
+ ds_type = config_dataset.ds_type
+ elif ".parquet" in config_dataset.path:
+ ds_type = "parquet"
+ elif ".arrow" in config_dataset.path:
+ ds_type = "arrow"
+ elif ".csv" in config_dataset.path:
+ ds_type = "csv"
+ elif ".txt" in config_dataset.path:
+ ds_type = "text"
+ return ds_type
+
+
+def load_prepare_datasets(
+ tokenizer: PreTrainedTokenizerBase,
+ cfg,
+ default_dataset_prepared_path,
+ split="train",
+) -> Tuple[Dataset, Dataset, List[Prompter]]:
+ dataset, prompters = load_tokenized_prepared_datasets(
+ tokenizer, cfg, default_dataset_prepared_path, split=split
+ )
+
+ if cfg.dataset_shard_num and cfg.dataset_shard_idx is not None:
+ LOG.info(
+ f"Using index #{cfg.dataset_shard_idx} of {cfg.dataset_shard_num} shards"
+ )
+ dataset = dataset.shard(
+ num_shards=cfg.dataset_shard_num,
+ index=cfg.dataset_shard_idx,
+ )
+
+ if split == "train" and cfg.val_set_size:
+ # ensure we end up with the same fingerprint by doing rank0 first and being able to cache
+ to_hash_train = (
+ dataset._fingerprint # pylint: disable=protected-access
+ + "|"
+ + str(cfg.val_set_size)
+ + "|"
+ + "train"
+ + "|"
+ + str(cfg.seed or 42)
+ )
+ to_hash_test = (
+ dataset._fingerprint # pylint: disable=protected-access
+ + "|"
+ + str(cfg.val_set_size)
+ + "|"
+ + "test"
+ + "|"
+ + str(cfg.seed or 42)
+ )
+ train_fingerprint = md5(to_hash_train)
+ test_fingerprint = md5(to_hash_test)
+
+ dataset = dataset.train_test_split(
+ test_size=cfg.val_set_size,
+ shuffle=False,
+ seed=cfg.seed or 42,
+ train_new_fingerprint=train_fingerprint,
+ test_new_fingerprint=test_fingerprint,
+ )
+
+ train_dataset = dataset["train"]
+ eval_dataset = dataset["test"]
+ elif split == "test":
+ train_dataset = None
+ eval_dataset = dataset
+ else:
+ train_dataset = dataset
+ eval_dataset = None
+
+ return train_dataset, eval_dataset, prompters
+
+
+def get_dataset_wrapper(
+ config_dataset,
+ tokenizer,
+ cfg,
+ d_base_type,
+ dataset,
+ d_prompt_style=None,
+):
+ dataset_wrapper = None
+ dataset_prompter = None
+
+ ds_kwargs = {
+ "process_count": cfg.dataset_processes,
+ "keep_in_memory": cfg.dataset_keep_in_memory is True,
+ }
+
+ if (
+ isinstance(dataset, Dataset)
+ and "input_ids" in dataset.features
+ and "attention_mask" in dataset.features
+ and "labels" in dataset.features
+ ):
+ # dataset is already tokenized, just drop it straight in
+ dataset_prompter = UnsupportedPrompter()
+ dataset_wrapper = dataset
+ elif isinstance(config_dataset.type, DictDefault):
+ ds_strategy = load(
+ "user_defined", tokenizer, cfg, config_dataset.type.to_dict()
+ )
+ dataset_prompter = UnsupportedPrompter()
+ dataset_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ elif ds_strategy := load(config_dataset.type, tokenizer, cfg, config_dataset):
+ dataset_prompter = UnsupportedPrompter()
+ dataset_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ elif d_base_type == "alpaca":
+ dataset_prompter = AlpacaPrompter(d_prompt_style)
+ ds_strategy = AlpacaPromptTokenizingStrategy(
+ dataset_prompter,
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ ds_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ dataset_wrapper = ds_wrapper
+ elif d_base_type == "explainchoice":
+ dataset_prompter = MultipleChoiceExplainPrompter(d_prompt_style)
+ ds_strategy = AlpacaMultipleChoicePromptTokenizingStrategy(
+ dataset_prompter,
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ ds_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ dataset_wrapper = ds_wrapper
+ elif d_base_type == "concisechoice":
+ dataset_prompter = MultipleChoiceConcisePrompter(d_prompt_style)
+ ds_strategy = AlpacaMultipleChoicePromptTokenizingStrategy(
+ dataset_prompter,
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ ds_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ dataset_wrapper = ds_wrapper
+ elif d_base_type == "summarizetldr":
+ dataset_prompter = SummarizeTLDRPrompter(d_prompt_style)
+ ds_strategy = SummarizeTLDRPromptTokenizingStrategy(
+ dataset_prompter,
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ ds_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ dataset_wrapper = ds_wrapper
+ elif d_base_type == "jeopardy":
+ dataset_prompter = JeopardyPrompter(d_prompt_style)
+ ds_strategy = JeopardyPromptTokenizingStrategy(
+ dataset_prompter,
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ ds_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ dataset_wrapper = ds_wrapper
+ elif d_base_type == "oasst":
+ dataset_prompter = AlpacaPrompter(d_prompt_style)
+ ds_strategy = OpenAssistantPromptTokenizingStrategy(
+ dataset_prompter,
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ ds_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ dataset_wrapper = ds_wrapper
+ elif d_base_type == "gpteacher":
+ dataset_prompter = GPTeacherPrompter(d_prompt_style)
+ ds_strategy = GPTeacherPromptTokenizingStrategy(
+ dataset_prompter,
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ ds_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ dataset_wrapper = ds_wrapper
+ elif d_base_type == "reflection":
+ dataset_prompter = ReflectAlpacaPrompter(d_prompt_style)
+ ds_strategy = AlpacaReflectionPTStrategy(
+ dataset_prompter,
+ tokenizer,
+ cfg.train_on_inputs,
+ cfg.sequence_len,
+ )
+ ds_wrapper = TokenizedPromptDataset(
+ ds_strategy,
+ dataset,
+ **ds_kwargs,
+ )
+ dataset_wrapper = ds_wrapper
+ else:
+ suffix = ""
+ if ":load_" in config_dataset.type:
+ suffix = f" Did you mean {config_dataset.type.replace(':load_', '.load_')}?"
+ LOG.error(
+ f"unhandled prompt tokenization strategy: {config_dataset.type}. {suffix}"
+ )
+ raise ValueError(
+ f"unhandled prompt tokenization strategy: {config_dataset.type} {suffix}"
+ )
+
+ return dataset_wrapper, dataset_prompter
diff --git a/src/axolotl/utils/data/utils.py b/src/axolotl/utils/data/utils.py
new file mode 100644
index 0000000000000000000000000000000000000000..e05701e7b0c99ba6bd2d56d0421c11c231c6d3c6
--- /dev/null
+++ b/src/axolotl/utils/data/utils.py
@@ -0,0 +1,10 @@
+"""data handling helpers"""
+
+import hashlib
+
+
+def md5(to_hash: str, encoding: str = "utf-8") -> str:
+ try:
+ return hashlib.md5(to_hash.encode(encoding), usedforsecurity=False).hexdigest()
+ except TypeError:
+ return hashlib.md5(to_hash.encode(encoding)).hexdigest() # nosec
diff --git a/src/axolotl/utils/dict.py b/src/axolotl/utils/dict.py
new file mode 100644
index 0000000000000000000000000000000000000000..409d088e6d2c6d63ce7d1b323c4c0fcf344ab143
--- /dev/null
+++ b/src/axolotl/utils/dict.py
@@ -0,0 +1,15 @@
+"""Module containing the DictDefault class"""
+
+from addict import Dict
+
+
+class DictDefault(Dict):
+ """
+ A Dict that returns None instead of returning empty Dict for missing keys.
+ """
+
+ def __missing__(self, key):
+ return None
+
+ def __or__(self, other):
+ return DictDefault(super().__ror__(other))
diff --git a/src/axolotl/utils/distributed.py b/src/axolotl/utils/distributed.py
new file mode 100644
index 0000000000000000000000000000000000000000..ecb1bcc9ec8701ed4332c3e7ffdb2fb6adb89057
--- /dev/null
+++ b/src/axolotl/utils/distributed.py
@@ -0,0 +1,228 @@
+"""
+utility helpers for distributed checks
+"""
+import os
+import pickle # nosec
+from contextlib import contextmanager
+from datetime import timedelta
+
+import torch
+import torch.distributed as dist
+from accelerate import PartialState
+
+distributed_state = None # pylint: disable=invalid-name
+
+
+def is_distributed():
+ """
+ Check if distributed training is initialized.
+ """
+ global distributed_state # pylint: disable=global-statement
+ if not distributed_state:
+ timeout = int(os.environ.get("AXOLOTL_NCCL_TIMEOUT", 1800))
+ distributed_state = PartialState(timeout=timedelta(seconds=timeout))
+
+ return distributed_state.use_distributed and distributed_state.initialized
+
+
+def barrier():
+ """
+ Acts as a barrier to wait for all processes. This ensures that all processes
+ reach the barrier before proceeding further.
+ """
+ if is_distributed():
+ dist.barrier()
+
+
+def is_main_process():
+ """
+ Check if the current process is the main process.
+ If not in distributed mode, always return True.
+ """
+ if not is_distributed():
+ return True
+ return dist.get_rank() == 0
+
+
+def get_world_size():
+ return int(os.getenv("WORLD_SIZE", "1"))
+
+
+@contextmanager
+def zero_only():
+ """
+ Context manager that only runs the enclosed block on the main rank.
+ """
+ if is_main_process():
+ yield
+ else:
+ yield None
+
+
+@contextmanager
+def zero_first(is_main):
+ """
+ runs the wrapped context so that rank 0 runs first before other ranks
+ """
+ if not is_main: # other ranks wait first
+ barrier()
+ yield
+ if is_main: # then rank 0 waits after it has run the context
+ barrier()
+
+
+def gather_scalar_from_all_ranks(fn, world_size=1): # pylint: disable=invalid-name
+ """
+ Run a callable 'fn' on all ranks and gather the results on the specified rank.
+
+ Args:
+ - fn (callable): A function that computes the value. This should not have any side effects.
+ - rank (int, optional): The rank that gathers the values. Default is 0.
+ - world_size (int, optional): Total number of processes in the current distributed setup.
+
+ Returns:
+ - A list of computed values from all ranks if on the gathering rank, otherwise None.
+ """
+ value_scalar = fn()
+ if not is_distributed():
+ return [value_scalar]
+ value_tensor = torch.tensor(
+ value_scalar, device=torch.cuda.current_device()
+ ).float()
+
+ if not is_main_process():
+ dist.gather(value_tensor, dst=0)
+ else:
+ gathered_tensors = [torch.zeros_like(value_tensor) for _ in range(world_size)]
+ dist.gather(value_tensor, gather_list=gathered_tensors, dst=0)
+
+ # Convert tensors back to their original type (int or float)
+ gathered_values = []
+ for tensor in gathered_tensors:
+ if tensor == tensor.int():
+ gathered_values.append(int(tensor.item()))
+ else:
+ gathered_values.append(float(tensor.item()))
+ return gathered_values
+ return None
+
+
+def broadcast_dict(vals: dict):
+ if not is_distributed():
+ return vals
+
+ if is_main_process():
+ data_byte = pickle.dumps(vals)
+ data_tensor = torch.ByteTensor(list(data_byte)).to("cuda")
+ data_size = torch.IntTensor([len(data_byte)]).to("cuda")
+ else:
+ data_tensor = torch.empty([1024], dtype=torch.uint8, device="cuda")
+ data_size = torch.IntTensor([0]).to("cuda")
+
+ dist.broadcast(data_size, 0)
+ if not is_main_process():
+ # resize
+ data_tensor = data_tensor.new_empty([data_size.item()])
+
+ dist.broadcast(data_tensor, 0)
+
+ if not is_main_process():
+ data_list = data_tensor.cpu().tolist()
+ data_byte = bytes(data_list[: data_size.item()])
+ vals = pickle.loads(data_byte) # nosec
+
+ return vals
+
+
+def compute_and_broadcast(fn): # pylint: disable=invalid-name
+ """
+ Compute a value using the function 'fn' only on the specified rank (default is 0).
+ The value is then broadcasted to all other ranks.
+
+ Args:
+ - fn (callable): A function that computes the value. This should not have any side effects.
+ - rank (int, optional): The rank that computes the value. Default is 0.
+
+ Returns:
+ - The computed value (int or float).
+ """
+ if is_main_process():
+ value_scalar = fn()
+ value_tensor = torch.tensor(
+ value_scalar, device=torch.cuda.current_device()
+ ).float()
+ else:
+ value_tensor = torch.tensor(
+ 0.0, device=torch.cuda.current_device()
+ ) # Placeholder tensor
+
+ # Broadcast the tensor to all processes.
+ barrier()
+ dist.broadcast(value_tensor, src=0)
+
+ # Convert the tensor back to its original type (int or float)
+ if value_tensor == value_tensor.int():
+ return int(value_tensor.item())
+ return float(value_tensor.item())
+
+
+def gather_from_all_ranks(fn, world_size=1): # pylint: disable=invalid-name
+ """
+ Run a callable 'fn' on all ranks and gather the results on the specified rank.
+
+ Args:
+ - fn (callable): A function that computes the value. This should not have any side effects.
+ - rank (int, optional): The rank that gathers the values. Default is 0.
+ - world_size (int, optional): Total number of processes in the current distributed setup.
+
+ Returns:
+ - A list of computed values from all ranks if on the gathering rank, otherwise None.
+ """
+ value_scalar = fn()
+ value_tensor = torch.tensor(
+ value_scalar, device=torch.cuda.current_device()
+ ).float()
+
+ # Placeholder tensor for gathering results
+ if is_main_process():
+ gathered_tensors = [torch.zeros_like(value_tensor) for _ in range(world_size)]
+ else:
+ gathered_tensors = None
+
+ dist.gather(value_tensor, gather_list=gathered_tensors, dst=0)
+
+ if is_main_process():
+ # Convert tensors back to their original type (int or float)
+ gathered_values = []
+ for tensor in gathered_tensors:
+ if tensor == tensor.int():
+ gathered_values.append(int(tensor.item()))
+ else:
+ gathered_values.append(float(tensor.item()))
+ return gathered_values
+ return None
+
+
+def reduce_and_broadcast(fn1, fn2):
+ """
+ Run a callable 'fn1' on all ranks, gather the results, reduce them using 'fn2',
+ and then broadcast the reduced result to all ranks.
+
+ Args:
+ - fn1 (callable): A function that computes the value on each rank.
+ - fn2 (callable): A reduction function that takes a list of values and returns a single value.
+ - world_size (int, optional): Total number of processes in the current distributed setup.
+
+ Returns:
+ - The reduced and broadcasted value.
+ """
+
+ # Gather values from all ranks using fn1
+ if not is_distributed():
+ return fn2([fn1()])
+
+ gathered_values = gather_from_all_ranks(fn1, world_size=dist.get_world_size())
+
+ # Use compute_and_broadcast to compute the reduced value on the main process
+ # and then broadcast it to all ranks
+ return compute_and_broadcast(lambda: fn2(gathered_values))
diff --git a/src/axolotl/utils/freeze.py b/src/axolotl/utils/freeze.py
new file mode 100644
index 0000000000000000000000000000000000000000..e3d0fd1446fe33246d3cf9b792d093e1b3e2c147
--- /dev/null
+++ b/src/axolotl/utils/freeze.py
@@ -0,0 +1,228 @@
+"""
+module to freeze/unfreeze parameters by name
+"""
+import logging
+import re
+from typing import Callable, List, Tuple, Union
+
+from axolotl.utils.distributed import is_main_process
+
+LOG = logging.getLogger("axolotl.utils.freeze")
+
+
+def freeze_layers_except(model, regex_patterns):
+ """
+ Freezes all layers of the given model except for the layers that match given regex patterns.
+ Periods in the patterns are treated as literal periods, not as wildcard characters.
+
+ Parameters:
+ - model (nn.Module): The PyTorch model to be modified.
+ - regex_patterns (list of str): List of regex patterns to match layer names to keep unfrozen.
+ Note that you cannot use a dot as a wildcard character in the patterns since it is reserved for separating layer names.
+ Also, to match the entire layer name, the pattern should start with "^" and end with "$", otherwise it will match any part of the layer name.
+ The range pattern part is optional and it is not compiled as a regex pattern which means you must put "$" before the range pattern if you want to match the entire layer name.
+ E.g., ["^model.embed_tokens.weight$[:32000]", "layers.2[0-9]+.block_sparse_moe.gate.[a-z]+$"]
+
+ Returns:
+ None; the model is modified in place.
+ """
+ if isinstance(regex_patterns, str):
+ regex_patterns = [regex_patterns]
+
+ patterns = [LayerNamePattern(pattern) for pattern in regex_patterns]
+
+ # Unfreeze layers that match the regex patterns
+ for name, param in model.named_parameters():
+ param.requires_grad = False
+ unfrozen_ranges = []
+ for pattern in patterns:
+ if not pattern.match(name):
+ continue
+
+ param.requires_grad = True
+
+ if pattern.range is not None:
+ unfrozen_ranges.append(pattern.range)
+
+ merged_unfrozen_ranges = _merge_ranges(unfrozen_ranges, len(param))
+
+ if param.requires_grad and is_main_process():
+ unfrozen_ranges = (
+ f" with ranges {merged_unfrozen_ranges}"
+ if merged_unfrozen_ranges
+ else ""
+ )
+ LOG.debug(f"Unfrozen {name}{unfrozen_ranges}")
+
+ if not merged_unfrozen_ranges:
+ continue
+
+ # The range list we need is actually the inverted of the merged ranges
+ ranges_to_freeze = _invert_ranges(merged_unfrozen_ranges, len(param))
+
+ param.register_hook(_create_freeze_parameters_hook(ranges_to_freeze))
+
+ if is_main_process() and all(
+ not param.requires_grad for param in model.parameters()
+ ):
+ LOG.warning("All parameters are frozen. Model will not be trained.")
+
+
+def _invert_ranges(
+ given_ranges: List[Tuple[int, int]], layer_size: int
+) -> List[Tuple[int, int]]:
+ """
+ Inverts a list of ranges to obtain the ranges not covered by the given ranges.
+
+ Parameters:
+ - given_ranges (List[Tuple[int, int]]): List of ranges to invert. Each range is represented as a tuple of start (inclusive) and end (exclusive) indices.
+ - layer_size (int): The length of the layer. E.g., len(model.layer.weight)
+ Returns:
+ - List[Tuple[int, int]]: List of inverted ranges, where each range is represented as a tuple of start (inclusive) and end (exclusive) indices.
+ """
+ if not given_ranges:
+ return [(0, layer_size)]
+
+ inverted_ranges = []
+ current_start = 0
+
+ for start, end in sorted(given_ranges):
+ if start > current_start:
+ inverted_ranges.append((current_start, start))
+ current_start = max(current_start, end)
+
+ # Handle the case where the last given range does not reach the end of the total_size
+ if current_start < layer_size:
+ inverted_ranges.append((current_start, layer_size))
+
+ return inverted_ranges
+
+
+def _merge_ranges(
+ given_ranges: List[Tuple[int, Union[int, None]]], layer_size: int
+) -> List[Tuple[int, int]]:
+ """
+ Merges overlapping ranges and sorts the given ranges.
+
+ This function takes a list of ranges and merges any overlapping ranges. The ranges are represented
+ as tuples, where the first element is the start index (inclusive) and the second element is the end
+ index (exclusive). The end index can be None, indicating that the range extends to the end of the
+ sequence.
+
+ Parameters:
+ - given_ranges (List[Tuple[int, int | None]]): List of ranges to merge.
+ - layer_size (int): The length of the layer. E.g., len(model.layer.weight)
+
+ Returns:
+ - List[Tuple[int, int]]: List of merged ranges, as start (inclusive) and end (exclusive) indices.
+ """
+ # End of each range can be determined now since we have the total size
+ processed_ranges = [
+ (start, end if end is not None else layer_size) for start, end in given_ranges
+ ]
+
+ # No need to merge if there's only one or no ranges
+ if len(processed_ranges) <= 1:
+ return processed_ranges
+
+ sorted_ranges = sorted(processed_ranges)
+
+ merged_ranges = [sorted_ranges[0]]
+ for start, end in sorted_ranges[1:]:
+ prev_start, prev_end = merged_ranges[-1]
+ if start <= prev_end:
+ merged_ranges[-1] = (prev_start, max(prev_end, end))
+ else:
+ merged_ranges.append((start, end))
+
+ return merged_ranges
+
+
+def _create_freeze_parameters_hook(ranges_to_freeze: List[Tuple[int, int]]) -> Callable:
+ """
+ Create a hook to freeze parameters in specified ranges by setting their gradients to zero.
+
+ This function takes a list of tuples representing the ranges of indices to freeze. Each tuple should contain
+ two integers representing the start and end indices of the range.
+
+ Parameters:
+ - ranges_to_freeze (List[Tuple[int, int]]): Ranges of indices to freeze.
+
+ Returns:
+ - Callable: A hook function to be used with `register_hook` on parameters.
+
+ Example usage:
+ ```
+ ranges_to_freeze = [(0, 10), (20, 30)]
+ hook = _create_freeze_parameters_hook(ranges_to_freeze)
+ model.register_hook(hook)
+ ```
+ """
+
+ def freeze_parameters_hook(gradients):
+ for start, end in ranges_to_freeze:
+ gradients[start:end].zero_()
+
+ return freeze_parameters_hook
+
+
+class LayerNamePattern:
+ """
+ Represents a regex pattern for layer names, potentially including a parameter index range.
+ """
+
+ def __init__(self, pattern: str):
+ """
+ Initializes a new instance of the LayerNamePattern class.
+
+ Parameters:
+ - pattern (str): The regex pattern for layer names, potentially including a parameter index range.
+ """
+ self.raw_pattern = pattern
+ name_pattern, self.range = self._parse_pattern(pattern)
+ self.name_regex = re.compile(name_pattern.replace(".", "\\."))
+
+ def match(self, name: str) -> bool:
+ """
+ Checks if the given layer name matches the regex pattern.
+
+ Parameters:
+ - name (str): The layer name to check.
+
+ Returns:
+ - bool: True if the layer name matches the pattern, False otherwise.
+ """
+ return self.name_regex.match(name) is not None
+
+ def _parse_pattern(
+ self, pattern: str
+ ) -> Tuple[str, Union[Tuple[int, Union[int, None]], None]]:
+ """
+ Extracts the range pattern from the given pattern.
+
+ Parameters:
+ - pattern (str): The pattern to extract the range from.
+
+ Returns:
+ - Tuple[str, Tuple[int, int | None] | None]: A tuple containing the regex pattern to match the layer name without the range pattern and the range of layer indices to match, if specified.
+ """
+ match = re.match(r"^(.+)\[([0-9]*)(?::([0-9]*))?\]$", pattern)
+ if not match:
+ return pattern, None
+
+ base_pattern, start_part, end_part = match.groups()
+
+ if end_part is None and start_part.isdecimal():
+ index = int(start_part)
+ return base_pattern, (index, index + 1)
+
+ # [:end] or [start:] or [start:end]
+ start = int(start_part) if start_part else 0
+ end = int(end_part) if end_part else None
+
+ if end is not None and start >= end:
+ raise ValueError(
+ f"Invalid range in layer name pattern: {pattern}."
+ "End of range must be greater than start."
+ )
+ return base_pattern, (start, end)
diff --git a/src/axolotl/utils/gradient_checkpointing/__init__.py b/src/axolotl/utils/gradient_checkpointing/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..4639fc266cb56f874fc2c6e86bd0e8bb0395ec3e
--- /dev/null
+++ b/src/axolotl/utils/gradient_checkpointing/__init__.py
@@ -0,0 +1,13 @@
+"""custom checkpointing utils"""
+from axolotl.utils.gradient_checkpointing.unsloth import (
+ Unsloth_Offloaded_Gradient_Checkpointer,
+)
+
+
+def hf_grad_checkpoint_unsloth_wrapper(
+ decoder_layer, *args, use_reentrant=None
+): # pylint: disable=unused-argument
+ return Unsloth_Offloaded_Gradient_Checkpointer.apply(
+ decoder_layer.__self__,
+ *args,
+ )
diff --git a/src/axolotl/utils/gradient_checkpointing/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/gradient_checkpointing/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..f52344bf19b3ade4556d91265ee5e81742f49f1a
Binary files /dev/null and b/src/axolotl/utils/gradient_checkpointing/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/gradient_checkpointing/__pycache__/unsloth.cpython-310.pyc b/src/axolotl/utils/gradient_checkpointing/__pycache__/unsloth.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..82f8371ec6b1ec1b418420e955ab4aade6e1db2a
Binary files /dev/null and b/src/axolotl/utils/gradient_checkpointing/__pycache__/unsloth.cpython-310.pyc differ
diff --git a/src/axolotl/utils/gradient_checkpointing/unsloth.py b/src/axolotl/utils/gradient_checkpointing/unsloth.py
new file mode 100644
index 0000000000000000000000000000000000000000..fbe8346be274d5b4c945e54d332471aa8843bd8d
--- /dev/null
+++ b/src/axolotl/utils/gradient_checkpointing/unsloth.py
@@ -0,0 +1,52 @@
+"""Unsloth checkpointing"""
+
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+import torch
+
+
+class Unsloth_Offloaded_Gradient_Checkpointer( # pylint: disable=invalid-name
+ torch.autograd.Function
+):
+ """
+ Saves VRAM by smartly offloading to RAM.
+ Tiny hit to performance, since we mask the movement via non blocking calls.
+ """
+
+ @staticmethod
+ @torch.cuda.amp.custom_fwd
+ def forward(ctx, forward_function, hidden_states, *args):
+ saved_hidden_states = hidden_states.to("cpu", non_blocking=True)
+ with torch.no_grad():
+ output = forward_function(hidden_states, *args)
+ ctx.save_for_backward(saved_hidden_states)
+ ctx.forward_function = forward_function
+ ctx.args = args
+ return output
+
+ @staticmethod
+ @torch.cuda.amp.custom_bwd
+ def backward(ctx, dY):
+ (hidden_states,) = ctx.saved_tensors
+ hidden_states = hidden_states.to("cuda", non_blocking=True).detach()
+ hidden_states.requires_grad = True
+ with torch.enable_grad():
+ (output,) = ctx.forward_function(hidden_states, *ctx.args)
+ torch.autograd.backward(output, dY)
+ return (
+ None,
+ hidden_states.grad,
+ ) + (
+ None,
+ ) * len(ctx.args)
diff --git a/src/axolotl/utils/lora_embeddings.py b/src/axolotl/utils/lora_embeddings.py
new file mode 100644
index 0000000000000000000000000000000000000000..70f56655ea14cddcecc2dd0d9f781077243ccabe
--- /dev/null
+++ b/src/axolotl/utils/lora_embeddings.py
@@ -0,0 +1,14 @@
+"""
+helpers for lora embeddings
+"""
+
+
+def get_linear_embedding_layers(model_type):
+ """
+ returns the linear embedding layers needed for loras, dependent on the model arch
+ """
+ if model_type == "gpt_neox":
+ return ["embed_in", "embed_out"]
+ if model_type == "falcon":
+ return ["word_embeddings", "lm_head"]
+ return ["embed_tokens", "lm_head"]
diff --git a/src/axolotl/utils/mlflow_.py b/src/axolotl/utils/mlflow_.py
new file mode 100644
index 0000000000000000000000000000000000000000..ce77390342788a9df9ccd84e4f52bd71abc949f0
--- /dev/null
+++ b/src/axolotl/utils/mlflow_.py
@@ -0,0 +1,18 @@
+"""Module for mlflow utilities"""
+
+import os
+
+from axolotl.utils.dict import DictDefault
+
+
+def setup_mlflow_env_vars(cfg: DictDefault):
+ for key in cfg.keys():
+ if key.startswith("mlflow_") or key.startswith("hf_mlflow_"):
+ value = cfg.get(key, "")
+
+ if value and isinstance(value, str) and len(value) > 0:
+ os.environ[key.upper()] = value
+
+ # Enable mlflow if experiment name is present
+ if cfg.mlflow_experiment_name and len(cfg.mlflow_experiment_name) > 0:
+ cfg.use_mlflow = True
diff --git a/src/axolotl/utils/model_shard_quant.py b/src/axolotl/utils/model_shard_quant.py
new file mode 100644
index 0000000000000000000000000000000000000000..65f23b9e0f6222e0c02283d61fc154891aa9c742
--- /dev/null
+++ b/src/axolotl/utils/model_shard_quant.py
@@ -0,0 +1,259 @@
+"""
+module to handle loading model on cpu/meta device for FSDP
+"""
+import os
+import time
+from typing import List, Optional, Type, Union
+
+import safetensors
+import torch
+from accelerate import init_empty_weights
+from bitsandbytes.nn import Linear4bit, Params4bit
+from fastcore.parallel import parallel
+from torch import Tensor, nn
+from tqdm import tqdm
+from transformers import AutoModelForCausalLM
+from transformers.utils import SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, hub
+
+
+def _replace_linear(
+ model: nn.Module,
+ linear_replacement: Type[nn.Module],
+ quant_config: Union[dict, None] = None,
+ skip_modules=None,
+ **kwargs,
+):
+ """
+ Replace linear modules with a new Linear module.
+ Parameters:
+ model (`torch.nn.Module`):
+ Input model or `torch.nn.Module` as the function is run recursively.
+ linear_replacement (`torch.nn.Module`):
+ The linear module that replaces the old one. Only expects standard arguments.
+ If other arguments need to be passed, use a lambda.
+ skip_modules (`List[str]`, *optional*, defaults to `lm_head`):
+ List of modules names not to convert. Defaults to `lm_head`.
+ """
+ if skip_modules is None:
+ skip_modules = ["lm_head"]
+ for name, module in model.named_children():
+ if len(list(module.children())) > 0:
+ _replace_linear(
+ module, linear_replacement, quant_config, skip_modules, **kwargs
+ )
+
+ if isinstance(module, torch.nn.Linear) and name not in skip_modules:
+ if issubclass(linear_replacement, Linear4bit):
+ model._modules[ # pylint: disable=protected-access
+ name
+ ] = linear_replacement(
+ module.in_features,
+ module.out_features,
+ module.bias is not None,
+ **kwargs,
+ )
+ else:
+ raise ValueError(
+ f"Unsupported linear replacement: {type(linear_replacement)}"
+ )
+ return model
+
+
+def load_and_quantize(
+ module: nn.Module,
+ name: str,
+ value: Tensor,
+ device: torch.device = None,
+ dtype: torch.dtype = None,
+ skip_names: Optional[List[str]] = None,
+ to_cpu: bool = False,
+ to_meta: bool = False,
+ verbose: bool = False,
+ quant_method: str = "bnb",
+):
+ """
+ Loads `value` tensor into submodule of `module`, optionally skipping `skip_names` and converting to `dtype`.
+
+ Quantizes `Params4bit` on `device` then places on "cpu" if to_cpu=True or "meta" if to_meta=True.
+ """
+
+ if not skip_names:
+ skip_names = []
+
+ def place_on_device(value):
+ if to_meta:
+ device = "meta"
+ elif to_cpu:
+ device = "cpu"
+ return value.to(device=device, dtype=dtype)
+
+ if any(skip_name in name for skip_name in skip_names):
+ if verbose:
+ print(f"Skipping {name} because it is in skip_names")
+ return
+
+ module_key, _, value_key = name.rpartition(".")
+ try:
+ submodule = module.get_submodule(module_key)
+ except AttributeError as exc:
+ print(f"Module {module_key} not found:\n{exc}")
+ return
+
+ try:
+ if quant_method == "bnb":
+ param = submodule.get_parameter(value_key)
+ if isinstance(param, Params4bit):
+ # With `sync_module_states=True`, a meta device Params4bit needs to be the same
+ # shape as the quantized Params4bit with an initialized quant_state. However,
+ # FSDP only syncs parameters and buffers, so the quant_state isn't copied. This
+ # workaround quantizes Params4bit to initialize quant_state on all ranks, then
+ # replaces Params4bit's data with a meta tensor to free memory on non-rank 0.
+ value = type(param)(
+ value.to(device=device, dtype=dtype).data, **param.__dict__
+ ).cuda(device)
+ if to_meta:
+ value = type(param)(value.data.to("meta"), **value.__dict__)
+ elif to_cpu:
+ value = type(param)(value.data.to("cpu"), **value.__dict__)
+ else:
+ value = type(param)(place_on_device(value).data)
+
+ except AttributeError:
+ # it's a buffer
+ value = place_on_device(value)
+
+ setattr(submodule, value_key, value)
+
+
+def n_loading_workers(quant_method: str, param_count: float):
+ devprops = torch.cuda.get_device_properties(torch.cuda.current_device())
+ left = int(os.cpu_count() / torch.cuda.device_count())
+ model_params_b = 70
+ right = int(
+ (4 if quant_method == "hqq" else 8)
+ * (devprops.total_memory / 1e9 / 40)
+ * (model_params_b / (param_count / 1e9))
+ )
+ return min(left, right)
+
+
+def load_sharded_model(
+ model_name,
+ model_config,
+ cfg,
+ torch_dtype=torch.bfloat16,
+ low_memory=True,
+):
+ if (low_memory and cfg.local_rank == 0) or not low_memory:
+ model = AutoModelForCausalLM.from_pretrained(
+ model_name,
+ use_cache=False,
+ torch_dtype=torch.float32,
+ _attn_implementation=model_config._attn_implementation, # pylint: disable=protected-access
+ trust_remote_code=cfg.trust_remote_code,
+ )
+ dtype = torch_dtype if not cfg.float32 else None
+ model.to(dtype=dtype, device="cpu" if low_memory else cfg.local_rank)
+ else:
+ with init_empty_weights():
+ model = AutoModelForCausalLM.from_config(
+ model_config,
+ torch_dtype=torch_dtype,
+ trust_remote_code=cfg.trust_remote_code,
+ )
+ return model
+
+
+def load_sharded_model_quant(
+ model_name,
+ model_config,
+ cfg,
+ compute_dtype=torch.bfloat16,
+ quant_storage=torch.float32,
+ low_memory=True,
+ verbose=False,
+ loading_workers=2,
+):
+ with init_empty_weights():
+ model = AutoModelForCausalLM.from_config(
+ model_config,
+ trust_remote_code=cfg.trust_remote_code,
+ )
+ if hasattr(model, "transformer"):
+ model.transformer = _replace_linear(
+ model.transformer,
+ Linear4bit,
+ compute_dtype=compute_dtype,
+ quant_type="nf4",
+ quant_storage=quant_storage,
+ )
+ else:
+ # this is the more common case with HF transformers
+ model.model = _replace_linear(
+ model.model,
+ Linear4bit,
+ compute_dtype=compute_dtype,
+ quant_type="nf4",
+ quant_storage=quant_storage,
+ )
+ model.is_loaded_in_4bit = True
+
+ # Grab the safetensors files that hold the weights
+ try:
+ idx = hub.cached_file(model_name, SAFE_WEIGHTS_INDEX_NAME)
+ files, _ = hub.get_checkpoint_shard_files(model_name, idx)
+ except OSError:
+ try:
+ # This means the model doesn't have a model.safetensors.index.json because it is not sharded
+ files = []
+ files.append(hub.cached_file(model_name, SAFE_WEIGHTS_NAME))
+ except OSError as exc:
+ # This means the model probably doesn't have a safetensors file
+ raise exc
+
+ # Load in the weights, using our custom load_and_quantize method which quantizes Params4bit on the fly
+ # and then places each layer on CPU or meta if using low_memory to minimize GPU memory usage
+ def load_and_quantize_parallel(name_param, model, **kwargs):
+ name, param = name_param
+ load_and_quantize(model, name, param, **kwargs)
+
+ quant_method = "bnb"
+ param_count = sum((p.numel() for n, p in model.named_parameters()))
+
+ n_workers = (
+ n_loading_workers(quant_method, param_count)
+ if loading_workers == -1
+ else loading_workers
+ )
+ if cfg.local_rank == 0 and verbose:
+ print(f"Using n_workers: {n_workers} for loading")
+
+ start = time.time()
+ for filename in tqdm(
+ files,
+ desc="Loading & Quantizing Model Shards",
+ disable=cfg.local_rank != 0,
+ position=0,
+ ):
+ weights = safetensors.torch.load_file(filename)
+ parallel(
+ load_and_quantize_parallel,
+ iter(weights.items()),
+ n_workers=n_workers,
+ threadpool=True,
+ model=model,
+ dtype=quant_storage,
+ device=cfg.local_rank,
+ skip_names=[],
+ to_cpu=(low_memory and cfg.local_rank == 0),
+ to_meta=(low_memory and cfg.local_rank != 0),
+ verbose=verbose,
+ quant_method=quant_method,
+ )
+
+ if cfg.local_rank == 0 and verbose:
+ print(f"Loaded model weights in {time.time()-start:.3f} seconds")
+ # cleanup any extra memory usage from parallel loading
+ torch.cuda.empty_cache()
+
+ return model
diff --git a/src/axolotl/utils/models.py b/src/axolotl/utils/models.py
new file mode 100644
index 0000000000000000000000000000000000000000..c4ca6730eccc39151386e39ed77bb77c5ef56a91
--- /dev/null
+++ b/src/axolotl/utils/models.py
@@ -0,0 +1,1028 @@
+"""Module for models and model loading"""
+
+# pylint: disable=too-many-lines
+
+import logging
+import math
+import os
+import types
+from typing import Any, Dict, Optional, Tuple, Union # noqa: F401
+
+import addict
+import bitsandbytes as bnb
+import torch
+import transformers
+import transformers.modeling_utils
+from accelerate import init_empty_weights
+from bitsandbytes.nn import Params4bit
+from peft import (
+ LoftQConfig,
+ PeftConfig,
+ PeftModel,
+ PeftModelForCausalLM,
+ prepare_model_for_kbit_training,
+)
+from peft.tuners.lora import QuantLinear
+from torch import nn
+from transformers import ( # noqa: F401
+ AddedToken,
+ AutoConfig,
+ AutoModelForCausalLM,
+ AutoTokenizer,
+ BitsAndBytesConfig,
+ GPTQConfig,
+ PreTrainedModel,
+ PreTrainedTokenizerBase,
+)
+from transformers.integrations.deepspeed import is_deepspeed_zero3_enabled
+
+from axolotl.models.mamba import fix_mamba_attn_for_loss
+from axolotl.monkeypatch.multipack import (
+ SUPPORTED_MULTIPACK_MODEL_TYPES,
+ patch_for_multipack,
+)
+from axolotl.prompt_tokenizers import LLAMA_DEFAULT_EOS_TOKEN
+from axolotl.utils.bench import log_gpu_memory_usage
+from axolotl.utils.chat_templates import chat_templates
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.distributed import zero_only
+from axolotl.utils.gradient_checkpointing import hf_grad_checkpoint_unsloth_wrapper
+from axolotl.utils.lora_embeddings import get_linear_embedding_layers
+from axolotl.utils.model_shard_quant import load_sharded_model, load_sharded_model_quant
+
+LOG = logging.getLogger("axolotl")
+
+
+# copied from accelerator.FullyShardedDataParallelPlugin
+def get_module_class_from_name(module, name):
+ """
+ Gets a class from a module by its name.
+
+ Args:
+ module (`torch.nn.Module`): The module to get the class from.
+ name (`str`): The name of the class.
+ """
+ modules_children = list(module.children())
+ if module.__class__.__name__ == name:
+ return module.__class__
+
+ if len(modules_children) == 0:
+ return None
+
+ for child_module in modules_children:
+ module_class = get_module_class_from_name(child_module, name)
+ if module_class is not None:
+ return module_class
+
+ return None
+
+
+def check_model_config(cfg: DictDefault, model_config: Union[AutoConfig, DictDefault]):
+ quant_config_exists = (
+ hasattr(model_config, "quantization_config")
+ and model_config.quantization_config
+ )
+ quant_config_method_is_gptq = (
+ quant_config_exists
+ and "quant_method" in model_config.quantization_config
+ and model_config.quantization_config["quant_method"] == "gptq"
+ )
+
+ if cfg.gptq and not quant_config_method_is_gptq:
+ raise ValueError(
+ "model_config.quantization_config is not set or quant_method is not set to gptq. "
+ "Please make sure to point to a GPTQ model."
+ )
+
+ if not cfg.gptq and quant_config_exists:
+ raise ValueError(
+ "model_config.quantization_config is set but `gptq` flag is not. "
+ "Please use the `gptq` flag to train quantized model or point to a non-quantized model."
+ )
+
+ lora_modules_to_save = get_linear_embedding_layers(model_config.model_type)
+ if (
+ cfg.adapter
+ and cfg.tokens
+ and (
+ not cfg.lora_modules_to_save
+ or not all(x in cfg.lora_modules_to_save for x in lora_modules_to_save)
+ )
+ ):
+ lora_modules_to_save = ", ".join(map(lambda x: f"`{x}`", lora_modules_to_save))
+ raise ValueError(
+ f"`lora_modules_to_save` not properly set when adding new tokens. Please include [{lora_modules_to_save}] in `lora_modules_to_save`."
+ )
+
+
+def load_model_config(cfg):
+ model_config_name = cfg.base_model_config or cfg.base_model
+ if not model_config_name and cfg.tokenizer_config:
+ model_config_name = cfg.tokenizer_config
+ trust_remote_code = cfg.trust_remote_code is True
+ config_kwargs = {}
+ if cfg.revision_of_model:
+ config_kwargs["revision"] = cfg.revision_of_model
+
+ try:
+ model_config = AutoConfig.from_pretrained(
+ model_config_name,
+ trust_remote_code=trust_remote_code,
+ **config_kwargs,
+ )
+ except ValueError as err:
+ if "mamba" in model_config_name:
+ return addict.Dict(
+ {
+ "model_type": "mamba",
+ }
+ )
+ raise err
+
+ if cfg.overrides_of_model_config:
+ for key, val in cfg.overrides_of_model_config.items():
+ setattr(model_config, key, val)
+
+ check_model_config(cfg, model_config)
+
+ return model_config
+
+
+def load_tokenizer(cfg):
+ model_config = load_model_config(cfg)
+ tokenizer_kwargs = {}
+ use_fast = True # this is the default
+
+ if cfg.tokenizer_use_fast is not None:
+ use_fast = cfg.tokenizer_use_fast
+ if cfg.tokenizer_legacy is not None:
+ # True is the default w/ https://github.com/huggingface/transformers/pull/25224
+ tokenizer_kwargs["legacy"] = cfg.tokenizer_legacy
+
+ tokenizer_cls = AutoTokenizer
+ if cfg.tokenizer_type:
+ tokenizer_cls = getattr(transformers, cfg.tokenizer_type)
+
+ tokenizer = tokenizer_cls.from_pretrained(
+ cfg.tokenizer_config,
+ trust_remote_code=cfg.trust_remote_code or False,
+ use_fast=use_fast,
+ **tokenizer_kwargs,
+ )
+
+ if (
+ tokenizer.__class__.__name__
+ in [
+ "LlamaTokenizer",
+ "LlamaTokenizerFast",
+ "CodeLlamaTokenizer",
+ "CodeLlamaTokenizerFast",
+ ]
+ and hasattr(tokenizer, "pad_token")
+ and not tokenizer.pad_token
+ ):
+ # set a pad_token, but use eos_token so we don't add a new token
+ tokenizer.pad_token = LLAMA_DEFAULT_EOS_TOKEN
+
+ if tokenizer.__class__.__name__ == "GPTNeoXTokenizerFast":
+ tokenizer.add_special_tokens({"pad_token": "[PAD]"})
+ os.environ["TOKENIZERS_PARALLELISM"] = "false"
+
+ # Mistral's official FA implementation requires left padding
+ if cfg.is_mistral_derived_model and cfg.flash_attention and not cfg.sample_packing:
+ tokenizer.padding_side = "left"
+
+ # Qwen base only has single token, so we need to set the special tokens
+ if cfg.is_qwen_derived_model:
+ token_ids = ["bos_token_id", "eos_token_id", "pad_token_id", "unk_token_id"]
+ for attr_name in token_ids:
+ if getattr(tokenizer, attr_name) is None:
+ setattr(tokenizer, attr_name, tokenizer.eod_id)
+
+ token_names = ["bos_token", "eos_token", "pad_token", "unk_token"]
+ for attr_name in token_names:
+ if getattr(tokenizer, attr_name) is None:
+ setattr(tokenizer, attr_name, "<|endoftext|>")
+
+ additional_special_tokens = None
+ if cfg.special_tokens:
+ special_tokens = cfg.special_tokens.to_dict()
+ additional_special_tokens = special_tokens.pop(
+ "additional_special_tokens", None
+ )
+ lora_modules_to_save = get_linear_embedding_layers(model_config.model_type)
+ for k, val in special_tokens.items():
+ # check if new special token is not already in tokenizer and
+ # is adapter training to make sure lora_modules_to_save is set
+ # pylint: disable=too-many-boolean-expressions
+ if (
+ (getattr(tokenizer, k) is None or getattr(tokenizer, k) != val)
+ and (len(tokenizer.encode(val, add_special_tokens=False)) > 2)
+ and cfg.adapter
+ and (
+ not cfg.lora_modules_to_save
+ or not all(
+ x in cfg.lora_modules_to_save for x in lora_modules_to_save
+ )
+ )
+ ):
+ lora_modules_to_save = ", ".join(
+ [f"`{x}`" for x in lora_modules_to_save]
+ )
+ raise ValueError(
+ f"Please set lora_modules_to_save to [{lora_modules_to_save}] when using an adapter and changing the special tokens."
+ )
+
+ tokenizer.add_special_tokens(
+ {k: AddedToken(val, rstrip=False, lstrip=False, normalized=False)}
+ )
+
+ # If we add bos_token and eos_token, we need to update the post processor to
+ # handle them correctly.
+ # https://github.com/huggingface/transformers/pull/24132
+ bos_or_eos_in_special_tokens = (
+ "bos_token" in cfg.special_tokens and "eos_token" in cfg.special_tokens
+ )
+ if (
+ tokenizer.__class__.__name__
+ in (
+ "LlamaTokenizerFast",
+ "CodeLlamaTokenizerFast",
+ )
+ and bos_or_eos_in_special_tokens
+ ):
+ tokenizer.update_post_processor()
+
+ if cfg.tokens:
+ tokenizer.add_tokens(
+ [
+ AddedToken(token, rstrip=False, lstrip=False, normalized=False)
+ for token in cfg.tokens
+ ]
+ )
+
+ # Additional special tokens are a List, and need to be treated differently than regular special
+ # tokens. We add them after we have called `add_tokens` in case these additional special tokens
+ # are new tokens.
+ #
+ # Usage:
+ #
+ # ```py
+ # special_tokens:
+ # additional_special_tokens: ["<|im_start|>", "<|im_end|>"]
+ # ```
+ if additional_special_tokens is not None:
+ tokenizer.add_special_tokens(
+ {"additional_special_tokens": additional_special_tokens}
+ )
+
+ with zero_only():
+ LOG.debug(f"EOS: {tokenizer.eos_token_id} / {tokenizer.eos_token}")
+ LOG.debug(f"BOS: {tokenizer.bos_token_id} / {tokenizer.bos_token}")
+ LOG.debug(f"PAD: {tokenizer.pad_token_id} / {tokenizer.pad_token}")
+ LOG.debug(f"UNK: {tokenizer.unk_token_id} / {tokenizer.unk_token}")
+
+ if cfg.chat_template:
+ chat_template_string = chat_templates(cfg.chat_template)
+ if cfg.default_system_message and cfg.chat_template == "chatml":
+ chat_template_string = chat_template_string.replace(
+ "you are a highly intelligent sentient being", cfg.default_system_message
+ )
+
+ tokenizer.chat_template = chat_template_string
+ else:
+ LOG.info(
+ "No Chat template selected. Consider adding a chat template for easier inference."
+ )
+ return tokenizer
+
+
+def load_model(
+ cfg: DictDefault,
+ tokenizer: PreTrainedTokenizerBase,
+ inference: bool = False,
+ reference_model: bool = False,
+) -> Tuple[PreTrainedModel, Optional[PeftConfig]]:
+ """
+ Load a model for a given configuration and tokenizer.
+ """
+ base_model = cfg.base_model
+ model_type = cfg.type_of_model
+ model_config = load_model_config(cfg)
+
+ # TODO refactor as a kwarg
+ load_in_8bit = cfg.load_in_8bit
+
+ if cfg.gradient_checkpointing == "unsloth":
+ transformers.modeling_utils.checkpoint = hf_grad_checkpoint_unsloth_wrapper
+
+ if hasattr(model_config, "model_type") and model_config.model_type == "btlm":
+ if cfg.flash_attention:
+ from axolotl.monkeypatch.btlm_attn_hijack_flash import (
+ replace_btlm_attn_with_flash_attn,
+ )
+
+ replace_btlm_attn_with_flash_attn(cfg.base_model)
+
+ if (
+ hasattr(model_config, "model_type")
+ and model_config.model_type == "stablelm_epoch"
+ ):
+ if cfg.flash_attention and cfg.sample_packing:
+ from axolotl.monkeypatch.stablelm_attn_hijack_flash import (
+ replace_stablelm_attn_with_flash_attn,
+ )
+
+ replace_stablelm_attn_with_flash_attn(cfg.base_model)
+
+ if cfg.sample_packing and cfg.s2_attention:
+ raise ValueError(
+ "Received `sample_packing=true` and `s2_attention=true`; however, \
+ shifted-sparse attention does not currently support sample packing."
+ )
+
+ if (
+ cfg.model_config_type in SUPPORTED_MULTIPACK_MODEL_TYPES
+ and cfg.flash_attention
+ and cfg.sample_packing
+ ):
+ patch_for_multipack(cfg.model_config_type, model_name=cfg.base_model)
+ elif cfg.is_llama_derived_model:
+ # Modify all llama derived models in one block
+
+ if cfg.flash_attention:
+ from axolotl.monkeypatch.llama_attn_hijack_flash import (
+ replace_llama_attn_with_flash_attn,
+ )
+
+ if cfg.sample_packing:
+ if cfg.device not in ["mps", "cpu"] and not inference:
+ LOG.info("patching with flash attention for sample packing")
+ replace_llama_attn_with_flash_attn(
+ packed=True,
+ cross_entropy=cfg.flash_attn_cross_entropy,
+ rms_norm=cfg.flash_attn_rms_norm,
+ )
+ elif cfg.s2_attention:
+ LOG.info("patching w/ flash-enabled, shifted-sparse attention")
+ replace_llama_attn_with_flash_attn(
+ packed=False,
+ cross_entropy=cfg.flash_attn_cross_entropy,
+ rms_norm=cfg.flash_attn_rms_norm,
+ use_shifted_sparse_attn=True,
+ )
+ elif cfg.xformers_attention:
+ from axolotl.monkeypatch.llama_attn_hijack_xformers import (
+ hijack_llama_attention,
+ )
+
+ LOG.info("patching with xformers attention")
+ hijack_llama_attention()
+ elif cfg.sample_packing:
+ from axolotl.monkeypatch.llama_patch_multipack import (
+ hijack_llama_prepare_4d_mask,
+ )
+
+ LOG.info("patching llama _prepare_4d_causal_attention_mask*")
+ hijack_llama_prepare_4d_mask()
+ elif cfg.s2_attention:
+ raise NotImplementedError(
+ "Shifted-sparse attention not currently implemented without flash attention."
+ )
+
+ if cfg.unsloth_cross_entropy_loss:
+ from axolotl.monkeypatch.unsloth_ import integrate_cross_entropy_loss_patch
+
+ integrate_cross_entropy_loss_patch()
+
+ if cfg.unsloth_lora_qkv or cfg.unsloth_lora_o:
+ from axolotl.monkeypatch.unsloth_ import patch_self_attn_lora
+
+ patch_self_attn_lora()
+
+ # Modify mistral derived models
+ if (
+ cfg.model_config_type == "mistral"
+ and cfg.flash_attention
+ and cfg.sample_packing
+ ):
+ from axolotl.monkeypatch.mistral_attn_hijack_flash import (
+ replace_mistral_attn_with_flash_attn,
+ )
+
+ LOG.info("patching mistral with flash attention")
+ replace_mistral_attn_with_flash_attn(packed=cfg.sample_packing)
+
+ if cfg.is_llama_derived_model and cfg.sample_packing and not inference:
+ from axolotl.monkeypatch.llama_expand_mask import hijack_expand_mask
+
+ LOG.info("patching _expand_mask")
+ hijack_expand_mask()
+
+ model_kwargs: Dict[str, Any] = {}
+
+ if cfg.model_kwargs:
+ for key, val in cfg.model_kwargs.items():
+ model_kwargs[key] = val
+
+ max_memory = cfg.max_memory
+ device_map = cfg.device_map
+
+ if cfg.gpu_memory_limit:
+ gpu_memory_limit = (
+ str(cfg.gpu_memory_limit) + "GiB"
+ if isinstance(cfg.gpu_memory_limit, int)
+ else cfg.gpu_memory_limit
+ )
+
+ max_memory = {}
+ for i in range(torch.cuda.device_count()):
+ max_memory[i] = gpu_memory_limit
+ max_memory["cpu"] = "256GiB" # something sufficiently large to fit anything
+
+ if max_memory is not None:
+ # Based on https://github.com/togethercomputer/OpenChatKit/blob/main/inference/bot.py
+ from accelerate import infer_auto_device_map
+
+ with init_empty_weights():
+ model_canvas = AutoModelForCausalLM.from_config(
+ model_config, trust_remote_code=cfg.trust_remote_code or False
+ )
+ model_canvas.tie_weights()
+ device_map = infer_auto_device_map(
+ model_canvas,
+ max_memory=max_memory,
+ dtype=cfg.torch_dtype,
+ )
+ # We can discard max_memory now as we have a device map set up for us
+ max_memory = None
+
+ model_kwargs["device_map"] = device_map
+ model_kwargs["torch_dtype"] = cfg.torch_dtype
+
+ if torch.backends.mps.is_available():
+ model_kwargs["device_map"] = "mps:0"
+
+ # TODO can we put the reference model on it's own gpu? I think we have to move logits around to calculate loss
+ # if cfg.rl:
+ # if torch.cuda.device_count() > 1:
+ # if reference_model:
+ # model_kwargs["device_map"] = "cuda:" + str(
+ # torch.cuda.current_device() + 1
+ # )
+ # else:
+ # model_kwargs["device_map"] = "cuda:" + str(torch.cuda.current_device())
+
+ if is_deepspeed_zero3_enabled():
+ del model_kwargs["device_map"]
+
+ if cfg.revision_of_model:
+ model_kwargs["revision"] = cfg.revision_of_model
+
+ if cfg.gptq:
+ if not hasattr(model_config, "quantization_config"):
+ LOG.warning("model config does not contain quantization_config information")
+ else:
+ if cfg.gptq_disable_exllama is not None:
+ model_config.quantization_config[
+ "disable_exllama"
+ ] = cfg.gptq_disable_exllama
+ model_kwargs["quantization_config"] = GPTQConfig(
+ **model_config.quantization_config
+ )
+ if cfg.adapter == "qlora" and cfg.load_in_4bit:
+ bnb_config = {
+ "load_in_4bit": True,
+ "llm_int8_threshold": 6.0,
+ "llm_int8_has_fp16_weight": False,
+ "bnb_4bit_compute_dtype": cfg.torch_dtype,
+ "bnb_4bit_use_double_quant": True,
+ "bnb_4bit_quant_type": "nf4",
+ "bnb_4bit_quant_storage": torch.bfloat16,
+ }
+ if cfg.model_config_type in ["jamba", "qwen2_moe"] and not cfg.deepspeed:
+ # for some reason, this causes the loss to be off by an order of magnitude
+ # but deepspeed needs this still in bfloat16
+ bnb_config["bnb_4bit_quant_storage"] = torch.float32
+
+ if cfg.bnb_config_kwargs:
+ bnb_config.update(cfg.bnb_config_kwargs)
+
+ model_kwargs["quantization_config"] = BitsAndBytesConfig(
+ **bnb_config,
+ )
+ elif cfg.adapter == "lora" and cfg.load_in_8bit:
+ bnb_config = {
+ "load_in_8bit": True,
+ }
+ # Exclude mamba blocks from int8 quantization for jamba
+ if cfg.model_config_type == "jamba":
+ bnb_config["llm_int8_skip_modules"] = ["mamba"]
+ model_kwargs["quantization_config"] = BitsAndBytesConfig(
+ **bnb_config,
+ )
+
+ if cfg.load_in_8bit and cfg.adapter is not None:
+ model_kwargs["load_in_8bit"] = True
+ if cfg.load_in_4bit and cfg.adapter is not None:
+ model_kwargs["load_in_4bit"] = True
+
+ # no longer needed per https://github.com/huggingface/transformers/pull/26610
+ if "quantization_config" in model_kwargs or cfg.gptq:
+ if "load_in_8bit" in model_kwargs:
+ del model_kwargs["load_in_8bit"]
+ if "load_in_4bit" in model_kwargs:
+ del model_kwargs["load_in_4bit"]
+
+ # sample packing uses custom FA2 patch
+ if cfg.flash_attention:
+ if not cfg.sample_packing:
+ if cfg.s2_attention:
+ pass
+ # most other models support flash attention, we can define exceptions as they come up
+ model_kwargs["attn_implementation"] = "flash_attention_2"
+ model_config._attn_implementation = ( # pylint: disable=protected-access
+ "flash_attention_2"
+ )
+ else:
+ if model_config.model_type in SUPPORTED_MULTIPACK_MODEL_TYPES:
+ model_kwargs["attn_implementation"] = "flash_attention_2"
+ model_config._attn_implementation = ( # pylint: disable=protected-access
+ "flash_attention_2"
+ )
+ else:
+ model_kwargs["attn_implementation"] = "eager"
+ model_config._attn_implementation = ( # pylint: disable=protected-access
+ "eager"
+ )
+ elif cfg.sdp_attention:
+ model_kwargs["attn_implementation"] = "sdpa"
+ model_config._attn_implementation = "sdpa" # pylint: disable=protected-access
+ elif cfg.eager_attention:
+ model_kwargs["attn_implementation"] = "eager"
+ model_config._attn_implementation = "eager" # pylint: disable=protected-access
+
+ if cfg.low_cpu_mem_usage:
+ model_kwargs["low_cpu_mem_usage"] = True
+
+ qlora_fsdp = cfg.fsdp and cfg.adapter == "qlora"
+
+ try:
+ skip_move_to_device = False
+ if (
+ cfg.fsdp and cfg.fsdp_config.fsdp_cpu_ram_efficient_loading
+ ) and not qlora_fsdp:
+ model = load_sharded_model(
+ base_model,
+ model_config,
+ cfg,
+ torch_dtype=cfg.torch_dtype,
+ )
+ skip_move_to_device = True
+ elif (
+ qlora_fsdp
+ and cfg.fsdp_config.fsdp_cpu_ram_efficient_loading
+ and cfg.model_config_type == "dbrx"
+ ):
+ quant_storage = cfg.torch_dtype
+ model = load_sharded_model_quant(
+ base_model,
+ model_config,
+ cfg,
+ quant_storage=quant_storage,
+ )
+ skip_move_to_device = True
+ elif (
+ model_config.model_type == "llama"
+ and not cfg.trust_remote_code
+ and not cfg.gptq
+ ):
+ from transformers import LlamaForCausalLM
+
+ model = LlamaForCausalLM.from_pretrained(
+ base_model,
+ config=model_config,
+ **model_kwargs,
+ )
+
+ if cfg.flash_attention and not inference:
+ from axolotl.monkeypatch.llama_attn_hijack_flash import (
+ is_xformers_swiglu_available,
+ replace_llama_mlp_with_swiglu,
+ replace_llama_qkv_with_fused,
+ )
+
+ if cfg.flash_attn_fuse_mlp and is_xformers_swiglu_available():
+ LOG.info("patching with SwiGLU")
+ replace_llama_mlp_with_swiglu(model)
+
+ if cfg.flash_attn_fuse_qkv:
+ LOG.info("patching with fused QKV")
+ replace_llama_qkv_with_fused(model)
+ elif model_type == "MambaLMHeadModel":
+ # FIXME this is janky at best and hacked together to make it work
+ MambaLMHeadModel = fix_mamba_attn_for_loss() # pylint: disable=invalid-name
+
+ model_kwargs["dtype"] = model_kwargs["torch_dtype"]
+ model_kwargs["device"] = torch.cuda.current_device()
+ del model_kwargs["torch_dtype"]
+ del model_kwargs["device_map"]
+
+ model = MambaLMHeadModel.from_pretrained(
+ base_model,
+ **model_kwargs,
+ )
+ elif model_type and not cfg.trust_remote_code:
+ if cfg.gptq:
+ model = AutoModelForCausalLM.from_pretrained(
+ base_model,
+ config=model_config,
+ trust_remote_code=cfg.trust_remote_code or False,
+ **model_kwargs,
+ )
+ else:
+ model = getattr(transformers, model_type).from_pretrained(
+ base_model,
+ config=model_config,
+ trust_remote_code=cfg.trust_remote_code or False,
+ **model_kwargs,
+ )
+ else:
+ # Shouldn't be a problem most of the time. will obviously error if the model doesn't support this
+ # when training starts
+ if (
+ hasattr(model_config, "max_seq_len")
+ and model_config.max_seq_len
+ and cfg.sequence_len > model_config.max_seq_len
+ ):
+ model_config.max_seq_len = cfg.sequence_len
+ LOG.warning(f"increasing context length to {cfg.sequence_len}")
+ elif (
+ hasattr(model_config, "max_sequence_length")
+ and model_config.max_sequence_length
+ and cfg.sequence_len > model_config.max_sequence_length
+ ):
+ model_config.max_sequence_length = cfg.sequence_len
+ LOG.warning(f"increasing context length to {cfg.sequence_len}")
+ if cfg.gptq:
+ model = AutoModelForCausalLM.from_pretrained(
+ base_model,
+ config=model_config,
+ trust_remote_code=cfg.trust_remote_code or False,
+ **model_kwargs,
+ )
+ else:
+ if qlora_fsdp and cfg.fsdp_config.fsdp_cpu_ram_efficient_loading:
+ skip_move_to_device = True
+ if "device_map" in model_kwargs:
+ del model_kwargs["device_map"]
+
+ model = AutoModelForCausalLM.from_pretrained(
+ base_model,
+ config=model_config,
+ trust_remote_code=cfg.trust_remote_code or False,
+ **model_kwargs,
+ )
+ except Exception as err: # pylint: disable=broad-exception-caught
+ LOG.exception(err)
+ raise err
+
+ if isinstance(model, (PeftModel, PeftModelForCausalLM)) and not qlora_fsdp:
+ model = model.merge_and_unload()
+
+ embeddings_len = (
+ math.ceil(len(tokenizer) / 32) * 32
+ if cfg.resize_token_embeddings_to_32x
+ else len(tokenizer)
+ )
+ if (
+ hasattr(model, "get_input_embeddings")
+ and model.get_input_embeddings().num_embeddings < embeddings_len
+ ):
+ model.resize_token_embeddings(embeddings_len)
+ else:
+ model.tie_weights()
+
+ if (
+ hasattr(model, "config")
+ and hasattr(model.config, "max_position_embeddings")
+ and model.config.max_position_embeddings
+ and cfg.sequence_len > model.config.max_position_embeddings
+ ):
+ LOG.warning(
+ f"increasing model.config.max_position_embeddings from {model.config.max_position_embeddings} to {cfg.sequence_len}"
+ )
+ model.config.max_position_embeddings = cfg.sequence_len
+
+ if (
+ hasattr(model, "config")
+ and hasattr(model.config, "bos_token_id")
+ and model.config.bos_token_id
+ and model.config.bos_token_id != tokenizer.bos_token_id
+ ):
+ model.config.bos_token_id = tokenizer.bos_token_id
+
+ if (
+ hasattr(model, "config")
+ and hasattr(model.config, "eos_token_id")
+ and model.config.eos_token_id
+ and model.config.eos_token_id != tokenizer.eos_token_id
+ ):
+ model.config.eos_token_id = tokenizer.eos_token_id
+
+ if hasattr(model, "device") and model.device.type in ("cuda", "mps"):
+ log_gpu_memory_usage(LOG, "after model load", model.device)
+
+ # make sure these are fp32 per Ramesh et al. (2021)
+ embedding_modules = get_linear_embedding_layers(cfg.model_config_type)
+ if not cfg.fsdp:
+ # FSDP doesn't like mixed Float and BFloat16
+ for name, module in model.named_modules():
+ if "norm" in name or name.endswith(".gate"):
+ module.to(torch.float32)
+ if model_config.model_type == "btlm":
+ # don't upcast lm_head for btlm
+ continue
+ if any(m in name for m in embedding_modules):
+ if hasattr(module, "weight"):
+ module.to(torch.float32)
+
+ needs_fa2_dtype = cfg.adapter or cfg.fsdp
+ skip_prepare_model_for_kbit_training = False
+
+ if is_deepspeed_zero3_enabled():
+ from deepspeed.utils import ( # pylint: disable=no-name-in-module
+ set_z3_leaf_modules,
+ )
+
+ if cfg.model_config_type == "mixtral":
+ moe_block = get_module_class_from_name(model, "MixtralSparseMoeBlock")
+ set_z3_leaf_modules(model, [moe_block])
+ elif cfg.model_config_type == "dbrx":
+ moe_block = get_module_class_from_name(model, "DbrxFFN")
+ set_z3_leaf_modules(model, [moe_block])
+
+ if cfg.model_config_type == "qwen" and cfg.adapter == "lora":
+ # Qwen doesn't play nicely with LoRA if this is enabled
+ skip_prepare_model_for_kbit_training = True
+
+ loftq_bits = cfg.peft and cfg.peft.loftq_config and cfg.peft.loftq_config.loftq_bits
+ if cfg.adapter == "lora" and loftq_bits:
+ skip_prepare_model_for_kbit_training = True
+
+ if qlora_fsdp or (cfg.fsdp and cfg.fsdp_config.fsdp_cpu_ram_efficient_loading):
+ # make sure everything is in the same dtype
+ skip_prepare_model_for_kbit_training = True
+
+ if cfg.adapter in ["lora", "qlora"]:
+ if cfg.gradient_checkpointing:
+ model.gradient_checkpointing_enable(
+ gradient_checkpointing_kwargs=cfg.gradient_checkpointing_kwargs
+ )
+ if (
+ cfg.load_in_8bit or cfg.load_in_4bit
+ ) and not skip_prepare_model_for_kbit_training:
+ LOG.info("converting PEFT model w/ prepare_model_for_kbit_training")
+ model = prepare_model_for_kbit_training(
+ model, use_gradient_checkpointing=cfg.gradient_checkpointing
+ )
+ needs_fa2_dtype = True
+
+ # LlamaRMSNorm layers are in fp32 after kbit_training or full finetune, so we need to
+ # convert them back to fp16/bf16 for flash-attn compatibility.
+ if (needs_fa2_dtype or cfg.flash_attention) and not qlora_fsdp:
+ LOG.info("converting modules to %s for flash attention", cfg.torch_dtype)
+ for name, module in model.named_modules():
+ if "norm" in name:
+ module.to(cfg.torch_dtype)
+ if any(m in name for m in embedding_modules):
+ if hasattr(module, "weight"):
+ module.to(cfg.torch_dtype)
+
+ lora_config = None
+ if not reference_model or cfg.lora_model_dir:
+ # if we're not loading the reference model, then we're loading the model for training
+ # then the dpo trainer doesn't want the peft model loaded over it, it just wants the lora/peft config
+ if cfg.adapter and cfg.rl in ["dpo", "ipo", "kto_pair"] and not cfg.merge_lora:
+ _, lora_config = load_lora(model, cfg, inference=False, config_only=True)
+ else:
+ model, lora_config = load_adapter(model, cfg, cfg.adapter)
+
+ if (
+ cfg.ddp
+ and not load_in_8bit
+ and not (cfg.rl and cfg.load_in_4bit)
+ and not skip_move_to_device
+ ):
+ # TODO revaldate this conditional
+ model.to(f"cuda:{cfg.local_rank}")
+
+ if torch.cuda.device_count() > 1 and int(os.getenv("WORLD_SIZE", "1")) == 1:
+ setattr(model, "is_parallelizable", True)
+ setattr(model, "model_parallel", True)
+
+ requires_grad = []
+ for name, param in model.named_parameters(recurse=True):
+ if param.requires_grad:
+ requires_grad.append(f"{name}: {param.requires_grad}")
+ if len(requires_grad) == 0:
+ LOG.warning("there are no parameters that require gradient updates")
+ if hasattr(model, "config"):
+ model.config.use_cache = False
+
+ if cfg.flash_optimum:
+ from optimum.bettertransformer import BetterTransformer
+
+ model = BetterTransformer.transform(model)
+
+ if cfg.adapter is not None:
+ log_gpu_memory_usage(LOG, "after adapters", model.device)
+
+ if cfg.unsloth_lora_mlp:
+ from axolotl.monkeypatch.unsloth_ import integrate_lora_mlp_patch
+
+ integrate_lora_mlp_patch(model)
+ if cfg.unsloth_lora_qkv or cfg.unsloth_lora_o:
+ from axolotl.monkeypatch.unsloth_ import integrate_lora_patch
+
+ integrate_lora_patch(model, cfg)
+
+ # TODO resume_from_checkpoint handling
+ return model, lora_config
+
+
+def load_adapter(model, cfg, adapter, inference=False):
+ # type: (PreTrainedModel, DictDefault, Optional[str], bool) -> Tuple[PreTrainedModel, Optional[PeftConfig]]
+
+ if adapter is None:
+ return model, None
+ if hasattr(model, "enable_input_require_grads"):
+ model.enable_input_require_grads()
+ if adapter in ["lora", "qlora"]:
+ return load_lora(model, cfg, inference=inference)
+ if adapter == "llama-adapter":
+ return load_llama_adapter(model, cfg)
+
+ raise NotImplementedError(f"{adapter} peft adapter not available")
+
+
+def load_llama_adapter(model, cfg):
+ # type: (PreTrainedModel, DictDefault) -> Tuple[PreTrainedModel, Optional[PeftConfig]]
+ from peft import AdaptionPromptConfig, get_peft_model
+
+ peft_config = AdaptionPromptConfig(
+ adapter_layers=cfg.peft_adapter.layers, # layers (L)
+ adapter_len=cfg.peft_adapter.len, # prompt length (K)
+ task_type="CAUSAL_LM",
+ )
+
+ if cfg.lora_model_dir:
+ LOG.debug("Loading pretrained PEFT - llama_adapter")
+ model = PeftModel.from_pretrained(
+ model,
+ cfg.lora_model_dir,
+ torch_dtype=torch.float16,
+ )
+ else:
+ model = get_peft_model(model, peft_config)
+
+ model.print_trainable_parameters()
+
+ return model, peft_config
+
+
+def find_all_linear_names(model):
+ cls = (bnb.nn.Linear4bit, bnb.nn.Linear8bitLt, torch.nn.Linear, QuantLinear)
+ lora_module_names = set()
+ for name, module in model.named_modules():
+ if (
+ isinstance(module, cls)
+ or "Linear" in module.__class__.__name__
+ and module.__class__.__name__ not in ("LlamaLinearScalingRotaryEmbedding",)
+ ):
+ names = name.split(".")
+ lora_module_names.add(names[0] if len(names) == 1 else names[-1])
+
+ embedding_modules = get_linear_embedding_layers(model.config.model_type)
+ output_embedding = embedding_modules[1]
+ if output_embedding in lora_module_names: # needed for 16-bit
+ lora_module_names.remove(output_embedding)
+
+ return list(lora_module_names)
+
+
+def setup_quantized_meta_for_peft(model: nn.Module):
+ """Replaces `quant_state.to` with a dummy function to prevent PEFT from moving `quant_state` to meta device"""
+
+ def temp_to_method(self, *args, **kwargs): # pylint: disable=unused-argument
+ return self
+
+ for param in model.parameters():
+ if isinstance(param, Params4bit):
+ param.quant_state._orig_to = ( # pylint: disable=protected-access
+ param.quant_state.to
+ )
+ param.quant_state.to = types.MethodType(temp_to_method, param.quant_state)
+
+
+def setup_quantized_peft_meta_for_training(model: nn.Module):
+ """Replaces dummy `quant_state.to` method with the original function to allow training to continue"""
+ for param in model.parameters():
+ if isinstance(param, Params4bit) and hasattr(param.quant_state, "_orig_to"):
+ param.quant_state.to = (
+ param.quant_state._orig_to # pylint: disable=protected-access
+ )
+ param.quant_state._orig_to = None # pylint: disable=protected-access
+
+
+def load_lora(model, cfg, inference=False, config_only=False):
+ # type: (PreTrainedModel, DictDefault, bool, bool) -> Tuple[Optional[PreTrainedModel], Optional[PeftConfig]]
+
+ from peft import LoraConfig, get_peft_model
+
+ lora_target_modules = list(cfg.lora_target_modules or [])
+
+ if cfg.lora_target_linear:
+ linear_names = find_all_linear_names(model)
+ LOG.info(f"found linear modules: {repr(linear_names)}")
+ lora_target_modules = list(set(lora_target_modules + linear_names))
+
+ lora_config_kwargs = {}
+ loftq_bits = cfg.peft and cfg.peft.loftq_config and cfg.peft.loftq_config.loftq_bits
+ if loftq_bits:
+ lora_config_kwargs["loftq_config"] = LoftQConfig(loftq_bits=loftq_bits)
+ lora_config_kwargs["init_lora_weights"] = "loftq"
+ if cfg.peft_use_dora:
+ lora_config_kwargs["use_dora"] = cfg.peft_use_dora
+ if cfg.peft_use_rslora:
+ lora_config_kwargs["use_rslora"] = cfg.peft_use_rslora
+ if cfg.peft_layer_replication:
+ lora_config_kwargs["layer_replication"] = cfg.peft_layer_replication
+
+ lora_config = LoraConfig(
+ r=cfg.lora_r,
+ lora_alpha=cfg.lora_alpha,
+ target_modules=lora_target_modules,
+ layers_to_transform=cfg.peft_layers_to_transform,
+ lora_dropout=cfg.lora_dropout,
+ fan_in_fan_out=cfg.lora_fan_in_fan_out,
+ modules_to_save=cfg.lora_modules_to_save if cfg.lora_modules_to_save else None,
+ bias="none",
+ task_type="CAUSAL_LM",
+ **lora_config_kwargs,
+ )
+
+ if config_only:
+ return None, lora_config
+
+ rank = int(os.environ.get("LOCAL_RANK", 0))
+
+ if (
+ cfg.fsdp
+ and cfg.adapter
+ and cfg.fsdp_config.fsdp_cpu_ram_efficient_loading
+ and rank != 0
+ ):
+ setup_quantized_meta_for_peft(model)
+
+ if cfg.lora_model_dir:
+ LOG.debug("Loading pretrained PEFT - LoRA")
+ model_kwargs: Any = {}
+ if cfg.lora_on_cpu:
+ model_kwargs["max_memory"] = {"cpu": "256GiB"}
+ model_kwargs["device_map"] = {"": "cpu"}
+ model = PeftModel.from_pretrained(
+ model,
+ cfg.lora_model_dir,
+ is_trainable=(not inference),
+ **model_kwargs,
+ )
+ else:
+ model = get_peft_model(model, lora_config)
+
+ if rank == 0:
+ try:
+ model.print_trainable_parameters()
+ except AttributeError as exc:
+ LOG.warning(
+ "Exception caught during model.print_trainable_parameters(): %s", exc
+ )
+ elif (
+ cfg.fsdp
+ and cfg.adapter
+ and cfg.fsdp_config.fsdp_cpu_ram_efficient_loading
+ and rank != 0
+ ):
+ setup_quantized_peft_meta_for_training(model)
+
+ return model, lora_config
+
+
+def ensure_dtype(model, dtype=torch.bfloat16):
+ for name, module in model.named_modules():
+ try:
+ if module.weight.dtype != dtype:
+ print(f"Converting module {name}: {module.weight.dtype} -> {dtype}")
+ module.to(dtype)
+ except AttributeError:
+ pass
diff --git a/src/axolotl/utils/samplers/__init__.py b/src/axolotl/utils/samplers/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..96e00a5d26fbd3359f3d23192c396dea8d184c71
--- /dev/null
+++ b/src/axolotl/utils/samplers/__init__.py
@@ -0,0 +1,5 @@
+"""
+axolotl samplers module
+"""
+from .multipack import MultipackBatchSampler # noqa: F401
+from .utils import get_dataset_lengths # noqa: F401
diff --git a/src/axolotl/utils/samplers/__pycache__/__init__.cpython-310.pyc b/src/axolotl/utils/samplers/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..0fc902648bebda5d389fadb6d97a854bfa12df92
Binary files /dev/null and b/src/axolotl/utils/samplers/__pycache__/__init__.cpython-310.pyc differ
diff --git a/src/axolotl/utils/samplers/__pycache__/multipack.cpython-310.pyc b/src/axolotl/utils/samplers/__pycache__/multipack.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..d0ab07db179c65272c3456f479bf485ab1929f58
Binary files /dev/null and b/src/axolotl/utils/samplers/__pycache__/multipack.cpython-310.pyc differ
diff --git a/src/axolotl/utils/samplers/__pycache__/utils.cpython-310.pyc b/src/axolotl/utils/samplers/__pycache__/utils.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..7c89645311b3cf529a3b30b2e9d76ace5d672b5b
Binary files /dev/null and b/src/axolotl/utils/samplers/__pycache__/utils.cpython-310.pyc differ
diff --git a/src/axolotl/utils/samplers/multipack.py b/src/axolotl/utils/samplers/multipack.py
new file mode 100644
index 0000000000000000000000000000000000000000..cf47d9639b76bb2677001b238b4198ed95c2ab11
--- /dev/null
+++ b/src/axolotl/utils/samplers/multipack.py
@@ -0,0 +1,202 @@
+# pylint: skip-file
+"""
+Multipack Batch Sampler
+"""
+import logging
+import math
+import os
+from typing import Any, Iterable, List, Union
+
+import numba
+import numpy as np
+from torch.utils.data import BatchSampler, Sampler
+
+LOG = logging.getLogger("axolotl.utils.samplers.multipack")
+
+
+@numba.njit
+def ffd_check(a: np.ndarray, c: int, n: int):
+ # First-fit-decreasing bin packing
+ # Check if a[] could fit in n bins with capacity c
+ # https://en.wikipedia.org/wiki/First-fit-decreasing_bin_packing
+
+ a = np.sort(a)[::-1]
+ bins = np.full((n,), c, dtype=a.dtype)
+ for size in a:
+ not_found = True
+ for idx in range(n):
+ if bins[idx] >= size:
+ bins[idx] -= size
+ not_found = False
+ break
+
+ if not_found:
+ return False
+
+ return True
+
+
+@numba.njit
+def ffd_with_result(a: np.ndarray, c: int, start_index: int):
+ # First-fit-decreasing bin packing (with result return)
+
+ indices = np.argsort(a)[::-1]
+ a = a[indices]
+
+ bins: List[Any] = []
+ bins_result: List[Any] = []
+ for a_id, size in enumerate(a):
+ add_new = True
+ for idx in range(len(bins)):
+ if bins[idx] >= size:
+ bins[idx] -= size
+ bins_result[idx].append(indices[a_id] + start_index)
+ add_new = False
+ break
+
+ if add_new:
+ bins.append(c - size)
+ bins_result.append([indices[a_id] + start_index])
+
+ return bins_result
+
+
+@numba.njit
+def allocate(
+ lengths: np.ndarray, lengths_cumsum: np.ndarray, rank: int, c: int, n: int
+):
+ # Dynamic batch allocator, similar to Multifit
+ # https://en.wikipedia.org/wiki/Multifit_algorithm
+ # ~99.5% efficiency on OpenChat training set (12 * 2048 ctx len)
+
+ s = 0
+ start_index = 0
+ result = []
+
+ while True:
+ # binary search [l, r)
+ left = 1
+ right = 1 + np.searchsorted(lengths_cumsum[start_index:], s + c * n, "right")
+
+ while right - left > 1:
+ mid = (left + right) // 2
+ if ffd_check(lengths[start_index : start_index + mid], c, n):
+ left = mid
+ else:
+ right = mid
+
+ # use length l
+ batch = ffd_with_result(
+ lengths[start_index : start_index + left], c, start_index
+ )
+ assert len(batch) <= n
+ if len(batch) < n:
+ break
+
+ start_index += left
+ s = lengths_cumsum[start_index - 1]
+
+ # add local rank
+ result.append(batch[rank])
+
+ return result, s, len(result) * c * n
+
+
+class MultipackBatchSampler(BatchSampler):
+ """
+ Batch Sampler class for multipack
+ """
+
+ def __init__(
+ self,
+ sampler: Union[Sampler[int], Iterable[int]],
+ batch_size: int,
+ drop_last: bool,
+ batch_max_len: int,
+ lengths: np.ndarray,
+ packing_efficiency_estimate: float = 1.0,
+ ):
+ super().__init__(sampler, batch_size, drop_last)
+ self.batch_size = batch_size
+ self.batch_max_len = batch_max_len
+ self.lengths: np.ndarray = lengths
+ self.packing_efficiency_estimate = packing_efficiency_estimate or 1.0
+
+ assert isinstance(self.lengths, np.ndarray)
+
+ self.epoch = 0
+
+ # statistics
+ self.eff_total_used = 0
+ self.eff_total_slots = 0
+
+ def set_epoch(self, epoch: int):
+ self.epoch = epoch
+
+ def generate_batches(self, set_stats=False):
+ indices = [idx for idx in self.sampler]
+
+ lengths = self.lengths[indices]
+ lengths_cumsum = np.cumsum(lengths)
+
+ batches, total_used, total_slots = allocate(
+ lengths=lengths,
+ lengths_cumsum=lengths_cumsum,
+ rank=0,
+ c=self.batch_max_len,
+ n=1,
+ )
+
+ batches = [
+ [
+ [indices[b_idx] for b_idx in batch]
+ for batch in batches[i : i + self.batch_size]
+ ]
+ for i in range(0, len(batches), self.batch_size)
+ ]
+
+ # statistics
+ if set_stats:
+ self.eff_total_used += total_used
+ self.eff_total_slots += total_slots
+
+ return batches
+
+ def __iter__(self):
+ batches = self.generate_batches(set_stats=True)
+ return iter(batches)
+
+ def num_batches(self):
+ batches = self.generate_batches(set_stats=True)
+ return len(batches)
+
+ def efficiency(self):
+ return self.eff_total_used / self.eff_total_slots
+
+ def __len__(self):
+ self.num_batches()
+ return self._len_est()
+
+ def _len_est(self):
+ world_size = int(os.getenv("WORLD_SIZE", "1"))
+ lengths_sum = np.sum(self.lengths)
+ lengths_sum_per_device = lengths_sum // world_size
+ LOG.info(
+ f"packing_efficiency_estimate: {self.packing_efficiency_estimate} "
+ f"total_num_tokens per device: {lengths_sum_per_device}"
+ )
+
+ # shave off 1% + 1 for dealing with variance in packing from random sampler to sampler
+ return max(
+ 0,
+ (
+ world_size
+ * math.floor(
+ 0.99
+ * lengths_sum_per_device
+ / self.packing_efficiency_estimate
+ // (self.batch_max_len * self.batch_size)
+ )
+ - 1
+ ),
+ )
diff --git a/src/axolotl/utils/samplers/utils.py b/src/axolotl/utils/samplers/utils.py
new file mode 100644
index 0000000000000000000000000000000000000000..e4af4e5f35f3cc25a8e3bb8fc908c3400cf60eff
--- /dev/null
+++ b/src/axolotl/utils/samplers/utils.py
@@ -0,0 +1,17 @@
+"""
+helper util to calculate dataset lengths
+"""
+import numpy as np
+
+
+def get_dataset_lengths(dataset):
+ if "length" in dataset.data.column_names:
+ lengths = np.array(dataset.data.column("length"))
+ elif "position_ids" in dataset.data.column_names:
+ position_ids = dataset.data.column("position_ids")
+ lengths = np.array([x[-1] + 1 for x in position_ids])
+ else:
+ input_ids = dataset.data.column("input_ids")
+ lengths = np.vectorize(len)(np.array(input_ids, dtype=object))
+ return lengths
+ return lengths
diff --git a/src/axolotl/utils/schedulers.py b/src/axolotl/utils/schedulers.py
new file mode 100644
index 0000000000000000000000000000000000000000..94387e5ab88af036e82db3066ef291aeaec01e11
--- /dev/null
+++ b/src/axolotl/utils/schedulers.py
@@ -0,0 +1,219 @@
+"""Module for custom LRScheduler class"""
+import math
+from functools import partial
+
+from torch.optim import Optimizer
+from torch.optim.lr_scheduler import LambdaLR, LRScheduler
+
+
+class InterpolatingLogScheduler(LRScheduler):
+ """
+ A scheduler that interpolates learning rates in a logarithmic fashion
+ """
+
+ def __init__(self, optimizer, num_steps, min_lr, max_lr, last_epoch=-1):
+ """A scheduler that interpolates learning rates in a logarithmic fashion
+
+ Args:
+ - optimizer: pytorch optimizer
+ - num_steps: int, the number of steps over which to increase from the min_lr to the max_lr
+ - min_lr: float, the minimum learning rate
+ - max_lr: float, the maximum learning rate
+
+ Usage:
+ fc = nn.Linear(1,1)
+ optimizer = optim.Adam(fc.parameters())
+ lr_scheduler = InterpolatingLogScheduler(optimizer, num_steps=400, min_lr=1e-6, max_lr=1e-4)
+ """
+ self.num_steps = num_steps
+ self.min_lr = min_lr
+ self.max_lr = max_lr
+ self.q = (max_lr / min_lr) ** ( # pylint: disable=invalid-name
+ 1 / (num_steps - 1)
+ )
+ super().__init__(optimizer, last_epoch)
+
+ def get_lr(self):
+ if self.last_epoch <= 0:
+ lrs = [self.min_lr for base_lr in self.base_lrs]
+ elif self.last_epoch < self.num_steps:
+ lrs = [
+ self.min_lr * (self.q ** (self.last_epoch - 1))
+ for base_lr in self.base_lrs
+ ]
+ else:
+ lrs = [self.max_lr for base_lr in self.base_lrs]
+
+ return lrs
+
+
+def _get_cosine_schedule_with_quadratic_warmup_lr_lambda(
+ current_step: int,
+ *,
+ num_warmup_steps: int,
+ num_training_steps: int,
+ num_cycles: float,
+):
+ if current_step < num_warmup_steps:
+ return (float(current_step) / float(max(1, num_warmup_steps))) ** 2
+ progress = float(current_step - num_warmup_steps) / float(
+ max(1, num_training_steps - num_warmup_steps)
+ )
+ return max(
+ 0.0, 0.5 * (1.0 + math.cos(math.pi * float(num_cycles) * 2.0 * progress))
+ )
+
+
+def get_cosine_schedule_with_quadratic_warmup(
+ optimizer: Optimizer,
+ num_warmup_steps: int,
+ num_training_steps: int,
+ num_cycles: float = 0.5,
+ last_epoch: int = -1,
+):
+ """
+ Create a schedule with a learning rate that decreases following the values of the cosine function between the
+ initial lr set in the optimizer to 0, after a warmup period during which it increases linearly between 0 and the
+ initial lr set in the optimizer.
+
+ Args:
+ optimizer ([`~torch.optim.Optimizer`]):
+ The optimizer for which to schedule the learning rate.
+ num_warmup_steps (`int`):
+ The number of steps for the warmup phase.
+ num_training_steps (`int`):
+ The total number of training steps.
+ num_cycles (`float`, *optional*, defaults to 0.5):
+ The number of waves in the cosine schedule (the defaults is to just decrease from the max value to 0
+ following a half-cosine).
+ last_epoch (`int`, *optional*, defaults to -1):
+ The index of the last epoch when resuming training.
+
+ Return:
+ `torch.optim.lr_scheduler.LambdaLR` with the appropriate schedule.
+ """
+
+ lr_lambda = partial(
+ _get_cosine_schedule_with_quadratic_warmup_lr_lambda,
+ num_warmup_steps=num_warmup_steps,
+ num_training_steps=num_training_steps,
+ num_cycles=num_cycles,
+ )
+ return LambdaLR(optimizer, lr_lambda, last_epoch)
+
+
+def _get_cosine_schedule_with_min_lr_lambda(
+ current_step: int,
+ *,
+ num_warmup_steps: int,
+ num_training_steps: int,
+ min_lr_ratio: float,
+):
+ # Warm up
+ if current_step < num_warmup_steps:
+ return float(current_step) / float(max(1, num_warmup_steps))
+
+ # Cosine learning rate decay
+ progress = float(current_step - num_warmup_steps) / float(
+ max(1, num_training_steps - num_warmup_steps)
+ )
+ scaling = 0.5 * (1.0 + math.cos(math.pi * progress))
+ return (1 - min_lr_ratio) * scaling + min_lr_ratio
+
+
+def get_cosine_schedule_with_min_lr(
+ optimizer: Optimizer,
+ num_warmup_steps: int,
+ num_training_steps: int,
+ min_lr_ratio: float = 0.0,
+):
+ """
+ Create a learning rate schedule which has:
+ - linear warmup from 0 -> `max_lr` over `num_warmup_steps`
+ - cosine learning rate annealing from `max_lr` -> `min_lr` over `num_training_steps`
+ """
+
+ lr_lambda = partial(
+ _get_cosine_schedule_with_min_lr_lambda,
+ num_warmup_steps=num_warmup_steps,
+ num_training_steps=num_training_steps,
+ min_lr_ratio=min_lr_ratio,
+ )
+ return LambdaLR(optimizer, lr_lambda)
+
+
+def _get_cosine_schedule_with_warmup_decay_constant_lr_lambda(
+ current_step: int,
+ *,
+ num_warmup_steps: int,
+ num_training_steps: int,
+ constant_lr_ratio: float,
+ min_lr_ratio: float,
+ num_cycles: float,
+):
+ if current_step < num_warmup_steps:
+ return float(current_step) / float(max(1, num_warmup_steps))
+
+ num_constant_steps = int(num_training_steps * constant_lr_ratio)
+ current_step = min(current_step, num_constant_steps)
+
+ progress = float(current_step - num_warmup_steps) / float(
+ max(1, num_constant_steps - num_warmup_steps)
+ )
+
+ return (
+ max(
+ 0,
+ (1 - min_lr_ratio)
+ * 0.5
+ * (1.0 + math.cos(math.pi * float(num_cycles) * 2.0 * progress)),
+ )
+ + min_lr_ratio
+ )
+
+
+def get_cosine_schedule_with_warmup_decay_constant(
+ optimizer: Optimizer,
+ num_warmup_steps: int,
+ num_training_steps: int,
+ constant_lr_ratio: float,
+ min_lr_ratio: float,
+ num_cycles: float = 0.5,
+ last_epoch: int = -1,
+):
+ """
+ Implementation of Continual Pre-Training of Large Language Models: How to (re)warm your model? (https://arxiv.org/pdf/2308.04014.pdf)
+ Create a schedule with a learning rate that decreases following the values of the cosine function between the
+ initial lr set in the optimizer to min_lr_ratio until num_training_steps * constant_lr_ratio, after constant_rate returns constant value of min_rate
+ , after a warmup period during which it increases linearly between 0 and the initial lr set in the optimizer.
+
+ Args:
+ optimizer ([`~torch.optim.Optimizer`]):
+ The optimizer for which to schedule the learning rate.
+ num_warmup_steps (`int`):
+ The number of steps for the warmup phase.
+ num_training_steps (`int`):
+ The total number of training steps.
+ constant_lr_ratio: (`float`):
+ The ratio of num_training_steps to decrease by cosine function.
+ min_lr_ratio: (`float):
+ The ratio of maximum learning rate for cosine function to decay to minimum learning rate.
+ num_cycles (`float`, *optional*, defaults to 0.5):
+ The number of waves in the cosine schedule (the defaults is to just decrease from the max value to 0
+ following a half-cosine).
+ last_epoch (`int`, *optional*, defaults to -1):
+ The index of the last epoch when resuming training.
+
+ Return:
+ `torch.optim.lr_scheduler.LambdaLR` with the appropriate schedule.
+ """
+
+ lr_lambda = partial(
+ _get_cosine_schedule_with_warmup_decay_constant_lr_lambda,
+ num_warmup_steps=num_warmup_steps,
+ num_training_steps=num_training_steps,
+ constant_lr_ratio=constant_lr_ratio,
+ min_lr_ratio=min_lr_ratio,
+ num_cycles=num_cycles,
+ )
+ return LambdaLR(optimizer, lr_lambda, last_epoch)
diff --git a/src/axolotl/utils/tokenization.py b/src/axolotl/utils/tokenization.py
new file mode 100644
index 0000000000000000000000000000000000000000..845296b7a61f868aa5065ce968b0942167768863
--- /dev/null
+++ b/src/axolotl/utils/tokenization.py
@@ -0,0 +1,157 @@
+"""Module for tokenization utilities"""
+
+import logging
+import re
+from typing import Dict, List
+
+from termcolor import colored
+
+LOG = logging.getLogger("axolotl")
+
+
+def check_dataset_labels(
+ dataset,
+ tokenizer,
+ num_examples=5,
+ text_only=False,
+ rl_mode=False,
+):
+ # the dataset is already shuffled, so let's just check the first 5 elements
+ for idx in range(num_examples):
+ if not rl_mode:
+ check_example_labels(dataset[idx], tokenizer, text_only=text_only)
+ else:
+ check_rl_example_labels(dataset[idx], tokenizer, text_only=text_only)
+
+
+def check_example_labels(example, tokenizer, text_only=False):
+ # Get the input_ids, labels, and attention_mask from the dataset
+ input_ids = example["input_ids"]
+ labels = example["labels"]
+
+ # You can compare the input_ids and labels element-wise
+ # Remember to ignore positions with IGNORE_TOKEN_ID (if you use it) or attention_mask equal to 0
+ colored_tokens = []
+ for _, (input_id, label_id) in enumerate(zip(input_ids, labels)):
+ decoded_input_token = tokenizer.decode(input_id)
+ # Choose the color based on whether the label has the ignore value or not
+ color = "red" if label_id == -100 else ("yellow" if label_id == 0 else "green")
+ colored_token = colored(decoded_input_token, color) + (
+ not text_only and colored(f"({label_id}, {input_id})", "white") or ""
+ )
+ colored_tokens.append(colored_token)
+
+ delimiter = "" if text_only else " "
+ LOG.info(delimiter.join(colored_tokens))
+ LOG.info("\n\n\n")
+
+ return " ".join(colored_tokens)
+
+
+def color_token_for_rl_debug(decoded_token, encoded_token, color, text_only):
+ """Helper function to color tokens based on their type."""
+ colored_text = colored(decoded_token, color)
+ return (
+ colored_text
+ if text_only
+ else f"{colored_text}{colored(f'({encoded_token})', 'white')}"
+ )
+
+
+def process_tokens_for_rl_debug(tokens, color, tokenizer, text_only):
+ """Helper function to process and color tokens."""
+ colored_tokens = [
+ color_token_for_rl_debug(tokenizer.decode(token), token, color, text_only)
+ for token in tokenizer.encode(tokens)
+ ]
+ return colored_tokens
+
+
+def check_rl_example_labels(example, tokenizer, text_only=False):
+ field_prompt, field_chosen, field_rejected = "prompt", "chosen", "rejected"
+
+ input_tokens = example[field_prompt]
+ labels_chosen, labels_rejected = example[field_chosen], example[field_rejected]
+
+ # Process and color each type of token
+ colored_tokens = process_tokens_for_rl_debug(
+ input_tokens, "yellow", tokenizer, text_only
+ )
+ colored_chosens = process_tokens_for_rl_debug(
+ labels_chosen, "green", tokenizer, text_only
+ )
+ colored_rejecteds = process_tokens_for_rl_debug(
+ labels_rejected, "red", tokenizer, text_only
+ )
+
+ # Create a delimiter based on text_only flag
+ delimiter = "" if text_only else " "
+
+ # Logging information
+ LOG.info(f"INPUT PROMPT: {delimiter.join(colored_tokens)}\n\n")
+ LOG.info(f"CHOSEN RESPONSE: {delimiter.join(colored_chosens)}\n\n")
+ LOG.info(f"REJECTED RESPONSE: {delimiter.join(colored_rejecteds)}\n\n\n")
+
+ return delimiter.join(colored_tokens)
+
+
+GLAIVE_ROLES = ["USER", "ASSISTANT", "FUNCTION RESPONSE"]
+GLAIVE_TO_SHAREGPT_ROLE = {
+ "SYSTEM": "system",
+ "USER": "human",
+ "ASSISTANT": "gpt",
+ "FUNCTION RESPONSE": "tool",
+}
+
+GLAIVE_MSG_REGEX = re.compile(rf"({'|'.join(GLAIVE_ROLES)}): ")
+
+
+def chatml_to_conversation(row: Dict[str, str]) -> List[Dict[str, str]]:
+ """
+ Converts a ChatML formatted row to a list of messages in ShareGPT format.
+ Initially based off https://github.com/lilacai/lilac/blob/main/notebooks/GlaiveToShareGPT.ipynb.
+ """
+
+ system_prompt = row.get("system")
+ if system_prompt:
+ system_prompt = system_prompt.removeprefix("SYSTEM: ")
+
+ chat_str = row["chat"]
+ chat_msgs = [s.strip() for s in GLAIVE_MSG_REGEX.split(chat_str) if s]
+
+ chat_msg_dicts = [
+ {"from": GLAIVE_TO_SHAREGPT_ROLE[role], "value": value}
+ for role, value in zip(chat_msgs[::2], chat_msgs[1::2])
+ ]
+
+ if system_prompt:
+ chat_msg_dicts = [
+ {"from": GLAIVE_TO_SHAREGPT_ROLE["SYSTEM"], "value": system_prompt}
+ ] + chat_msg_dicts
+
+ return chat_msg_dicts
+
+
+def merge_consecutive_messages(messages):
+ """
+ Merge consecutive messages from the same sender into a single message.
+ This can be useful with datasets that contain multiple consecutive tool calls.
+ """
+
+ merged_messages = []
+ current_from = None
+ current_message = ""
+
+ for msg in messages:
+ if current_from == msg["from"]:
+ current_message += msg["value"]
+ else:
+ if current_from is not None:
+ merged_messages.append({"from": current_from, "value": current_message})
+ current_from = msg["from"]
+ current_message = msg["value"]
+
+ if current_from is not None:
+ merged_messages.append({"from": current_from, "value": current_message})
+
+ return merged_messages
diff --git a/src/axolotl/utils/trainer.py b/src/axolotl/utils/trainer.py
new file mode 100644
index 0000000000000000000000000000000000000000..fdf86e5672becd64386ec8b79a7a66bf76857f93
--- /dev/null
+++ b/src/axolotl/utils/trainer.py
@@ -0,0 +1,441 @@
+"""Module containing the Trainer class and related functions"""
+import math
+import os
+import random
+from contextlib import contextmanager
+from functools import partial
+from typing import List, Optional
+
+import numpy as np
+import torch
+import torch.cuda
+from accelerate.logging import get_logger
+from datasets import set_caching_enabled
+from torch.utils.data import DataLoader, RandomSampler
+from transformers.utils import is_torch_bf16_gpu_available
+
+from axolotl.core.trainer_builder import HFCausalTrainerBuilder, HFRLTrainerBuilder
+from axolotl.utils.distributed import is_main_process, reduce_and_broadcast, zero_first
+from axolotl.utils.samplers import MultipackBatchSampler, get_dataset_lengths
+
+LOG = get_logger("axolotl")
+
+
+@torch.jit.script
+def weighted_cross_entropy(
+ logits: torch.Tensor, labels: torch.Tensor, weights: torch.Tensor
+):
+ # Flatten the logits, labels, and weights tensors
+ logits = logits.view(
+ -1, logits.size(-1)
+ ) # logits becomes of shape [batch_size*sequence_length, vocab_size]
+ labels = labels.view(-1) # labels becomes of shape [batch_size*sequence_length]
+ weights = weights.view(-1) # weights becomes of shape [batch_size*sequence_length]
+
+ # Compute the unweighted cross entropy loss
+ losses = torch.nn.functional.cross_entropy(logits, labels, reduction="none")
+
+ # Apply the weights to the losses and compute their sum
+ return (weights * losses).sum()
+
+
+@torch.jit.script
+def create_weighted_mask(labels: torch.Tensor):
+ # Check if the tensor is 2D. If not, unsqueeze it to make it 2D
+ if len(labels.shape) == 1:
+ labels = labels.unsqueeze(0)
+
+ weights = torch.zeros_like(labels).float()
+ for i in range(labels.shape[0]):
+ mask = labels[i] != -100
+
+ # Create a tensor to track group ids
+ group_ids = torch.zeros_like(labels[i]).int()
+ curr_group_id = 0
+
+ for j in range(1, len(labels[i])):
+ if mask[j] and not mask[j - 1]: # switch from masked to unmasked label
+ curr_group_id += 1 # start new group
+ group_ids[j] = (
+ curr_group_id if mask[j] else 0
+ ) # assign group id if unmasked label
+
+ # Count only unmasked labels in each group
+ group_counts = torch.bincount(group_ids[mask])
+
+ mask_weights = torch.zeros_like(labels[i]).float()
+ mask_weights[mask] = 1.0 / group_counts[group_ids[mask]]
+
+ weights[i] = mask_weights
+
+ return weights.squeeze() # squeeze the output to match the input dimension
+
+
+def trainer_weighted_loss(model_output, labels, shift_labels=True):
+ logits = (
+ model_output["logits"] if isinstance(model_output, dict) else model_output[0]
+ )
+ if shift_labels:
+ logits = logits[..., :-1, :].contiguous()
+ labels = labels[..., 1:].contiguous()
+
+ weights = create_weighted_mask(labels)
+ return weighted_cross_entropy(logits, labels, weights)
+
+
+@contextmanager
+def disable_datasets_caching():
+ try:
+ set_caching_enabled(False)
+ yield
+ finally:
+ set_caching_enabled(True)
+
+
+def add_position_ids(sample):
+ sample_len = len(sample["input_ids"])
+ sample["position_ids"] = torch.arange(len(sample["input_ids"]))
+ sample["length"] = sample_len
+ return sample
+
+
+def add_pose_position_ids(
+ sample,
+ max_context_len=32768,
+ split_on_token_ids: Optional[List[int]] = None,
+ chunks: int = 2,
+):
+ """
+ use the PoSE technique to extend the context length by randomly skipping
+ positions in the context. We only want to skip right before tokens in
+ the split_on_token_ids list. We should attempt to randomly distribute
+ the skips, but we don't need the final position_ids to be the full
+ context_len. There may be multiple turns in the context, so we want to
+ make sure we take into account the maximum possible number of skips
+ remaining in each sample.
+ """
+
+ input_ids = sample["input_ids"]
+ sample_len = len(input_ids)
+ max_skips = max_context_len - sample_len
+
+ if split_on_token_ids is None:
+ split_on_token_ids = []
+
+ if split_on_token_ids:
+ split_indices = [
+ i for i, token_id in enumerate(input_ids) if token_id in split_on_token_ids
+ ]
+ else:
+ chunk_len = sample_len // chunks
+ split_indices = [i * chunk_len for i in range(1, chunks)]
+ split_indices.append(len(input_ids)) # make sure we go to the end of the sample
+ if split_indices[0] < 2:
+ # drop the first split index if it's too close to the beginning
+ split_indices = split_indices[1:]
+
+ position_ids = []
+ prev_index = 0
+ total_skips = 0
+
+ for split_index in split_indices:
+ num_skips = (
+ random.randint(0, max_skips) # nosec B311
+ if prev_index != 0 and max_skips
+ else 0
+ )
+ max_skips -= num_skips
+ total_skips += num_skips
+
+ segment_position_ids = list(
+ range(prev_index + total_skips, split_index + total_skips)
+ )
+
+ position_ids.extend(segment_position_ids)
+ prev_index = split_index
+
+ sample["sequence_len"] = position_ids[-1]
+ position_ids = torch.tensor(position_ids)
+
+ sample["position_ids"] = position_ids
+ sample["length"] = len(position_ids)
+ assert len(position_ids) == len(input_ids)
+
+ return sample
+
+
+def add_length(sample):
+ sample["length"] = len(sample["input_ids"])
+ return sample
+
+
+def drop_long_seq(sample, sequence_len=2048, min_sequence_len=2):
+ return (
+ len(sample["input_ids"]) <= sequence_len
+ and len(sample["input_ids"]) >= min_sequence_len
+ )
+
+
+def process_datasets_for_packing(cfg, train_dataset, eval_dataset):
+ drop_long = partial(
+ drop_long_seq,
+ sequence_len=cfg.sequence_len,
+ min_sequence_len=cfg.min_sample_len or 2,
+ )
+ with zero_first(is_main_process()):
+ if cfg.is_preprocess:
+ min_input_len = np.min(get_dataset_lengths(train_dataset))
+ LOG.debug(f"min_input_len: {min_input_len}", main_process_only=True)
+ max_input_len = np.max(get_dataset_lengths(train_dataset))
+ LOG.debug(f"max_input_len: {max_input_len}", main_process_only=True)
+
+ if (
+ cfg.is_mistral_derived_model and cfg.flash_attention
+ ) or cfg.model_config_type == "mamba":
+ LOG.info("dropping attention_mask column")
+ train_dataset = train_dataset.remove_columns("attention_mask")
+ if eval_dataset:
+ eval_dataset = eval_dataset.remove_columns("attention_mask")
+
+ if cfg.model_config_type == "falcon":
+ LOG.info("dropping token_type_ids column if it exists")
+ if "token_type_ids" in train_dataset.column_names:
+ train_dataset = train_dataset.remove_columns("token_type_ids")
+ if eval_dataset and "token_type_ids" in eval_dataset.column_names:
+ eval_dataset = eval_dataset.remove_columns("token_type_ids")
+
+ train_dataset = train_dataset.filter(
+ drop_long,
+ num_proc=cfg.dataset_processes,
+ load_from_cache_file=not cfg.is_preprocess,
+ desc="Dropping Long Sequences",
+ )
+ if eval_dataset:
+ eval_dataset = eval_dataset.filter(
+ drop_long,
+ num_proc=cfg.dataset_processes,
+ load_from_cache_file=not cfg.is_preprocess,
+ desc="Dropping Long Sequences",
+ )
+
+ if cfg.group_by_length:
+ train_dataset = train_dataset.map(
+ add_length,
+ num_proc=cfg.dataset_processes,
+ load_from_cache_file=not cfg.is_preprocess,
+ desc="Group By Length",
+ )
+
+ if cfg.use_pose:
+ pose_kwargs = {}
+ if cfg.pose_num_chunks is not None:
+ pose_kwargs["chunks"] = cfg.pose_num_chunks
+ pose_fn = partial(
+ add_pose_position_ids,
+ max_context_len=cfg.pose_max_context_len,
+ split_on_token_ids=cfg.pose_split_on_token_ids,
+ **pose_kwargs,
+ )
+ train_dataset = train_dataset.map(
+ pose_fn,
+ num_proc=cfg.dataset_processes,
+ load_from_cache_file=not cfg.is_preprocess,
+ desc="Add position_id column (PoSE)",
+ )
+ train_dataset = train_dataset.sort("sequence_len")
+ if cfg.eval_sample_packing is not False:
+ if eval_dataset:
+ eval_dataset = eval_dataset.map(
+ pose_fn,
+ num_proc=cfg.dataset_processes,
+ load_from_cache_file=not cfg.is_preprocess,
+ desc="Add position_id column (PoSE)",
+ )
+ elif cfg.sample_packing:
+ train_dataset = train_dataset.map(
+ add_position_ids,
+ num_proc=cfg.dataset_processes,
+ load_from_cache_file=not cfg.is_preprocess,
+ desc="Add position_id column (Sample Packing)",
+ )
+ if cfg.eval_sample_packing is not False:
+ if eval_dataset:
+ eval_dataset = eval_dataset.map(
+ add_position_ids,
+ num_proc=cfg.dataset_processes,
+ load_from_cache_file=not cfg.is_preprocess,
+ desc="Add position_id column (Sample Packing)",
+ )
+
+ return train_dataset, eval_dataset
+
+
+def process_pretraining_datasets_for_packing(
+ train_dataset, sequence_len, skip_position_ids=True
+):
+ drop_long = partial(drop_long_seq, sequence_len=sequence_len)
+
+ train_dataset = train_dataset.filter(
+ drop_long,
+ desc="Dropping Long Sequences",
+ )
+ if skip_position_ids:
+ train_dataset = train_dataset.map(
+ add_position_ids,
+ desc="Add position_id column (Pretraining Sample Packing)",
+ )
+
+ return train_dataset
+
+
+def calculate_total_num_steps(cfg, train_dataset, update=True):
+ if not cfg.total_num_tokens:
+ total_num_tokens = np.sum(
+ train_dataset.data.column("input_ids")
+ .to_pandas()
+ .apply(lambda x: len(x)) # pylint: disable=unnecessary-lambda
+ .values
+ )
+ LOG.debug(f"total_num_tokens: {total_num_tokens:_}", main_process_only=True)
+ if update:
+ cfg.total_num_tokens = total_num_tokens
+
+ skip_estimates = cfg.model_config_type == "mamba"
+
+ if not skip_estimates and not cfg.total_supervised_tokens:
+ total_supervised_tokens = (
+ train_dataset.data.column("labels")
+ .to_pandas()
+ .apply(lambda x: np.sum(np.array(x) != -100))
+ .sum()
+ )
+ LOG.debug(
+ f"`total_supervised_tokens: {total_supervised_tokens:_}`",
+ main_process_only=True,
+ )
+ if update:
+ cfg.total_supervised_tokens = total_supervised_tokens
+
+ if not skip_estimates and cfg.sample_packing:
+ # we have to drop anything longer then sequence len otherwise
+ # flash attention with position ids fails
+
+ if cfg.sample_packing_eff_est:
+ total_num_steps = (
+ # match count to len est in dataloader
+ (
+ math.floor(
+ 0.99
+ * cfg.total_num_tokens
+ / cfg.sample_packing_eff_est
+ / cfg.sequence_len
+ // cfg.batch_size
+ )
+ - 1
+ )
+ * cfg.num_epochs
+ )
+ LOG.debug(
+ f"total_num_tokens: {cfg.total_num_tokens:_}, total_num_steps: {total_num_steps:_}",
+ main_process_only=True,
+ )
+ else:
+ if cfg.flash_attention:
+ batch_size = 1
+ batch_max_len = cfg.micro_batch_size * cfg.sequence_len
+ else:
+ batch_size = cfg.micro_batch_size
+ batch_max_len = cfg.sequence_len
+ sampler = MultipackBatchSampler(
+ sampler=RandomSampler(train_dataset),
+ batch_size=batch_size,
+ drop_last=True,
+ batch_max_len=batch_max_len,
+ lengths=get_dataset_lengths(train_dataset),
+ )
+
+ data_loader = DataLoader(
+ train_dataset.remove_columns(["length"]),
+ batch_sampler=sampler,
+ )
+ data_loader_len = len(data_loader) // (
+ cfg.world_size * cfg.gradient_accumulation_steps
+ )
+ actual_eff = sampler.efficiency()
+ LOG.debug(f"data_loader_len: {data_loader_len}", main_process_only=True)
+ # FIXME: is there a bug here somewhere? the total num steps depends
+ # on the agreed on value for sample_packing_eff_est
+ total_num_steps = int(math.floor(data_loader_len * cfg.num_epochs))
+
+ def calc_sample_packing_eff_est(estimates: List[float]):
+ LOG.info(f"sample_packing_eff_est across ranks: {repr(estimates)}")
+ return max(estimates)
+
+ sample_packing_actual_eff_all = reduce_and_broadcast(
+ lambda: actual_eff,
+ calc_sample_packing_eff_est,
+ )
+ sample_packing_eff_est = (
+ math.ceil(sample_packing_actual_eff_all * 100.0) / 100.0
+ )
+ if update:
+ cfg.sample_packing_eff_est = sample_packing_eff_est
+ LOG.debug(
+ f"sample_packing_eff_est: {cfg.sample_packing_eff_est}",
+ main_process_only=True,
+ )
+ else:
+ total_num_steps = int(
+ math.ceil(len(train_dataset) * cfg.num_epochs / cfg.batch_size)
+ )
+ LOG.debug(f"total_num_steps: {total_num_steps}", main_process_only=True)
+ return total_num_steps
+
+
+def setup_fsdp_envs(cfg):
+ os.environ["ACCELERATE_USE_FSDP"] = "true"
+ if cfg.fsdp_config.fsdp_activation_checkpointing:
+ os.environ["FSDP_ACTIVATION_CHECKPOINTING"] = "true"
+ if cfg.fsdp_config.fsdp_offload_params:
+ os.environ["FSDP_OFFLOAD_PARAMS"] = "true"
+ if cfg.fsdp_config.fsdp_sync_module_states:
+ os.environ["FSDP_SYNC_MODULE_STATES"] = "true"
+ if cfg.fsdp_config.fsdp_cpu_ram_efficient_loading:
+ os.environ["FSDP_CPU_RAM_EFFICIENT_LOADING"] = "true"
+ if cfg.fsdp_config.fsdp_use_orig_params:
+ os.environ["FSDP_USE_ORIG_PARAMS"] = "true"
+ if cfg.fsdp_config.fsdp_state_dict_type:
+ os.environ["FSDP_STATE_DICT_TYPE"] = cfg.fsdp_config.fsdp_state_dict_type
+ if cfg.fsdp_config.fsdp_auto_wrap_policy:
+ os.environ["FSDP_AUTO_WRAP_POLICY"] = cfg.fsdp_config.fsdp_auto_wrap_policy
+ if cfg.fsdp_config.fsdp_transformer_layer_cls_to_wrap:
+ os.environ[
+ "FSDP_TRANSFORMER_CLS_TO_WRAP"
+ ] = cfg.fsdp_config.fsdp_transformer_layer_cls_to_wrap
+
+
+def prepare_optim_env(cfg):
+ if cfg.fsdp:
+ setup_fsdp_envs(cfg)
+ elif cfg.deepspeed:
+ os.environ["ACCELERATE_USE_DEEPSPEED"] = "true"
+ os.environ["ACCELERATE_DEEPSPEED_CONFIG_FILE"] = cfg.deepspeed
+
+ if (cfg.bf16 == "auto" and is_torch_bf16_gpu_available()) or cfg.bf16 is True:
+ os.environ["ACCELERATE_MIXED_PRECISION"] = "bf16"
+ elif cfg.fp16:
+ os.environ["ACCELERATE_MIXED_PRECISION"] = "fp16"
+
+
+def setup_trainer(cfg, train_dataset, eval_dataset, model, tokenizer, total_num_steps):
+ if cfg.rl in ["dpo", "ipo", "kto_pair", "orpo"]:
+ trainer_builder = HFRLTrainerBuilder(cfg, model[0], tokenizer)
+ trainer_builder.model_ref = model[1]
+ trainer_builder.peft_config = model[2]
+ else:
+ trainer_builder = HFCausalTrainerBuilder(cfg, model[0], tokenizer)
+
+ trainer_builder.train_dataset = train_dataset
+ trainer_builder.eval_dataset = eval_dataset
+
+ return trainer_builder.build(total_num_steps)
diff --git a/src/axolotl/utils/wandb_.py b/src/axolotl/utils/wandb_.py
new file mode 100644
index 0000000000000000000000000000000000000000..327dd9b6348159942de6d991cd7fd1730d2f5792
--- /dev/null
+++ b/src/axolotl/utils/wandb_.py
@@ -0,0 +1,21 @@
+"""Module for wandb utilities"""
+
+import os
+
+from axolotl.utils.dict import DictDefault
+
+
+def setup_wandb_env_vars(cfg: DictDefault):
+ for key in cfg.keys():
+ if key.startswith("wandb_"):
+ value = cfg.get(key, "")
+
+ if value and isinstance(value, str) and len(value) > 0:
+ os.environ[key.upper()] = value
+
+ # Enable wandb if project name is present
+ if cfg.wandb_project and len(cfg.wandb_project) > 0:
+ cfg.use_wandb = True
+ os.environ.pop("WANDB_DISABLED", None) # Remove if present
+ else:
+ os.environ["WANDB_DISABLED"] = "true"
diff --git a/styles.css b/styles.css
new file mode 100644
index 0000000000000000000000000000000000000000..2ddf50c7b4236e4b67c3e9fc369f6a7a562cd27d
--- /dev/null
+++ b/styles.css
@@ -0,0 +1 @@
+/* css styles */
diff --git a/tests/core/test_trainer_builder.py b/tests/core/test_trainer_builder.py
new file mode 100644
index 0000000000000000000000000000000000000000..82455922ef7dfce62b44f0f54d82c2c5b8763f0e
--- /dev/null
+++ b/tests/core/test_trainer_builder.py
@@ -0,0 +1,66 @@
+"""
+unit tests for axolotl.core.trainer_builder
+"""
+
+import pytest
+
+from axolotl.core.trainer_builder import HFRLTrainerBuilder
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.models import load_model, load_tokenizer
+
+
+@pytest.fixture(name="cfg")
+def fixture_cfg():
+ cfg = DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "model_type": "AutoModelForCausalLM",
+ "tokenizer_type": "LlamaTokenizer",
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "learning_rate": 0.00005,
+ "save_steps": 100,
+ "output_dir": "./model-out",
+ "warmup_steps": 10,
+ "gradient_checkpointing": False,
+ "optimizer": "adamw_torch",
+ "sequence_len": 2048,
+ "rl": True,
+ "adam_beta1": 0.998,
+ "adam_beta2": 0.9,
+ "adam_epsilon": 0.00001,
+ "dataloader_num_workers": 1,
+ "dataloader_pin_memory": True,
+ "model_config_type": "llama",
+ }
+ )
+
+ normalize_config(cfg)
+
+ return cfg
+
+
+@pytest.fixture(name="tokenizer")
+def fixture_tokenizer(cfg):
+ return load_tokenizer(cfg)
+
+
+@pytest.fixture(name="model")
+def fixture_model(cfg, tokenizer):
+ return load_model(cfg, tokenizer)
+
+
+class TestHFRLTrainerBuilder:
+ """
+ TestCase class for DPO trainer builder
+ """
+
+ def test_build_training_arguments(self, cfg, model, tokenizer):
+ builder = HFRLTrainerBuilder(cfg, model, tokenizer)
+ training_arguments = builder.build_training_arguments(100)
+ assert training_arguments.adam_beta1 == 0.998
+ assert training_arguments.adam_beta2 == 0.9
+ assert training_arguments.adam_epsilon == 0.00001
+ assert training_arguments.dataloader_num_workers == 1
+ assert training_arguments.dataloader_pin_memory is True
diff --git a/tests/e2e/.gitignore b/tests/e2e/.gitignore
new file mode 100644
index 0000000000000000000000000000000000000000..ad1727ec598318511c9b0620ec5a85e4a243243c
--- /dev/null
+++ b/tests/e2e/.gitignore
@@ -0,0 +1 @@
+last_run_prepared
diff --git a/tests/e2e/__init__.py b/tests/e2e/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/tests/e2e/patched/__init__.py b/tests/e2e/patched/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/tests/e2e/patched/test_4d_multipack_llama.py b/tests/e2e/patched/test_4d_multipack_llama.py
new file mode 100644
index 0000000000000000000000000000000000000000..d74d097237f279da40e8615a02dfc6ba4adcd733
--- /dev/null
+++ b/tests/e2e/patched/test_4d_multipack_llama.py
@@ -0,0 +1,114 @@
+"""
+E2E tests for multipack fft llama using 4d attention masks
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from ..utils import require_torch_2_1_1, with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class Test4dMultipackLlama(unittest.TestCase):
+ """
+ Test case for Llama models using 4d attention with multipack
+ """
+
+ @require_torch_2_1_1
+ @with_temp_dir
+ def test_sdp_lora_packing(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "flash_attention": False,
+ "sdp_attention": True,
+ "sample_packing": True,
+ "pad_to_sequence_len": True,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 16,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "sequence_len": 1024,
+ "val_set_size": 0.1,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "fp16": True,
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_torch_lora_packing(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "flash_attention": False,
+ "sdp_attention": False,
+ "sample_packing": True,
+ "pad_to_sequence_len": True,
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 16,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "fp16": True,
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
diff --git a/tests/e2e/patched/test_falcon_samplepack.py b/tests/e2e/patched/test_falcon_samplepack.py
new file mode 100644
index 0000000000000000000000000000000000000000..ae6a4973918618cd66a5f4c78386dced05585388
--- /dev/null
+++ b/tests/e2e/patched/test_falcon_samplepack.py
@@ -0,0 +1,112 @@
+"""
+E2E tests for falcon
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from ..utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestFalconPatched(unittest.TestCase):
+ """
+ Test case for Falcon models
+ """
+
+ @with_temp_dir
+ def test_qlora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "illuin/tiny-random-FalconForCausalLM",
+ "flash_attention": True,
+ "sample_packing": True,
+ "sequence_len": 2048,
+ "load_in_4bit": True,
+ "adapter": "qlora",
+ "lora_r": 16,
+ "lora_alpha": 32,
+ "lora_dropout": 0.1,
+ "lora_target_linear": True,
+ "lora_modules_to_save": ["word_embeddings", "lm_head"],
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "bos_token": "<|endoftext|>",
+ "pad_token": "<|endoftext|>",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_ft(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "illuin/tiny-random-FalconForCausalLM",
+ "flash_attention": True,
+ "sample_packing": True,
+ "sequence_len": 2048,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "bos_token": "<|endoftext|>",
+ "pad_token": "<|endoftext|>",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
diff --git a/tests/e2e/patched/test_fused_llama.py b/tests/e2e/patched/test_fused_llama.py
new file mode 100644
index 0000000000000000000000000000000000000000..de1195c368e634e344c1df631b6d0e7df22c2a57
--- /dev/null
+++ b/tests/e2e/patched/test_fused_llama.py
@@ -0,0 +1,74 @@
+"""
+E2E tests for lora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from transformers.utils import is_torch_bf16_gpu_available
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from ..utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestFusedLlama(unittest.TestCase):
+ """
+ Test case for Llama models using Fused layers
+ """
+
+ @with_temp_dir
+ def test_fft_packing(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "flash_attention": True,
+ "pad_to_sequence_len": True,
+ "flash_attn_fuse_qkv": True,
+ "flash_attn_fuse_mlp": True,
+ "sample_packing": True,
+ "sequence_len": 1024,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "unk_token": "",
+ "bos_token": "",
+ "eos_token": " ",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 10,
+ "save_steps": 5,
+ "eval_steps": 5,
+ }
+ )
+ if is_torch_bf16_gpu_available():
+ cfg.bf16 = True
+ else:
+ cfg.fp16 = True
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
diff --git a/tests/e2e/patched/test_llama_s2_attention.py b/tests/e2e/patched/test_llama_s2_attention.py
new file mode 100644
index 0000000000000000000000000000000000000000..f1d37eb3ca879d351a0562fa272e95888239dc76
--- /dev/null
+++ b/tests/e2e/patched/test_llama_s2_attention.py
@@ -0,0 +1,111 @@
+"""
+E2E tests for llama w/ S2 attn
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from ..utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestLlamaShiftedSparseAttention(unittest.TestCase):
+ """
+ Test case for Llama models using S2 Attn
+ """
+
+ @with_temp_dir
+ def test_lora_s2_attn(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 16384,
+ "sample_packing": False,
+ "flash_attention": True,
+ "s2_attention": True,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 16,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "Yukang/LongAlpaca-12k",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 10,
+ "save_steps": 5,
+ "eval_steps": 5,
+ "bf16": "auto",
+ }
+ )
+
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_fft_s2_attn(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 16384,
+ "sample_packing": False,
+ "flash_attention": True,
+ "s2_attention": True,
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "Yukang/LongAlpaca-12k",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 10,
+ "save_steps": 5,
+ "eval_steps": 5,
+ "bf16": "auto",
+ }
+ )
+
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
diff --git a/tests/e2e/patched/test_lora_llama_multipack.py b/tests/e2e/patched/test_lora_llama_multipack.py
new file mode 100644
index 0000000000000000000000000000000000000000..f251f9b66165884e6b3da4e6fa12659e04c8d1c8
--- /dev/null
+++ b/tests/e2e/patched/test_lora_llama_multipack.py
@@ -0,0 +1,126 @@
+"""
+E2E tests for lora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+import pytest
+from transformers.utils import is_auto_gptq_available, is_torch_bf16_gpu_available
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from ..utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestLoraLlama(unittest.TestCase):
+ """
+ Test case for Llama models using LoRA w multipack
+ """
+
+ @with_temp_dir
+ def test_lora_packing(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 1024,
+ "sample_packing": True,
+ "flash_attention": True,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 64,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.2,
+ "special_tokens": {
+ "unk_token": "",
+ "bos_token": "",
+ "eos_token": " ",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 8,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ }
+ )
+ if is_torch_bf16_gpu_available():
+ cfg.bf16 = True
+ else:
+ cfg.fp16 = True
+
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @pytest.mark.skipif(not is_auto_gptq_available(), reason="auto-gptq not available")
+ @with_temp_dir
+ def test_lora_gptq_packed(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "TheBlokeAI/jackfram_llama-68m-GPTQ",
+ "model_type": "AutoModelForCausalLM",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 1024,
+ "sample_packing": True,
+ "flash_attention": True,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "gptq": True,
+ "gptq_disable_exllama": True,
+ "lora_r": 32,
+ "lora_alpha": 64,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "unk_token": "",
+ "bos_token": "",
+ "eos_token": " ",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "save_steps": 0.5,
+ "micro_batch_size": 8,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
diff --git a/tests/e2e/patched/test_mistral_samplepack.py b/tests/e2e/patched/test_mistral_samplepack.py
new file mode 100644
index 0000000000000000000000000000000000000000..a56c530b219b19f2315c9868c4c19780eb27953b
--- /dev/null
+++ b/tests/e2e/patched/test_mistral_samplepack.py
@@ -0,0 +1,113 @@
+"""
+E2E tests for lora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from ..utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestMistral(unittest.TestCase):
+ """
+ Test case for Llama models using LoRA
+ """
+
+ @with_temp_dir
+ def test_lora_packing(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "openaccess-ai-collective/tiny-mistral",
+ "flash_attention": True,
+ "sample_packing": True,
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 64,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "unk_token": "",
+ "bos_token": "",
+ "eos_token": " ",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_ft_packing(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "openaccess-ai-collective/tiny-mistral",
+ "flash_attention": True,
+ "sample_packing": True,
+ "sequence_len": 1024,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "unk_token": "",
+ "bos_token": "",
+ "eos_token": " ",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
diff --git a/tests/e2e/patched/test_mixtral_samplepack.py b/tests/e2e/patched/test_mixtral_samplepack.py
new file mode 100644
index 0000000000000000000000000000000000000000..8baba030735cc7af17b186e5d0787a0d80c02e80
--- /dev/null
+++ b/tests/e2e/patched/test_mixtral_samplepack.py
@@ -0,0 +1,111 @@
+"""
+E2E tests for mixtral
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from ..utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestMixtral(unittest.TestCase):
+ """
+ Test case for Llama models using LoRA
+ """
+
+ @with_temp_dir
+ def test_qlora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "hf-internal-testing/Mixtral-tiny",
+ "tokenizer_config": "LoneStriker/Mixtral-8x7B-v0.1-HF",
+ "flash_attention": True,
+ "sample_packing": True,
+ "sequence_len": 2048,
+ "load_in_4bit": True,
+ "adapter": "qlora",
+ "lora_r": 16,
+ "lora_alpha": 32,
+ "lora_dropout": 0.1,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_ft(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "hf-internal-testing/Mixtral-tiny",
+ "tokenizer_config": "LoneStriker/Mixtral-8x7B-v0.1-HF",
+ "flash_attention": True,
+ "sample_packing": True,
+ "sequence_len": 2048,
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ model, _ = train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (
+ "MixtralFlashAttention2"
+ in model.model.layers[0].self_attn.__class__.__name__
+ )
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
diff --git a/tests/e2e/patched/test_model_patches.py b/tests/e2e/patched/test_model_patches.py
new file mode 100644
index 0000000000000000000000000000000000000000..eecd1b3c117594782ccf3b98e1cb69140e0b6d20
--- /dev/null
+++ b/tests/e2e/patched/test_model_patches.py
@@ -0,0 +1,95 @@
+"""
+E2E smoke tests to check that the monkeypatches are in place for certain configurations
+"""
+
+import unittest
+
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.models import load_model, load_tokenizer
+
+from ..utils import with_temp_dir
+
+
+class TestModelPatches(unittest.TestCase):
+ """
+ TestCases for the multipack monkey patches
+ """
+
+ @with_temp_dir
+ def test_mixtral_multipack(self, temp_dir):
+ cfg = DictDefault(
+ {
+ "base_model": "hf-internal-testing/Mixtral-tiny",
+ "tokenizer_config": "LoneStriker/Mixtral-8x7B-v0.1-HF",
+ "flash_attention": True,
+ "sample_packing": True,
+ "sequence_len": 2048,
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ tokenizer = load_tokenizer(cfg)
+ model, _ = load_model(cfg, tokenizer, inference=cli_args.inference)
+
+ assert (
+ "MixtralFlashAttention2"
+ in model.model.layers[0].self_attn.__class__.__name__
+ )
+
+ @with_temp_dir
+ def test_mistral_multipack(self, temp_dir):
+ cfg = DictDefault(
+ {
+ "base_model": "openaccess-ai-collective/tiny-mistral",
+ "flash_attention": True,
+ "sample_packing": True,
+ "sequence_len": 2048,
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ tokenizer = load_tokenizer(cfg)
+ model, _ = load_model(cfg, tokenizer, inference=cli_args.inference)
+
+ assert (
+ "axolotl.monkeypatch.mistral_attn_hijack_flash"
+ in model.model.layers[0].self_attn.forward.__module__
+ )
diff --git a/tests/e2e/patched/test_phi_multipack.py b/tests/e2e/patched/test_phi_multipack.py
new file mode 100644
index 0000000000000000000000000000000000000000..5f30453c18b816861c2d98b4349f5dbde0362860
--- /dev/null
+++ b/tests/e2e/patched/test_phi_multipack.py
@@ -0,0 +1,123 @@
+"""
+E2E tests for lora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from ..utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestPhiMultipack(unittest.TestCase):
+ """
+ Test case for Phi2 models
+ """
+
+ @with_temp_dir
+ def test_ft_packed(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "microsoft/phi-1_5",
+ "model_type": "PhiForCausalLM",
+ "tokenizer_type": "AutoTokenizer",
+ "sequence_len": 1024,
+ "sample_packing": True,
+ "flash_attention": True,
+ "pad_to_sequence_len": True,
+ "load_in_8bit": False,
+ "adapter": None,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "pad_token": "<|endoftext|>",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "dataset_shard_num": 10,
+ "dataset_shard_idx": 0,
+ "num_epochs": 1,
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "eval_steps": 10,
+ "save_steps": 10,
+ "bf16": "auto",
+ }
+ )
+
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
+
+ @with_temp_dir
+ def test_qlora_packed(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "microsoft/phi-1_5",
+ "model_type": "PhiForCausalLM",
+ "tokenizer_type": "AutoTokenizer",
+ "sequence_len": 1024,
+ "sample_packing": True,
+ "flash_attention": True,
+ "pad_to_sequence_len": True,
+ "load_in_8bit": False,
+ "adapter": "qlora",
+ "lora_r": 64,
+ "lora_alpha": 32,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "pad_token": "<|endoftext|>",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "dataset_shard_num": 10,
+ "dataset_shard_idx": 0,
+ "num_epochs": 1,
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "eval_steps": 10,
+ "save_steps": 10,
+ "bf16": "auto",
+ }
+ )
+
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
diff --git a/tests/e2e/patched/test_resume.py b/tests/e2e/patched/test_resume.py
new file mode 100644
index 0000000000000000000000000000000000000000..dfe9e86252d7bccfaf8cf98d7c0d9e762e9aa774
--- /dev/null
+++ b/tests/e2e/patched/test_resume.py
@@ -0,0 +1,95 @@
+"""
+E2E tests for resuming training
+"""
+
+import logging
+import os
+import re
+import subprocess
+import unittest
+from pathlib import Path
+
+from transformers.utils import is_torch_bf16_gpu_available
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from ..utils import most_recent_subdir, with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestResumeLlama(unittest.TestCase):
+ """
+ Test case for resuming training of llama models
+ """
+
+ @with_temp_dir
+ def test_resume_qlora_packed(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 1024,
+ "sample_packing": True,
+ "flash_attention": True,
+ "load_in_4bit": True,
+ "adapter": "qlora",
+ "lora_r": 32,
+ "lora_alpha": 64,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "vicgalle/alpaca-gpt4",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "save_steps": 10,
+ "save_total_limit": 5,
+ "max_steps": 40,
+ }
+ )
+ if is_torch_bf16_gpu_available():
+ cfg.bf16 = True
+ else:
+ cfg.fp16 = True
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+
+ resume_cfg = cfg | DictDefault(
+ {
+ "resume_from_checkpoint": f"{temp_dir}/checkpoint-30/",
+ }
+ )
+ normalize_config(resume_cfg)
+ cli_args = TrainerCliArgs()
+
+ train(cfg=resume_cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ tb_log_path_1 = most_recent_subdir(temp_dir + "/runs")
+ cmd = f"tensorboard --inspect --logdir {tb_log_path_1}"
+ res = subprocess.run(
+ cmd, shell=True, text=True, capture_output=True, check=True
+ )
+ pattern = r"first_step\s+(\d+)"
+ first_steps = int(re.findall(pattern, res.stdout)[0])
+ assert first_steps == 31
diff --git a/tests/e2e/test_dpo.py b/tests/e2e/test_dpo.py
new file mode 100644
index 0000000000000000000000000000000000000000..9596b1873f009a3a924bf9379d56d9142290f482
--- /dev/null
+++ b/tests/e2e/test_dpo.py
@@ -0,0 +1,207 @@
+"""
+E2E tests for lora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+import pytest
+
+from axolotl.cli import load_rl_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from .utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+@pytest.mark.skip(reason="doesn't seem to work on modal")
+class TestDPOLlamaLora(unittest.TestCase):
+ """
+ Test case for DPO Llama models using LoRA
+ """
+
+ @with_temp_dir
+ def test_dpo_lora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 64,
+ "lora_alpha": 32,
+ "lora_dropout": 0.1,
+ "lora_target_linear": True,
+ "special_tokens": {},
+ "rl": "dpo",
+ "datasets": [
+ {
+ "path": "Intel/orca_dpo_pairs",
+ "type": "chatml.intel",
+ "split": "train",
+ },
+ ],
+ "num_epochs": 1,
+ "micro_batch_size": 4,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "paged_adamw_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "warmup_steps": 5,
+ "gradient_checkpointing": True,
+ "gradient_checkpointing_kwargs": {"use_reentrant": True},
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_rl_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "checkpoint-20/adapter_model.safetensors").exists()
+
+ @with_temp_dir
+ def test_kto_pair_lora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 64,
+ "lora_alpha": 32,
+ "lora_dropout": 0.1,
+ "lora_target_linear": True,
+ "special_tokens": {},
+ "rl": "kto_pair",
+ "datasets": [
+ {
+ "path": "Intel/orca_dpo_pairs",
+ "type": "chatml.intel",
+ "split": "train",
+ },
+ ],
+ "num_epochs": 1,
+ "micro_batch_size": 4,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "paged_adamw_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "warmup_steps": 5,
+ "gradient_checkpointing": True,
+ "gradient_checkpointing_kwargs": {"use_reentrant": True},
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_rl_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "checkpoint-20/adapter_model.safetensors").exists()
+
+ @with_temp_dir
+ def test_ipo_lora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 64,
+ "lora_alpha": 32,
+ "lora_dropout": 0.1,
+ "lora_target_linear": True,
+ "special_tokens": {},
+ "rl": "ipo",
+ "datasets": [
+ {
+ "path": "Intel/orca_dpo_pairs",
+ "type": "chatml.intel",
+ "split": "train",
+ },
+ ],
+ "num_epochs": 1,
+ "micro_batch_size": 4,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "paged_adamw_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "warmup_steps": 5,
+ "gradient_checkpointing": True,
+ "gradient_checkpointing_kwargs": {"use_reentrant": True},
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_rl_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "checkpoint-20/adapter_model.safetensors").exists()
+
+ @with_temp_dir
+ def test_orpo_lora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 64,
+ "lora_alpha": 32,
+ "lora_dropout": 0.1,
+ "lora_target_linear": True,
+ "special_tokens": {},
+ "rl": "orpo",
+ "orpo_alpha": 0.1,
+ "remove_unused_columns": False,
+ "chat_template": "chatml",
+ "datasets": [
+ {
+ "path": "argilla/ultrafeedback-binarized-preferences-cleaned",
+ "type": "chat_template.argilla",
+ "split": "train",
+ },
+ ],
+ "num_epochs": 1,
+ "micro_batch_size": 4,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "paged_adamw_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "warmup_steps": 5,
+ "gradient_checkpointing": True,
+ "gradient_checkpointing_kwargs": {"use_reentrant": True},
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_rl_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "checkpoint-20/adapter_model.safetensors").exists()
diff --git a/tests/e2e/test_falcon.py b/tests/e2e/test_falcon.py
new file mode 100644
index 0000000000000000000000000000000000000000..c76699a7c86e373d41da46396b07267bdb6e6b3f
--- /dev/null
+++ b/tests/e2e/test_falcon.py
@@ -0,0 +1,166 @@
+"""
+E2E tests for falcon
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from .utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestFalcon(unittest.TestCase):
+ """
+ Test case for falcon
+ """
+
+ @with_temp_dir
+ def test_lora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "illuin/tiny-random-FalconForCausalLM",
+ "flash_attention": True,
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 64,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "lora_modules_to_save": [
+ "word_embeddings",
+ "lm_head",
+ ],
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "bos_token": "<|endoftext|>",
+ "pad_token": "<|endoftext|>",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_lora_added_vocab(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "illuin/tiny-random-FalconForCausalLM",
+ "flash_attention": True,
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 64,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "lora_modules_to_save": [
+ "word_embeddings",
+ "lm_head",
+ ],
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "bos_token": "<|endoftext|>",
+ "pad_token": "<|endoftext|>",
+ },
+ "tokens": [
+ "<|im_start|>",
+ "<|im_end|>",
+ ],
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_ft(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "illuin/tiny-random-FalconForCausalLM",
+ "flash_attention": True,
+ "sequence_len": 1024,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "bos_token": "<|endoftext|>",
+ "pad_token": "<|endoftext|>",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
diff --git a/tests/e2e/test_lora_llama.py b/tests/e2e/test_lora_llama.py
new file mode 100644
index 0000000000000000000000000000000000000000..c79652bef76c8682381cac7ad9fe9eceee9464e3
--- /dev/null
+++ b/tests/e2e/test_lora_llama.py
@@ -0,0 +1,67 @@
+"""
+E2E tests for lora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from .utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestLoraLlama(unittest.TestCase):
+ """
+ Test case for Llama models using LoRA
+ """
+
+ @with_temp_dir
+ def test_lora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 64,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "unk_token": "",
+ "bos_token": "",
+ "eos_token": " ",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 8,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
diff --git a/tests/e2e/test_mamba.py b/tests/e2e/test_mamba.py
new file mode 100644
index 0000000000000000000000000000000000000000..8755fa4d51fcf45c678d6f84e94e0b82a8367b65
--- /dev/null
+++ b/tests/e2e/test_mamba.py
@@ -0,0 +1,68 @@
+"""
+E2E tests for lora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+import pytest
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from .utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+@pytest.mark.skip(reason="skipping until upstreamed into transformers")
+class TestMamba(unittest.TestCase):
+ """
+ Test case for Mamba models
+ """
+
+ @with_temp_dir
+ def test_fft(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "state-spaces/mamba-130m",
+ "model_type": "MambaLMHeadModel",
+ "tokenizer_type": "AutoTokenizer",
+ "tokenizer_config": "EleutherAI/gpt-neox-20b",
+ "flash_attention": False,
+ "sequence_len": 1024,
+ "load_in_8bit": False,
+ "val_set_size": 0.0,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "gradient_checkpointing": False,
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": None,
+ "save_safetensors": False,
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
diff --git a/tests/e2e/test_mistral.py b/tests/e2e/test_mistral.py
new file mode 100644
index 0000000000000000000000000000000000000000..57d85e51eb9137f3032757f7da8972e9b69f779e
--- /dev/null
+++ b/tests/e2e/test_mistral.py
@@ -0,0 +1,115 @@
+"""
+E2E tests for lora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from transformers.utils import is_torch_bf16_gpu_available
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from .utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestMistral(unittest.TestCase):
+ """
+ Test case for Llama models using LoRA
+ """
+
+ @with_temp_dir
+ def test_lora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "openaccess-ai-collective/tiny-mistral",
+ "flash_attention": True,
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 64,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "unk_token": "",
+ "bos_token": "",
+ "eos_token": " ",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_ft(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "openaccess-ai-collective/tiny-mistral",
+ "flash_attention": True,
+ "sequence_len": 1024,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "unk_token": "",
+ "bos_token": "",
+ "eos_token": " ",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ }
+ )
+ if is_torch_bf16_gpu_available():
+ cfg.bf16 = True
+ else:
+ cfg.fp16 = True
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
diff --git a/tests/e2e/test_mixtral.py b/tests/e2e/test_mixtral.py
new file mode 100644
index 0000000000000000000000000000000000000000..d4dad14ef24b959b8a4208ea27972879a0414059
--- /dev/null
+++ b/tests/e2e/test_mixtral.py
@@ -0,0 +1,290 @@
+"""
+E2E tests for mixtral
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+import torch
+from transformers.utils import is_torch_bf16_gpu_available
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from .utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestMixtral(unittest.TestCase):
+ """
+ Test case for Llama models using LoRA
+ """
+
+ @with_temp_dir
+ def test_qlora_w_fa2(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "hf-internal-testing/Mixtral-tiny",
+ "tokenizer_config": "LoneStriker/Mixtral-8x7B-v0.1-HF",
+ "flash_attention": True,
+ "sequence_len": 1024,
+ "load_in_4bit": True,
+ "adapter": "qlora",
+ "lora_r": 4,
+ "lora_alpha": 8,
+ "lora_dropout": 0.1,
+ "lora_target_modules": [
+ "o_proj",
+ "w3",
+ "k_proj",
+ "v_proj",
+ "w1",
+ "q_proj",
+ "w2",
+ ],
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ model, _ = train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (
+ model.base_model.model.model.layers[0].block_sparse_moe.gate.weight.dtype
+ == torch.float32
+ )
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_qlora_wo_fa2(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "hf-internal-testing/Mixtral-tiny",
+ "tokenizer_config": "LoneStriker/Mixtral-8x7B-v0.1-HF",
+ "flash_attention": False,
+ "sequence_len": 1024,
+ "load_in_4bit": True,
+ "adapter": "qlora",
+ "lora_r": 4,
+ "lora_alpha": 8,
+ "lora_dropout": 0.1,
+ "lora_target_modules": [
+ "o_proj",
+ "w3",
+ "k_proj",
+ "v_proj",
+ "w1",
+ "q_proj",
+ "w2",
+ ],
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ model, _ = train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (
+ model.base_model.model.model.layers[0].block_sparse_moe.gate.weight.dtype
+ == torch.float32
+ )
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_16bit_lora_w_fa2(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "hf-internal-testing/Mixtral-tiny",
+ "tokenizer_config": "LoneStriker/Mixtral-8x7B-v0.1-HF",
+ "flash_attention": True,
+ "sequence_len": 1024,
+ "adapter": "lora",
+ "lora_r": 4,
+ "lora_alpha": 8,
+ "lora_dropout": 0.1,
+ "lora_target_modules": [
+ "o_proj",
+ "w3",
+ "k_proj",
+ "v_proj",
+ "w1",
+ "q_proj",
+ "w2",
+ ],
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ }
+ )
+ if is_torch_bf16_gpu_available():
+ cfg.bf16 = True
+ else:
+ cfg.fp16 = True
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ model, _ = train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (
+ model.base_model.model.model.layers[0].block_sparse_moe.gate.weight.dtype
+ == torch.float32
+ )
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_16bit_lora_wo_fa2(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "hf-internal-testing/Mixtral-tiny",
+ "tokenizer_config": "LoneStriker/Mixtral-8x7B-v0.1-HF",
+ "flash_attention": False,
+ "sequence_len": 1024,
+ "adapter": "lora",
+ "lora_r": 4,
+ "lora_alpha": 8,
+ "lora_dropout": 0.1,
+ "lora_target_modules": [
+ "o_proj",
+ "w3",
+ "k_proj",
+ "v_proj",
+ "w1",
+ "q_proj",
+ "w2",
+ ],
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ }
+ )
+ normalize_config(cfg)
+ if is_torch_bf16_gpu_available():
+ cfg.bf16 = True
+ else:
+ cfg.fp16 = True
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ model, _ = train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (
+ model.base_model.model.model.layers[0].block_sparse_moe.gate.weight.dtype
+ == torch.float32
+ )
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
+
+ @with_temp_dir
+ def test_ft(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "hf-internal-testing/Mixtral-tiny",
+ "tokenizer_config": "LoneStriker/Mixtral-8x7B-v0.1-HF",
+ "flash_attention": True,
+ "sequence_len": 1024,
+ "val_set_size": 0.1,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "num_epochs": 2,
+ "micro_batch_size": 2,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_bnb_8bit",
+ "lr_scheduler": "cosine",
+ "max_steps": 20,
+ "save_steps": 10,
+ "eval_steps": 10,
+ }
+ )
+ if is_torch_bf16_gpu_available():
+ cfg.bf16 = True
+ else:
+ cfg.fp16 = True
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
diff --git a/tests/e2e/test_phi.py b/tests/e2e/test_phi.py
new file mode 100644
index 0000000000000000000000000000000000000000..7abed8594538b101badc38832625b6567b0fa496
--- /dev/null
+++ b/tests/e2e/test_phi.py
@@ -0,0 +1,122 @@
+"""
+E2E tests for lora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+import pytest
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from .utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+@pytest.mark.skip(reason="doesn't seem to work on modal")
+class TestPhi(unittest.TestCase):
+ """
+ Test case for Phi2 models
+ """
+
+ @with_temp_dir
+ def test_phi_ft(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "microsoft/phi-1_5",
+ "model_type": "AutoModelForCausalLM",
+ "tokenizer_type": "AutoTokenizer",
+ "sequence_len": 2048,
+ "sample_packing": False,
+ "load_in_8bit": False,
+ "adapter": None,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "pad_token": "<|endoftext|>",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "dataset_shard_num": 10,
+ "dataset_shard_idx": 0,
+ "num_epochs": 1,
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "paged_adamw_8bit",
+ "lr_scheduler": "cosine",
+ "flash_attention": True,
+ "max_steps": 10,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "pytorch_model.bin").exists()
+
+ @with_temp_dir
+ def test_phi_qlora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "microsoft/phi-1_5",
+ "model_type": "AutoModelForCausalLM",
+ "tokenizer_type": "AutoTokenizer",
+ "sequence_len": 2048,
+ "sample_packing": False,
+ "load_in_8bit": False,
+ "adapter": "qlora",
+ "lora_r": 64,
+ "lora_alpha": 32,
+ "lora_dropout": 0.05,
+ "lora_target_linear": True,
+ "val_set_size": 0.1,
+ "special_tokens": {
+ "pad_token": "<|endoftext|>",
+ },
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "dataset_shard_num": 10,
+ "dataset_shard_idx": 0,
+ "num_epochs": 1,
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "paged_adamw_8bit",
+ "lr_scheduler": "cosine",
+ "flash_attention": True,
+ "max_steps": 10,
+ "save_steps": 10,
+ "eval_steps": 10,
+ "bf16": "auto",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "adapter_model.bin").exists()
diff --git a/tests/e2e/test_relora_llama.py b/tests/e2e/test_relora_llama.py
new file mode 100644
index 0000000000000000000000000000000000000000..4ba130c9dcae876770d63520c2220f2c4efa8180
--- /dev/null
+++ b/tests/e2e/test_relora_llama.py
@@ -0,0 +1,68 @@
+"""
+E2E tests for relora llama
+"""
+
+import logging
+import os
+import unittest
+from pathlib import Path
+
+from axolotl.cli import load_datasets
+from axolotl.common.cli import TrainerCliArgs
+from axolotl.train import train
+from axolotl.utils.config import normalize_config
+from axolotl.utils.dict import DictDefault
+
+from .utils import with_temp_dir
+
+LOG = logging.getLogger("axolotl.tests.e2e")
+os.environ["WANDB_DISABLED"] = "true"
+
+
+class TestReLoraLlama(unittest.TestCase):
+ """
+ Test case for Llama models using LoRA
+ """
+
+ @with_temp_dir
+ def test_relora(self, temp_dir):
+ # pylint: disable=duplicate-code
+ cfg = DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "sequence_len": 1024,
+ "load_in_8bit": True,
+ "adapter": "lora",
+ "lora_r": 32,
+ "lora_alpha": 16,
+ "lora_dropout": 0.05,
+ "lora_target_modules": ["q_proj", "v_proj"],
+ "relora_steps": 25,
+ "relora_warmup_steps": 5,
+ "relora_anneal_steps": 5,
+ "relora_cpu_offload": True,
+ "val_set_size": 0.0,
+ "special_tokens": {},
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ "warmup_steps": 15,
+ "num_epochs": 2,
+ "micro_batch_size": 4,
+ "gradient_accumulation_steps": 1,
+ "output_dir": temp_dir,
+ "learning_rate": 0.00001,
+ "optimizer": "adamw_torch",
+ "lr_scheduler": "cosine",
+ }
+ )
+ normalize_config(cfg)
+ cli_args = TrainerCliArgs()
+ dataset_meta = load_datasets(cfg=cfg, cli_args=cli_args)
+
+ train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ assert (Path(temp_dir) / "model.safetensors").exists()
diff --git a/tests/e2e/utils.py b/tests/e2e/utils.py
new file mode 100644
index 0000000000000000000000000000000000000000..837b4734fcaf3ace0f8720cf25175d636a87bc07
--- /dev/null
+++ b/tests/e2e/utils.py
@@ -0,0 +1,47 @@
+"""
+helper utils for tests
+"""
+import os
+import shutil
+import tempfile
+import unittest
+from functools import wraps
+from importlib.metadata import version
+from pathlib import Path
+
+
+def with_temp_dir(test_func):
+ @wraps(test_func)
+ def wrapper(*args, **kwargs):
+ # Create a temporary directory
+ temp_dir = tempfile.mkdtemp()
+ try:
+ # Pass the temporary directory to the test function
+ test_func(*args, temp_dir=temp_dir, **kwargs)
+ finally:
+ # Clean up the directory after the test
+ shutil.rmtree(temp_dir)
+
+ return wrapper
+
+
+def most_recent_subdir(path):
+ base_path = Path(path)
+ subdirectories = [d for d in base_path.iterdir() if d.is_dir()]
+ if not subdirectories:
+ return None
+ subdir = max(subdirectories, key=os.path.getctime)
+
+ return subdir
+
+
+def require_torch_2_1_1(test_case):
+ """
+ Decorator marking a test that requires torch >= 2.1.1
+ """
+
+ def is_min_2_1_1():
+ torch_version = version("torch")
+ return torch_version >= "2.1.1"
+
+ return unittest.skipUnless(is_min_2_1_1(), "test torch 2.1.1")(test_case)
diff --git a/tests/fixtures/alpaca/alpaca.json b/tests/fixtures/alpaca/alpaca.json
new file mode 100644
index 0000000000000000000000000000000000000000..912643d17fa230421cc64f1a1326a12b03ff4f4e
--- /dev/null
+++ b/tests/fixtures/alpaca/alpaca.json
@@ -0,0 +1,12 @@
+[
+ {
+ "instruction": "You will be given a series of words. Output these words in reverse order, with each word on its own line.",
+ "input": "Words: ['Hello', 'world'].",
+ "output": "['world', 'Hello']"
+ },
+ {
+ "instruction": "In this task, you're given a short description of an event. Your job is to order the steps involved in the event from first to last. Note that there may be multiple correct answers for each event.",
+ "input": "Description: A man walks into a bar and orders a drink. He pays for his drink and leaves the bar.",
+ "output": "1. The man walks into the bar.\n2. He orders a drink.\n3. He pays for his drink.\n4. He leaves the bar."
+ }
+]
diff --git a/tests/fixtures/conversation.json b/tests/fixtures/conversation.json
new file mode 100644
index 0000000000000000000000000000000000000000..bffb118a4196fcaa2e629c59c51e4c2c4e127c52
--- /dev/null
+++ b/tests/fixtures/conversation.json
@@ -0,0 +1,52 @@
+{
+ "conversations": [
+ {
+ "from": "human",
+ "value": "how could i implement a minesweeper algorithm that utilises algebraic topology to solve boards?"
+ },
+ {
+ "from": "gpt",
+ "value": "It's possible to use algebraic topology to solve Minesweeper, but it would likely be quite complex. Algebraic topology is a branch of mathematics that studies the properties of spaces that are preserved under continuous deformations, such as stretching and bending. In the context of Minesweeper, this could be used to analyze the structure of the game board and determine the locations of mines based on certain topological invariants. However, implementing this would require a deep understanding of both algebraic topology and the game of Minesweeper. It's not something that can be explained in a simple answer."
+ },
+ {
+ "from": "human",
+ "value": "could you give me the gist of how it could though?"
+ },
+ {
+ "from": "gpt",
+ "value": "Sure. The basic idea would be to use algebraic topology to analyze the structure of the Minesweeper game board and determine the locations of mines based on certain topological invariants. This could involve representing the game board as a topological space and using topological invariants such as homotopy groups and Betti numbers to infer information about the location of mines on the board. However, as I mentioned before, implementing this would require a deep understanding of both algebraic topology and the game of Minesweeper, so it's not something that can be explained in a simple answer."
+ },
+ {
+ "from": "human",
+ "value": "give me a specific example"
+ },
+ {
+ "from": "gpt",
+ "value": "As a specific example, suppose we have a Minesweeper game board with the following configuration:\n```\n 1 1 1 1 1\n 1 ? ? ? 1\n 1 ? ? ? 1\n 1 ? ? ? 1\n 1 1 1 1 1\n```\nIn this case, the game board can be represented as a topological space, where each square on the board corresponds to a point in the space. The mines on the board can be thought of as \"holes\" in the space, which are points that are not part of the space. Using algebraic topology, we can analyze the structure of this space and determine the locations of the mines based on certain topological invariants.\n\nFor instance, one way to do this is to compute the homotopy groups of the space. Homotopy groups are a way of describing the holes in a space, and they can be used to infer information about the locations of the mines on the Minesweeper board. In the case of the board above, we can compute the first homotopy group of the space, which describes the number and \"shape\" of the holes in the space. In this case, the first homotopy group would tell us that there are four holes in the space, arranged in a square shape.\n\nBased on this information, we can infer that the mines in the game are located in the four corners of the board, as shown below:\n```\n 1 1 1 1 1\n 1 M ? ? 1\n 1 ? ? ? 1\n 1 ? ? M 1\n 1 1 1 1 1\n```\nOf course, this is just one example of how algebraic topology could be used to solve Minesweeper. There are many other possible approaches, and the specific details of how to implement them would depend on the specific game board and the topological invariants being used. As I mentioned before, implementing a solution like this would require a deep understanding of both algebraic topology and the game of Minesweeper."
+ },
+ {
+ "from": "human",
+ "value": "assuming i was someone with a deep understanding of both, could you explain the same concept to me with that assumption?"
+ },
+ {
+ "from": "gpt",
+ "value": "Sure. Assuming you have a deep understanding of both algebraic topology and the game of Minesweeper, here is how you might use algebraic topology to solve a Minesweeper game board:\n\n1. Represent the Minesweeper game board as a topological space. Each square on the board corresponds to a point in the space, and the mines on the board can be thought of as \"holes\" in the space, which are points that are not part of the space.\n2. Use topological invariants to infer information about the locations of the mines on the board. There are many possible topological invariants that could be used for this, such as homotopy groups, Betti numbers, or persistent homology.\n3. Use the information inferred from the topological invariants to determine the locations of the mines on the board. For example, if the first homotopy group of the space indicates that there are four holes in the space, arranged in a square shape, then we can infer that the mines are located in the four corners of the board.\n4. Use this information to solve the Minesweeper game. Once you know the locations of the mines, you can use this information to avoid them and clear the board.\n\nOf course, this is just one possible approach to using algebraic topology to solve Minesweeper. There may be other ways to do it, depending on the specific game board and the topological invariants being used. However, this should give you a general idea of how algebraic topology can be used to solve Minesweeper."
+ },
+ {
+ "from": "human",
+ "value": "how would i do it with persistent homology or any of the other topological invariants?"
+ },
+ {
+ "from": "gpt",
+ "value": "To use persistent homology or any of the other topological invariants to solve a Minesweeper game board, you would follow the same basic steps as described above, but with some differences in the specific details.\n\nFor instance, here is how you might use persistent homology to solve a Minesweeper game board:\n\n1. Represent the Minesweeper game board as a topological space. Each square on the board corresponds to a point in the space, and the mines on the board can be thought of as \"holes\" in the space, which are points that are not part of the space.\n2. Compute the persistent homology of the space. Persistent homology is a way of describing the holes in a space over time, and it can be used to infer information about the locations of the mines on the Minesweeper board.\n3. Use the information from the persistent homology to determine the locations of the mines on the board. For example, if the persistent homology indicates that there are four holes in the space, arranged in a square shape, then we can infer that the mines are located in the four corners of the board.\n4. Use this information to solve the Minesweeper game. Once you know the locations of the mines, you can use this information to avoid them and clear the board.\n\nOf course, this is just one possible approach to using persistent homology to solve Minesweeper. There may be other ways to do it, depending on the specific game board and the specific details of the persistent homology computation. However, this should give you a general idea of how persistent homology can be used to solve Minesweeper."
+ },
+ {
+ "from": "human",
+ "value": "how could i implement this in python?"
+ },
+ {
+ "from": "gpt",
+ "value": "To implement a Minesweeper solver that uses algebraic topology in Python, you would need to do the following:\n\n1. Install a Python package for algebraic topology. There are several options available, such as Gudhi, Dionysus, or scikit-tda. These packages provide algorithms and data structures for computing topological invariants, such as homotopy groups, Betti numbers, or persistent homology.\n2. Represent the Minesweeper game board as a topological space. This could be done using a data structure provided by the algebraic topology package, such as a simplicial complex or a Vietoris-Rips complex. Each square on the board would correspond to a point in the space, and the mines on the board would be represented as \"holes\" in the space.\n3. Compute the topological invariants of the space. This could be done using algorithms provided by the algebraic topology package. For example, you could compute the persistent homology of the space using the persistent homology algorithm.\n4. Use the information from the topological invariants to determine the locations of the mines on the board. This could be done using a combination of mathematical reasoning and programming. For example, if the persistent homology indicates that there are four holes in the space, arranged in a square shape, then you could use a piece of code to infer that the mines are located in the four corners of the board.\n5. Use this information to solve the Minesweeper game. Once you know the locations of the mines, you could use a piece of code to automatically click on the safe squares and clear the board.\n\nOf course, this is just one possible approach to implementing a Minesweeper solver that uses algebraic topology in Python. There may be other ways to do it, depending on the specific details of the implementation. However, this should give you a general idea of how to get started with such a project."
+ }
+ ]
+}
diff --git a/tests/fixtures/conversation.missingturns.json b/tests/fixtures/conversation.missingturns.json
new file mode 100644
index 0000000000000000000000000000000000000000..b5b962d859bc94804885d2fbd6aca0910413b4c3
--- /dev/null
+++ b/tests/fixtures/conversation.missingturns.json
@@ -0,0 +1,52 @@
+{
+ "conversations": [
+ {
+ "from": "human",
+ "value": "how could i implement a minesweeper algorithm that utilises algebraic topology to solve boards?"
+ },
+ {
+ "from": "gpt",
+ "value": ""
+ },
+ {
+ "from": "human",
+ "value": "could you give me the gist of how it could though?"
+ },
+ {
+ "from": "gpt",
+ "value": "Sure. The basic idea would be to use algebraic topology to analyze the structure of the Minesweeper game board and determine the locations of mines based on certain topological invariants. This could involve representing the game board as a topological space and using topological invariants such as homotopy groups and Betti numbers to infer information about the location of mines on the board. However, as I mentioned before, implementing this would require a deep understanding of both algebraic topology and the game of Minesweeper, so it's not something that can be explained in a simple answer."
+ },
+ {
+ "from": "human",
+ "value": "give me a specific example"
+ },
+ {
+ "from": "gpt",
+ "value": "As a specific example, suppose we have a Minesweeper game board with the following configuration:\n```\n 1 1 1 1 1\n 1 ? ? ? 1\n 1 ? ? ? 1\n 1 ? ? ? 1\n 1 1 1 1 1\n```\nIn this case, the game board can be represented as a topological space, where each square on the board corresponds to a point in the space. The mines on the board can be thought of as \"holes\" in the space, which are points that are not part of the space. Using algebraic topology, we can analyze the structure of this space and determine the locations of the mines based on certain topological invariants.\n\nFor instance, one way to do this is to compute the homotopy groups of the space. Homotopy groups are a way of describing the holes in a space, and they can be used to infer information about the locations of the mines on the Minesweeper board. In the case of the board above, we can compute the first homotopy group of the space, which describes the number and \"shape\" of the holes in the space. In this case, the first homotopy group would tell us that there are four holes in the space, arranged in a square shape.\n\nBased on this information, we can infer that the mines in the game are located in the four corners of the board, as shown below:\n```\n 1 1 1 1 1\n 1 M ? ? 1\n 1 ? ? ? 1\n 1 ? ? M 1\n 1 1 1 1 1\n```\nOf course, this is just one example of how algebraic topology could be used to solve Minesweeper. There are many other possible approaches, and the specific details of how to implement them would depend on the specific game board and the topological invariants being used. As I mentioned before, implementing a solution like this would require a deep understanding of both algebraic topology and the game of Minesweeper."
+ },
+ {
+ "from": "human",
+ "value": "assuming i was someone with a deep understanding of both, could you explain the same concept to me with that assumption?"
+ },
+ {
+ "from": "gpt",
+ "value": "Sure. Assuming you have a deep understanding of both algebraic topology and the game of Minesweeper, here is how you might use algebraic topology to solve a Minesweeper game board:\n\n1. Represent the Minesweeper game board as a topological space. Each square on the board corresponds to a point in the space, and the mines on the board can be thought of as \"holes\" in the space, which are points that are not part of the space.\n2. Use topological invariants to infer information about the locations of the mines on the board. There are many possible topological invariants that could be used for this, such as homotopy groups, Betti numbers, or persistent homology.\n3. Use the information inferred from the topological invariants to determine the locations of the mines on the board. For example, if the first homotopy group of the space indicates that there are four holes in the space, arranged in a square shape, then we can infer that the mines are located in the four corners of the board.\n4. Use this information to solve the Minesweeper game. Once you know the locations of the mines, you can use this information to avoid them and clear the board.\n\nOf course, this is just one possible approach to using algebraic topology to solve Minesweeper. There may be other ways to do it, depending on the specific game board and the topological invariants being used. However, this should give you a general idea of how algebraic topology can be used to solve Minesweeper."
+ },
+ {
+ "from": "human",
+ "value": "how would i do it with persistent homology or any of the other topological invariants?"
+ },
+ {
+ "from": "gpt",
+ "value": "To use persistent homology or any of the other topological invariants to solve a Minesweeper game board, you would follow the same basic steps as described above, but with some differences in the specific details.\n\nFor instance, here is how you might use persistent homology to solve a Minesweeper game board:\n\n1. Represent the Minesweeper game board as a topological space. Each square on the board corresponds to a point in the space, and the mines on the board can be thought of as \"holes\" in the space, which are points that are not part of the space.\n2. Compute the persistent homology of the space. Persistent homology is a way of describing the holes in a space over time, and it can be used to infer information about the locations of the mines on the Minesweeper board.\n3. Use the information from the persistent homology to determine the locations of the mines on the board. For example, if the persistent homology indicates that there are four holes in the space, arranged in a square shape, then we can infer that the mines are located in the four corners of the board.\n4. Use this information to solve the Minesweeper game. Once you know the locations of the mines, you can use this information to avoid them and clear the board.\n\nOf course, this is just one possible approach to using persistent homology to solve Minesweeper. There may be other ways to do it, depending on the specific game board and the specific details of the persistent homology computation. However, this should give you a general idea of how persistent homology can be used to solve Minesweeper."
+ },
+ {
+ "from": "human",
+ "value": "how could i implement this in python?"
+ },
+ {
+ "from": "gpt",
+ "value": "To implement a Minesweeper solver that uses algebraic topology in Python, you would need to do the following:\n\n1. Install a Python package for algebraic topology. There are several options available, such as Gudhi, Dionysus, or scikit-tda. These packages provide algorithms and data structures for computing topological invariants, such as homotopy groups, Betti numbers, or persistent homology.\n2. Represent the Minesweeper game board as a topological space. This could be done using a data structure provided by the algebraic topology package, such as a simplicial complex or a Vietoris-Rips complex. Each square on the board would correspond to a point in the space, and the mines on the board would be represented as \"holes\" in the space.\n3. Compute the topological invariants of the space. This could be done using algorithms provided by the algebraic topology package. For example, you could compute the persistent homology of the space using the persistent homology algorithm.\n4. Use the information from the topological invariants to determine the locations of the mines on the board. This could be done using a combination of mathematical reasoning and programming. For example, if the persistent homology indicates that there are four holes in the space, arranged in a square shape, then you could use a piece of code to infer that the mines are located in the four corners of the board.\n5. Use this information to solve the Minesweeper game. Once you know the locations of the mines, you could use a piece of code to automatically click on the safe squares and clear the board.\n\nOf course, this is just one possible approach to implementing a Minesweeper solver that uses algebraic topology in Python. There may be other ways to do it, depending on the specific details of the implementation. However, this should give you a general idea of how to get started with such a project."
+ }
+ ]
+}
diff --git a/tests/fixtures/conversation.tokenized.json b/tests/fixtures/conversation.tokenized.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b7334cc154310025e28325e249f81ab0ac65652
--- /dev/null
+++ b/tests/fixtures/conversation.tokenized.json
@@ -0,0 +1 @@
+{"input_ids": [1, 319, 13563, 1546, 263, 12758, 1404, 322, 385, 23116, 21082, 20255, 29889, 450, 20255, 4076, 8444, 29892, 13173, 29892, 322, 1248, 568, 6089, 304, 278, 1404, 29915, 29879, 5155, 29889, 29871, 3148, 1001, 29901, 920, 1033, 474, 2334, 263, 29086, 705, 11356, 5687, 393, 3667, 4637, 21531, 20159, 304, 4505, 1045, 3163, 29973, 29871, 319, 1799, 9047, 13566, 29901, 739, 29915, 29879, 1950, 304, 671, 21531, 20159, 304, 4505, 341, 1475, 705, 11356, 29892, 541, 372, 723, 5517, 367, 3755, 4280, 29889, 838, 5869, 293, 20159, 338, 263, 5443, 310, 23964, 393, 11898, 278, 4426, 310, 8162, 393, 526, 21634, 1090, 9126, 316, 689, 800, 29892, 1316, 408, 16116, 292, 322, 289, 2548, 29889, 512, 278, 3030, 310, 341, 1475, 705, 11356, 29892, 445, 1033, 367, 1304, 304, 27599, 278, 3829, 310, 278, 3748, 7613, 322, 8161, 278, 14354, 310, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 2398, 29892, 16049, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29889, 739, 29915, 29879, 451, 1554, 393, 508, 367, 10824, 297, 263, 2560, 1234, 21106, 29879, 29958, 2, 3148, 1001, 29901, 1033, 366, 2367, 592, 278, 330, 391, 310, 920, 372, 1033, 2466, 29973, 29871, 319, 1799, 9047, 13566, 29901, 18585, 29889, 450, 6996, 2969, 723, 367, 304, 671, 21531, 20159, 304, 27599, 278, 3829, 310, 278, 341, 1475, 705, 11356, 3748, 7613, 322, 8161, 278, 14354, 310, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 910, 1033, 25135, 15783, 278, 3748, 7613, 408, 263, 25002, 2913, 322, 773, 25002, 15939, 1934, 1316, 408, 3632, 327, 2270, 6471, 322, 350, 9890, 3694, 304, 10115, 2472, 1048, 278, 4423, 310, 29086, 373, 278, 7613, 29889, 2398, 29892, 408, 306, 5276, 1434, 29892, 16049, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29892, 577, 372, 29915, 29879, 451, 1554, 393, 508, 367, 10824, 297, 263, 2560, 1234, 21106, 29879, 29958, 2, 3148, 1001, 29901, 2367, 592, 263, 2702, 1342, 29871, 319, 1799, 9047, 13566, 29901, 1094, 263, 2702, 1342, 29892, 7755, 591, 505, 263, 341, 1475, 705, 11356, 3748, 7613, 411, 278, 1494, 5285, 29901, 13, 28956, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 28956, 13, 797, 445, 1206, 29892, 278, 3748, 7613, 508, 367, 9875, 408, 263, 25002, 2913, 29892, 988, 1269, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29889, 450, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 5293, 21531, 20159, 29892, 591, 508, 27599, 278, 3829, 310, 445, 2913, 322, 8161, 278, 14354, 310, 278, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 13, 13, 2831, 2777, 29892, 697, 982, 304, 437, 445, 338, 304, 10272, 278, 3632, 327, 2270, 6471, 310, 278, 2913, 29889, 15089, 327, 2270, 6471, 526, 263, 982, 310, 20766, 278, 26532, 297, 263, 2913, 29892, 322, 896, 508, 367, 1304, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 341, 1475, 705, 11356, 7613, 29889, 512, 278, 1206, 310, 278, 7613, 2038, 29892, 591, 508, 10272, 278, 937, 3632, 327, 2270, 2318, 310, 278, 2913, 29892, 607, 16612, 278, 1353, 322, 376, 12181, 29908, 310, 278, 26532, 297, 278, 2913, 29889, 512, 445, 1206, 29892, 278, 937, 3632, 327, 2270, 2318, 723, 2649, 502, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29889, 13, 13, 29933, 1463, 373, 445, 2472, 29892, 591, 508, 10115, 393, 278, 29086, 297, 278, 3748, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29892, 408, 4318, 2400, 29901, 13, 28956, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 29871, 29896, 29871, 341, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 341, 259, 29896, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 28956, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1342, 310, 920, 21531, 20159, 1033, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 526, 1784, 916, 1950, 13501, 29892, 322, 278, 2702, 4902, 310, 920, 304, 2334, 963, 723, 8839, 373, 278, 2702, 3748, 7613, 322, 278, 25002, 15939, 1934, 1641, 1304, 29889, 1094, 306, 5276, 1434, 29892, 16049, 263, 1650, 763, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 21106, 29879, 29958, 2, 3148, 1001, 29901, 10241, 474, 471, 4856, 411, 263, 6483, 8004, 310, 1716, 29892, 1033, 366, 5649, 278, 1021, 6964, 304, 592, 411, 393, 11833, 29973, 29871, 319, 1799, 9047, 13566, 29901, 18585, 29889, 17090, 366, 505, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29892, 1244, 338, 920, 366, 1795, 671, 21531, 20159, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29901, 13, 13, 29896, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 7806, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 13, 29906, 29889, 4803, 25002, 15939, 1934, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1670, 526, 1784, 1950, 25002, 15939, 1934, 393, 1033, 367, 1304, 363, 445, 29892, 1316, 408, 3632, 327, 2270, 6471, 29892, 350, 9890, 3694, 29892, 470, 28152, 3632, 3002, 29889, 13, 29941, 29889, 4803, 278, 2472, 10115, 1127, 515, 278, 25002, 15939, 1934, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1152, 1342, 29892, 565, 278, 937, 3632, 327, 2270, 2318, 310, 278, 2913, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 591, 508, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29946, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 508, 671, 445, 2472, 304, 4772, 963, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 773, 21531, 20159, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 3748, 7613, 322, 278, 25002, 15939, 1934, 1641, 1304, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 21531, 20159, 508, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 21106, 29879, 29958, 2, 3148, 1001, 29901, 920, 723, 474, 437, 372, 411, 28152, 3632, 3002, 470, 738, 310, 278, 916, 25002, 15939, 1934, 29973, 29871, 319, 1799, 9047, 13566, 29901, 1763, 671, 28152, 3632, 3002, 470, 738, 310, 278, 916, 25002, 15939, 1934, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29892, 366, 723, 1101, 278, 1021, 6996, 6576, 408, 5439, 2038, 29892, 541, 411, 777, 12651, 297, 278, 2702, 4902, 29889, 13, 13, 2831, 2777, 29892, 1244, 338, 920, 366, 1795, 671, 28152, 3632, 3002, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29901, 13, 13, 29896, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 7806, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 13, 29906, 29889, 11796, 29872, 278, 28152, 3632, 3002, 310, 278, 2913, 29889, 9034, 9696, 3632, 3002, 338, 263, 982, 310, 20766, 278, 26532, 297, 263, 2913, 975, 931, 29892, 322, 372, 508, 367, 1304, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 341, 1475, 705, 11356, 7613, 29889, 13, 29941, 29889, 4803, 278, 2472, 515, 278, 28152, 3632, 3002, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1152, 1342, 29892, 565, 278, 28152, 3632, 3002, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 591, 508, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29946, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 508, 671, 445, 2472, 304, 4772, 963, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 773, 28152, 3632, 3002, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 3748, 7613, 322, 278, 2702, 4902, 310, 278, 28152, 3632, 3002, 16287, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 28152, 3632, 3002, 508, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 21106, 29879, 29958, 2, 3148, 1001, 29901, 920, 1033, 474, 2334, 445, 297, 3017, 29973, 29871, 319, 1799, 9047, 13566, 29901, 1763, 2334, 263, 341, 1475, 705, 11356, 899, 369, 393, 3913, 21531, 20159, 297, 5132, 29892, 366, 723, 817, 304, 437, 278, 1494, 29901, 13, 13, 29896, 29889, 16052, 263, 5132, 3577, 363, 21531, 20159, 29889, 1670, 526, 3196, 3987, 3625, 29892, 1316, 408, 402, 566, 2918, 29892, 360, 291, 952, 375, 29892, 470, 4560, 7354, 29899, 29873, 1388, 29889, 4525, 9741, 3867, 14009, 322, 848, 12286, 363, 20602, 25002, 15939, 1934, 29892, 1316, 408, 3632, 327, 2270, 6471, 29892, 350, 9890, 3694, 29892, 470, 28152, 3632, 3002, 29889, 13, 29906, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 910, 1033, 367, 2309, 773, 263, 848, 3829, 4944, 491, 278, 21531, 20159, 3577, 29892, 1316, 408, 263, 3053, 506, 616, 4280, 470, 263, 478, 2035, 29367, 29899, 29934, 4512, 4280, 29889, 7806, 6862, 373, 278, 7613, 723, 3928, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 723, 367, 9875, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29889, 13, 29941, 29889, 11796, 29872, 278, 25002, 15939, 1934, 310, 278, 2913, 29889, 910, 1033, 367, 2309, 773, 14009, 4944, 491, 278, 21531, 20159, 3577, 29889, 1152, 1342, 29892, 366, 1033, 10272, 278, 28152, 3632, 3002, 310, 278, 2913, 773, 278, 28152, 3632, 3002, 5687, 29889, 13, 29946, 29889, 4803, 278, 2472, 515, 278, 25002, 15939, 1934, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 910, 1033, 367, 2309, 773, 263, 10296, 310, 19475, 24481, 322, 8720, 29889, 1152, 1342, 29892, 565, 278, 28152, 3632, 3002, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 366, 1033, 671, 263, 8424, 310, 775, 304, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29945, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 1033, 671, 263, 8424, 310, 775, 304, 6336, 2828, 373, 278, 9109, 25256, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 16049, 263, 341, 1475, 705, 11356, 899, 369, 393, 3913, 21531, 20159, 297, 5132, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 4902, 310, 278, 5314, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 304, 679, 4687, 411, 1316, 263, 2060, 21106, 29879, 29958, 2], "attention_mask": [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], "labels": [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 739, 29915, 29879, 1950, 304, 671, 21531, 20159, 304, 4505, 341, 1475, 705, 11356, 29892, 541, 372, 723, 5517, 367, 3755, 4280, 29889, 838, 5869, 293, 20159, 338, 263, 5443, 310, 23964, 393, 11898, 278, 4426, 310, 8162, 393, 526, 21634, 1090, 9126, 316, 689, 800, 29892, 1316, 408, 16116, 292, 322, 289, 2548, 29889, 512, 278, 3030, 310, 341, 1475, 705, 11356, 29892, 445, 1033, 367, 1304, 304, 27599, 278, 3829, 310, 278, 3748, 7613, 322, 8161, 278, 14354, 310, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 2398, 29892, 16049, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29889, 739, 29915, 29879, 451, 1554, 393, 508, 367, 10824, 297, 263, 2560, 1234, 21106, 29879, 29958, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 18585, 29889, 450, 6996, 2969, 723, 367, 304, 671, 21531, 20159, 304, 27599, 278, 3829, 310, 278, 341, 1475, 705, 11356, 3748, 7613, 322, 8161, 278, 14354, 310, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 910, 1033, 25135, 15783, 278, 3748, 7613, 408, 263, 25002, 2913, 322, 773, 25002, 15939, 1934, 1316, 408, 3632, 327, 2270, 6471, 322, 350, 9890, 3694, 304, 10115, 2472, 1048, 278, 4423, 310, 29086, 373, 278, 7613, 29889, 2398, 29892, 408, 306, 5276, 1434, 29892, 16049, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29892, 577, 372, 29915, 29879, 451, 1554, 393, 508, 367, 10824, 297, 263, 2560, 1234, 21106, 29879, 29958, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 1094, 263, 2702, 1342, 29892, 7755, 591, 505, 263, 341, 1475, 705, 11356, 3748, 7613, 411, 278, 1494, 5285, 29901, 13, 28956, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 28956, 13, 797, 445, 1206, 29892, 278, 3748, 7613, 508, 367, 9875, 408, 263, 25002, 2913, 29892, 988, 1269, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29889, 450, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 5293, 21531, 20159, 29892, 591, 508, 27599, 278, 3829, 310, 445, 2913, 322, 8161, 278, 14354, 310, 278, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 13, 13, 2831, 2777, 29892, 697, 982, 304, 437, 445, 338, 304, 10272, 278, 3632, 327, 2270, 6471, 310, 278, 2913, 29889, 15089, 327, 2270, 6471, 526, 263, 982, 310, 20766, 278, 26532, 297, 263, 2913, 29892, 322, 896, 508, 367, 1304, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 341, 1475, 705, 11356, 7613, 29889, 512, 278, 1206, 310, 278, 7613, 2038, 29892, 591, 508, 10272, 278, 937, 3632, 327, 2270, 2318, 310, 278, 2913, 29892, 607, 16612, 278, 1353, 322, 376, 12181, 29908, 310, 278, 26532, 297, 278, 2913, 29889, 512, 445, 1206, 29892, 278, 937, 3632, 327, 2270, 2318, 723, 2649, 502, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29889, 13, 13, 29933, 1463, 373, 445, 2472, 29892, 591, 508, 10115, 393, 278, 29086, 297, 278, 3748, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29892, 408, 4318, 2400, 29901, 13, 28956, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 29871, 29896, 29871, 341, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 341, 259, 29896, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 28956, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1342, 310, 920, 21531, 20159, 1033, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 526, 1784, 916, 1950, 13501, 29892, 322, 278, 2702, 4902, 310, 920, 304, 2334, 963, 723, 8839, 373, 278, 2702, 3748, 7613, 322, 278, 25002, 15939, 1934, 1641, 1304, 29889, 1094, 306, 5276, 1434, 29892, 16049, 263, 1650, 763, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 21106, 29879, 29958, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 18585, 29889, 17090, 366, 505, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29892, 1244, 338, 920, 366, 1795, 671, 21531, 20159, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29901, 13, 13, 29896, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 7806, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 13, 29906, 29889, 4803, 25002, 15939, 1934, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1670, 526, 1784, 1950, 25002, 15939, 1934, 393, 1033, 367, 1304, 363, 445, 29892, 1316, 408, 3632, 327, 2270, 6471, 29892, 350, 9890, 3694, 29892, 470, 28152, 3632, 3002, 29889, 13, 29941, 29889, 4803, 278, 2472, 10115, 1127, 515, 278, 25002, 15939, 1934, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1152, 1342, 29892, 565, 278, 937, 3632, 327, 2270, 2318, 310, 278, 2913, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 591, 508, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29946, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 508, 671, 445, 2472, 304, 4772, 963, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 773, 21531, 20159, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 3748, 7613, 322, 278, 25002, 15939, 1934, 1641, 1304, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 21531, 20159, 508, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 21106, 29879, 29958, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 1763, 671, 28152, 3632, 3002, 470, 738, 310, 278, 916, 25002, 15939, 1934, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29892, 366, 723, 1101, 278, 1021, 6996, 6576, 408, 5439, 2038, 29892, 541, 411, 777, 12651, 297, 278, 2702, 4902, 29889, 13, 13, 2831, 2777, 29892, 1244, 338, 920, 366, 1795, 671, 28152, 3632, 3002, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29901, 13, 13, 29896, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 7806, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 13, 29906, 29889, 11796, 29872, 278, 28152, 3632, 3002, 310, 278, 2913, 29889, 9034, 9696, 3632, 3002, 338, 263, 982, 310, 20766, 278, 26532, 297, 263, 2913, 975, 931, 29892, 322, 372, 508, 367, 1304, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 341, 1475, 705, 11356, 7613, 29889, 13, 29941, 29889, 4803, 278, 2472, 515, 278, 28152, 3632, 3002, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1152, 1342, 29892, 565, 278, 28152, 3632, 3002, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 591, 508, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29946, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 508, 671, 445, 2472, 304, 4772, 963, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 773, 28152, 3632, 3002, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 3748, 7613, 322, 278, 2702, 4902, 310, 278, 28152, 3632, 3002, 16287, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 28152, 3632, 3002, 508, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 21106, 29879, 29958, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 1763, 2334, 263, 341, 1475, 705, 11356, 899, 369, 393, 3913, 21531, 20159, 297, 5132, 29892, 366, 723, 817, 304, 437, 278, 1494, 29901, 13, 13, 29896, 29889, 16052, 263, 5132, 3577, 363, 21531, 20159, 29889, 1670, 526, 3196, 3987, 3625, 29892, 1316, 408, 402, 566, 2918, 29892, 360, 291, 952, 375, 29892, 470, 4560, 7354, 29899, 29873, 1388, 29889, 4525, 9741, 3867, 14009, 322, 848, 12286, 363, 20602, 25002, 15939, 1934, 29892, 1316, 408, 3632, 327, 2270, 6471, 29892, 350, 9890, 3694, 29892, 470, 28152, 3632, 3002, 29889, 13, 29906, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 910, 1033, 367, 2309, 773, 263, 848, 3829, 4944, 491, 278, 21531, 20159, 3577, 29892, 1316, 408, 263, 3053, 506, 616, 4280, 470, 263, 478, 2035, 29367, 29899, 29934, 4512, 4280, 29889, 7806, 6862, 373, 278, 7613, 723, 3928, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 723, 367, 9875, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29889, 13, 29941, 29889, 11796, 29872, 278, 25002, 15939, 1934, 310, 278, 2913, 29889, 910, 1033, 367, 2309, 773, 14009, 4944, 491, 278, 21531, 20159, 3577, 29889, 1152, 1342, 29892, 366, 1033, 10272, 278, 28152, 3632, 3002, 310, 278, 2913, 773, 278, 28152, 3632, 3002, 5687, 29889, 13, 29946, 29889, 4803, 278, 2472, 515, 278, 25002, 15939, 1934, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 910, 1033, 367, 2309, 773, 263, 10296, 310, 19475, 24481, 322, 8720, 29889, 1152, 1342, 29892, 565, 278, 28152, 3632, 3002, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 366, 1033, 671, 263, 8424, 310, 775, 304, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29945, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 1033, 671, 263, 8424, 310, 775, 304, 6336, 2828, 373, 278, 9109, 25256, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 16049, 263, 341, 1475, 705, 11356, 899, 369, 393, 3913, 21531, 20159, 297, 5132, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 4902, 310, 278, 5314, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 304, 679, 4687, 411, 1316, 263, 2060, 21106, 29879, 29958, 2]}
diff --git a/tests/fixtures/conversation.tokenized_llama2chat.json b/tests/fixtures/conversation.tokenized_llama2chat.json
new file mode 100644
index 0000000000000000000000000000000000000000..a4109fc3e206a455b02f9607c7f325afa71435cb
--- /dev/null
+++ b/tests/fixtures/conversation.tokenized_llama2chat.json
@@ -0,0 +1 @@
+{"input_ids": [1, 518, 25580, 29962, 3532, 14816, 29903, 6778, 13, 3492, 526, 263, 8444, 29892, 3390, 1319, 322, 15993, 20255, 29889, 29849, 1234, 408, 1371, 3730, 408, 1950, 29892, 1550, 1641, 9109, 29889, 3575, 6089, 881, 451, 3160, 738, 10311, 1319, 29892, 443, 621, 936, 29892, 11021, 391, 29892, 7916, 391, 29892, 304, 27375, 29892, 18215, 29892, 470, 27302, 2793, 29889, 3529, 9801, 393, 596, 20890, 526, 5374, 635, 443, 5365, 1463, 322, 6374, 297, 5469, 29889, 13, 13, 3644, 263, 1139, 947, 451, 1207, 738, 4060, 29892, 470, 338, 451, 2114, 1474, 16165, 261, 296, 29892, 5649, 2020, 2012, 310, 22862, 1554, 451, 1959, 29889, 960, 366, 1016, 29915, 29873, 1073, 278, 1234, 304, 263, 1139, 29892, 3113, 1016, 29915, 29873, 6232, 2089, 2472, 29889, 13, 29966, 829, 14816, 29903, 6778, 13, 13, 3525, 1033, 474, 2334, 263, 29086, 705, 11356, 5687, 393, 3667, 4637, 21531, 20159, 304, 4505, 1045, 3163, 29973, 518, 29914, 25580, 29962, 739, 29915, 29879, 1950, 304, 671, 21531, 20159, 304, 4505, 341, 1475, 705, 11356, 29892, 541, 372, 723, 5517, 367, 3755, 4280, 29889, 838, 5869, 293, 20159, 338, 263, 5443, 310, 23964, 393, 11898, 278, 4426, 310, 8162, 393, 526, 21634, 1090, 9126, 316, 689, 800, 29892, 1316, 408, 16116, 292, 322, 289, 2548, 29889, 512, 278, 3030, 310, 341, 1475, 705, 11356, 29892, 445, 1033, 367, 1304, 304, 27599, 278, 3829, 310, 278, 3748, 7613, 322, 8161, 278, 14354, 310, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 2398, 29892, 16049, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29889, 739, 29915, 29879, 451, 1554, 393, 508, 367, 10824, 297, 263, 2560, 1234, 29889, 29871, 2, 1, 518, 25580, 29962, 1033, 366, 2367, 592, 278, 330, 391, 310, 920, 372, 1033, 2466, 29973, 518, 29914, 25580, 29962, 18585, 29889, 450, 6996, 2969, 723, 367, 304, 671, 21531, 20159, 304, 27599, 278, 3829, 310, 278, 341, 1475, 705, 11356, 3748, 7613, 322, 8161, 278, 14354, 310, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 910, 1033, 25135, 15783, 278, 3748, 7613, 408, 263, 25002, 2913, 322, 773, 25002, 15939, 1934, 1316, 408, 3632, 327, 2270, 6471, 322, 350, 9890, 3694, 304, 10115, 2472, 1048, 278, 4423, 310, 29086, 373, 278, 7613, 29889, 2398, 29892, 408, 306, 5276, 1434, 29892, 16049, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29892, 577, 372, 29915, 29879, 451, 1554, 393, 508, 367, 10824, 297, 263, 2560, 1234, 29889, 29871, 2, 1, 518, 25580, 29962, 2367, 592, 263, 2702, 1342, 518, 29914, 25580, 29962, 1094, 263, 2702, 1342, 29892, 7755, 591, 505, 263, 341, 1475, 705, 11356, 3748, 7613, 411, 278, 1494, 5285, 29901, 13, 28956, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 28956, 13, 797, 445, 1206, 29892, 278, 3748, 7613, 508, 367, 9875, 408, 263, 25002, 2913, 29892, 988, 1269, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29889, 450, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 5293, 21531, 20159, 29892, 591, 508, 27599, 278, 3829, 310, 445, 2913, 322, 8161, 278, 14354, 310, 278, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 13, 13, 2831, 2777, 29892, 697, 982, 304, 437, 445, 338, 304, 10272, 278, 3632, 327, 2270, 6471, 310, 278, 2913, 29889, 15089, 327, 2270, 6471, 526, 263, 982, 310, 20766, 278, 26532, 297, 263, 2913, 29892, 322, 896, 508, 367, 1304, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 341, 1475, 705, 11356, 7613, 29889, 512, 278, 1206, 310, 278, 7613, 2038, 29892, 591, 508, 10272, 278, 937, 3632, 327, 2270, 2318, 310, 278, 2913, 29892, 607, 16612, 278, 1353, 322, 376, 12181, 29908, 310, 278, 26532, 297, 278, 2913, 29889, 512, 445, 1206, 29892, 278, 937, 3632, 327, 2270, 2318, 723, 2649, 502, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29889, 13, 13, 29933, 1463, 373, 445, 2472, 29892, 591, 508, 10115, 393, 278, 29086, 297, 278, 3748, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29892, 408, 4318, 2400, 29901, 13, 28956, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 29871, 29896, 29871, 341, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 341, 259, 29896, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 28956, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1342, 310, 920, 21531, 20159, 1033, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 526, 1784, 916, 1950, 13501, 29892, 322, 278, 2702, 4902, 310, 920, 304, 2334, 963, 723, 8839, 373, 278, 2702, 3748, 7613, 322, 278, 25002, 15939, 1934, 1641, 1304, 29889, 1094, 306, 5276, 1434, 29892, 16049, 263, 1650, 763, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29889, 29871, 2, 1, 518, 25580, 29962, 10241, 474, 471, 4856, 411, 263, 6483, 8004, 310, 1716, 29892, 1033, 366, 5649, 278, 1021, 6964, 304, 592, 411, 393, 11833, 29973, 518, 29914, 25580, 29962, 18585, 29889, 17090, 366, 505, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29892, 1244, 338, 920, 366, 1795, 671, 21531, 20159, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29901, 13, 13, 29896, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 7806, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 13, 29906, 29889, 4803, 25002, 15939, 1934, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1670, 526, 1784, 1950, 25002, 15939, 1934, 393, 1033, 367, 1304, 363, 445, 29892, 1316, 408, 3632, 327, 2270, 6471, 29892, 350, 9890, 3694, 29892, 470, 28152, 3632, 3002, 29889, 13, 29941, 29889, 4803, 278, 2472, 10115, 1127, 515, 278, 25002, 15939, 1934, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1152, 1342, 29892, 565, 278, 937, 3632, 327, 2270, 2318, 310, 278, 2913, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 591, 508, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29946, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 508, 671, 445, 2472, 304, 4772, 963, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 773, 21531, 20159, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 3748, 7613, 322, 278, 25002, 15939, 1934, 1641, 1304, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 21531, 20159, 508, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 29889, 29871, 2, 1, 518, 25580, 29962, 920, 723, 474, 437, 372, 411, 28152, 3632, 3002, 470, 738, 310, 278, 916, 25002, 15939, 1934, 29973, 518, 29914, 25580, 29962, 1763, 671, 28152, 3632, 3002, 470, 738, 310, 278, 916, 25002, 15939, 1934, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29892, 366, 723, 1101, 278, 1021, 6996, 6576, 408, 5439, 2038, 29892, 541, 411, 777, 12651, 297, 278, 2702, 4902, 29889, 13, 13, 2831, 2777, 29892, 1244, 338, 920, 366, 1795, 671, 28152, 3632, 3002, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29901, 13, 13, 29896, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 7806, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 13, 29906, 29889, 11796, 29872, 278, 28152, 3632, 3002, 310, 278, 2913, 29889, 9034, 9696, 3632, 3002, 338, 263, 982, 310, 20766, 278, 26532, 297, 263, 2913, 975, 931, 29892, 322, 372, 508, 367, 1304, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 341, 1475, 705, 11356, 7613, 29889, 13, 29941, 29889, 4803, 278, 2472, 515, 278, 28152, 3632, 3002, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1152, 1342, 29892, 565, 278, 28152, 3632, 3002, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 591, 508, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29946, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 508, 671, 445, 2472, 304, 4772, 963, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 773, 28152, 3632, 3002, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 3748, 7613, 322, 278, 2702, 4902, 310, 278, 28152, 3632, 3002, 16287, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 28152, 3632, 3002, 508, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 29889, 29871, 2, 1, 518, 25580, 29962, 920, 1033, 474, 2334, 445, 297, 3017, 29973, 518, 29914, 25580, 29962, 1763, 2334, 263, 341, 1475, 705, 11356, 899, 369, 393, 3913, 21531, 20159, 297, 5132, 29892, 366, 723, 817, 304, 437, 278, 1494, 29901, 13, 13, 29896, 29889, 16052, 263, 5132, 3577, 363, 21531, 20159, 29889, 1670, 526, 3196, 3987, 3625, 29892, 1316, 408, 402, 566, 2918, 29892, 360, 291, 952, 375, 29892, 470, 4560, 7354, 29899, 29873, 1388, 29889, 4525, 9741, 3867, 14009, 322, 848, 12286, 363, 20602, 25002, 15939, 1934, 29892, 1316, 408, 3632, 327, 2270, 6471, 29892, 350, 9890, 3694, 29892, 470, 28152, 3632, 3002, 29889, 13, 29906, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 910, 1033, 367, 2309, 773, 263, 848, 3829, 4944, 491, 278, 21531, 20159, 3577, 29892, 1316, 408, 263, 3053, 506, 616, 4280, 470, 263, 478, 2035, 29367, 29899, 29934, 4512, 4280, 29889, 7806, 6862, 373, 278, 7613, 723, 3928, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 723, 367, 9875, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29889, 13, 29941, 29889, 11796, 29872, 278, 25002, 15939, 1934, 310, 278, 2913, 29889, 910, 1033, 367, 2309, 773, 14009, 4944, 491, 278, 21531, 20159, 3577, 29889, 1152, 1342, 29892, 366, 1033, 10272, 278, 28152, 3632, 3002, 310, 278, 2913, 773, 278, 28152, 3632, 3002, 5687, 29889, 13, 29946, 29889, 4803, 278, 2472, 515, 278, 25002, 15939, 1934, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 910, 1033, 367, 2309, 773, 263, 10296, 310, 19475, 24481, 322, 8720, 29889, 1152, 1342, 29892, 565, 278, 28152, 3632, 3002, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 366, 1033, 671, 263, 8424, 310, 775, 304, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29945, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 1033, 671, 263, 8424, 310, 775, 304, 6336, 2828, 373, 278, 9109, 25256, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 16049, 263, 341, 1475, 705, 11356, 899, 369, 393, 3913, 21531, 20159, 297, 5132, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 4902, 310, 278, 5314, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 304, 679, 4687, 411, 1316, 263, 2060, 29889, 29871, 2, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], "labels": [-100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 739, 29915, 29879, 1950, 304, 671, 21531, 20159, 304, 4505, 341, 1475, 705, 11356, 29892, 541, 372, 723, 5517, 367, 3755, 4280, 29889, 838, 5869, 293, 20159, 338, 263, 5443, 310, 23964, 393, 11898, 278, 4426, 310, 8162, 393, 526, 21634, 1090, 9126, 316, 689, 800, 29892, 1316, 408, 16116, 292, 322, 289, 2548, 29889, 512, 278, 3030, 310, 341, 1475, 705, 11356, 29892, 445, 1033, 367, 1304, 304, 27599, 278, 3829, 310, 278, 3748, 7613, 322, 8161, 278, 14354, 310, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 2398, 29892, 16049, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29889, 739, 29915, 29879, 451, 1554, 393, 508, 367, 10824, 297, 263, 2560, 1234, 29889, 29871, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 18585, 29889, 450, 6996, 2969, 723, 367, 304, 671, 21531, 20159, 304, 27599, 278, 3829, 310, 278, 341, 1475, 705, 11356, 3748, 7613, 322, 8161, 278, 14354, 310, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 910, 1033, 25135, 15783, 278, 3748, 7613, 408, 263, 25002, 2913, 322, 773, 25002, 15939, 1934, 1316, 408, 3632, 327, 2270, 6471, 322, 350, 9890, 3694, 304, 10115, 2472, 1048, 278, 4423, 310, 29086, 373, 278, 7613, 29889, 2398, 29892, 408, 306, 5276, 1434, 29892, 16049, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29892, 577, 372, 29915, 29879, 451, 1554, 393, 508, 367, 10824, 297, 263, 2560, 1234, 29889, 29871, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 1094, 263, 2702, 1342, 29892, 7755, 591, 505, 263, 341, 1475, 705, 11356, 3748, 7613, 411, 278, 1494, 5285, 29901, 13, 28956, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 28956, 13, 797, 445, 1206, 29892, 278, 3748, 7613, 508, 367, 9875, 408, 263, 25002, 2913, 29892, 988, 1269, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29889, 450, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 5293, 21531, 20159, 29892, 591, 508, 27599, 278, 3829, 310, 445, 2913, 322, 8161, 278, 14354, 310, 278, 29086, 2729, 373, 3058, 25002, 15939, 1934, 29889, 13, 13, 2831, 2777, 29892, 697, 982, 304, 437, 445, 338, 304, 10272, 278, 3632, 327, 2270, 6471, 310, 278, 2913, 29889, 15089, 327, 2270, 6471, 526, 263, 982, 310, 20766, 278, 26532, 297, 263, 2913, 29892, 322, 896, 508, 367, 1304, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 341, 1475, 705, 11356, 7613, 29889, 512, 278, 1206, 310, 278, 7613, 2038, 29892, 591, 508, 10272, 278, 937, 3632, 327, 2270, 2318, 310, 278, 2913, 29892, 607, 16612, 278, 1353, 322, 376, 12181, 29908, 310, 278, 26532, 297, 278, 2913, 29889, 512, 445, 1206, 29892, 278, 937, 3632, 327, 2270, 2318, 723, 2649, 502, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29889, 13, 13, 29933, 1463, 373, 445, 2472, 29892, 591, 508, 10115, 393, 278, 29086, 297, 278, 3748, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29892, 408, 4318, 2400, 29901, 13, 28956, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 29871, 29896, 29871, 341, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 1577, 259, 29896, 13, 29871, 29896, 29871, 1577, 29871, 1577, 29871, 341, 259, 29896, 13, 29871, 29896, 259, 29896, 259, 29896, 259, 29896, 259, 29896, 13, 28956, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1342, 310, 920, 21531, 20159, 1033, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 526, 1784, 916, 1950, 13501, 29892, 322, 278, 2702, 4902, 310, 920, 304, 2334, 963, 723, 8839, 373, 278, 2702, 3748, 7613, 322, 278, 25002, 15939, 1934, 1641, 1304, 29889, 1094, 306, 5276, 1434, 29892, 16049, 263, 1650, 763, 445, 723, 1996, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29889, 29871, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 18585, 29889, 17090, 366, 505, 263, 6483, 8004, 310, 1716, 21531, 20159, 322, 278, 3748, 310, 341, 1475, 705, 11356, 29892, 1244, 338, 920, 366, 1795, 671, 21531, 20159, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29901, 13, 13, 29896, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 7806, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 13, 29906, 29889, 4803, 25002, 15939, 1934, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1670, 526, 1784, 1950, 25002, 15939, 1934, 393, 1033, 367, 1304, 363, 445, 29892, 1316, 408, 3632, 327, 2270, 6471, 29892, 350, 9890, 3694, 29892, 470, 28152, 3632, 3002, 29889, 13, 29941, 29889, 4803, 278, 2472, 10115, 1127, 515, 278, 25002, 15939, 1934, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1152, 1342, 29892, 565, 278, 937, 3632, 327, 2270, 2318, 310, 278, 2913, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 591, 508, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29946, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 508, 671, 445, 2472, 304, 4772, 963, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 773, 21531, 20159, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 3748, 7613, 322, 278, 25002, 15939, 1934, 1641, 1304, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 21531, 20159, 508, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 29889, 29871, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 1763, 671, 28152, 3632, 3002, 470, 738, 310, 278, 916, 25002, 15939, 1934, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29892, 366, 723, 1101, 278, 1021, 6996, 6576, 408, 5439, 2038, 29892, 541, 411, 777, 12651, 297, 278, 2702, 4902, 29889, 13, 13, 2831, 2777, 29892, 1244, 338, 920, 366, 1795, 671, 28152, 3632, 3002, 304, 4505, 263, 341, 1475, 705, 11356, 3748, 7613, 29901, 13, 13, 29896, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 7806, 6862, 373, 278, 7613, 16161, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 508, 367, 2714, 310, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29892, 607, 526, 3291, 393, 526, 451, 760, 310, 278, 2913, 29889, 13, 29906, 29889, 11796, 29872, 278, 28152, 3632, 3002, 310, 278, 2913, 29889, 9034, 9696, 3632, 3002, 338, 263, 982, 310, 20766, 278, 26532, 297, 263, 2913, 975, 931, 29892, 322, 372, 508, 367, 1304, 304, 10115, 2472, 1048, 278, 14354, 310, 278, 29086, 373, 278, 341, 1475, 705, 11356, 7613, 29889, 13, 29941, 29889, 4803, 278, 2472, 515, 278, 28152, 3632, 3002, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 1152, 1342, 29892, 565, 278, 28152, 3632, 3002, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 591, 508, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29946, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 508, 671, 445, 2472, 304, 4772, 963, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 773, 28152, 3632, 3002, 304, 4505, 341, 1475, 705, 11356, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 3748, 7613, 322, 278, 2702, 4902, 310, 278, 28152, 3632, 3002, 16287, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 28152, 3632, 3002, 508, 367, 1304, 304, 4505, 341, 1475, 705, 11356, 29889, 29871, 2, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, 1763, 2334, 263, 341, 1475, 705, 11356, 899, 369, 393, 3913, 21531, 20159, 297, 5132, 29892, 366, 723, 817, 304, 437, 278, 1494, 29901, 13, 13, 29896, 29889, 16052, 263, 5132, 3577, 363, 21531, 20159, 29889, 1670, 526, 3196, 3987, 3625, 29892, 1316, 408, 402, 566, 2918, 29892, 360, 291, 952, 375, 29892, 470, 4560, 7354, 29899, 29873, 1388, 29889, 4525, 9741, 3867, 14009, 322, 848, 12286, 363, 20602, 25002, 15939, 1934, 29892, 1316, 408, 3632, 327, 2270, 6471, 29892, 350, 9890, 3694, 29892, 470, 28152, 3632, 3002, 29889, 13, 29906, 29889, 16314, 278, 341, 1475, 705, 11356, 3748, 7613, 408, 263, 25002, 2913, 29889, 910, 1033, 367, 2309, 773, 263, 848, 3829, 4944, 491, 278, 21531, 20159, 3577, 29892, 1316, 408, 263, 3053, 506, 616, 4280, 470, 263, 478, 2035, 29367, 29899, 29934, 4512, 4280, 29889, 7806, 6862, 373, 278, 7613, 723, 3928, 304, 263, 1298, 297, 278, 2913, 29892, 322, 278, 29086, 373, 278, 7613, 723, 367, 9875, 408, 376, 5391, 267, 29908, 297, 278, 2913, 29889, 13, 29941, 29889, 11796, 29872, 278, 25002, 15939, 1934, 310, 278, 2913, 29889, 910, 1033, 367, 2309, 773, 14009, 4944, 491, 278, 21531, 20159, 3577, 29889, 1152, 1342, 29892, 366, 1033, 10272, 278, 28152, 3632, 3002, 310, 278, 2913, 773, 278, 28152, 3632, 3002, 5687, 29889, 13, 29946, 29889, 4803, 278, 2472, 515, 278, 25002, 15939, 1934, 304, 8161, 278, 14354, 310, 278, 29086, 373, 278, 7613, 29889, 910, 1033, 367, 2309, 773, 263, 10296, 310, 19475, 24481, 322, 8720, 29889, 1152, 1342, 29892, 565, 278, 28152, 3632, 3002, 14088, 393, 727, 526, 3023, 26532, 297, 278, 2913, 29892, 21050, 297, 263, 6862, 8267, 29892, 769, 366, 1033, 671, 263, 8424, 310, 775, 304, 10115, 393, 278, 29086, 526, 5982, 297, 278, 3023, 26995, 310, 278, 7613, 29889, 13, 29945, 29889, 4803, 445, 2472, 304, 4505, 278, 341, 1475, 705, 11356, 3748, 29889, 9038, 366, 1073, 278, 14354, 310, 278, 29086, 29892, 366, 1033, 671, 263, 8424, 310, 775, 304, 6336, 2828, 373, 278, 9109, 25256, 322, 2821, 278, 7613, 29889, 13, 13, 2776, 3236, 29892, 445, 338, 925, 697, 1950, 2948, 304, 16049, 263, 341, 1475, 705, 11356, 899, 369, 393, 3913, 21531, 20159, 297, 5132, 29889, 1670, 1122, 367, 916, 5837, 304, 437, 372, 29892, 8679, 373, 278, 2702, 4902, 310, 278, 5314, 29889, 2398, 29892, 445, 881, 2367, 366, 263, 2498, 2969, 310, 920, 304, 679, 4687, 411, 1316, 263, 2060, 29889, 29871, 2, 1, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100], "attention_mask": [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false]}
diff --git a/tests/monkeypatch/test_llama_attn_hijack_flash.py b/tests/monkeypatch/test_llama_attn_hijack_flash.py
new file mode 100644
index 0000000000000000000000000000000000000000..4521cd07bc39e6bf8c3486ea6fdf6af258b8f3cb
--- /dev/null
+++ b/tests/monkeypatch/test_llama_attn_hijack_flash.py
@@ -0,0 +1,113 @@
+"""
+Unit tests for the monkeypatch utils
+"""
+import unittest
+
+import torch
+
+from axolotl.monkeypatch.utils import (
+ get_cu_seqlens,
+ get_cu_seqlens_from_pos_ids,
+ get_max_seqlen_in_batch,
+ get_unpad_data,
+)
+
+
+class TestMonkeyPatchUtils(unittest.TestCase):
+ """
+ Unit test class for monkeypatch utils
+ """
+
+ def test_get_cu_seqlens_1d(self):
+ attn_mask = torch.tensor([[1, 1, 1, 1, 2, 2, 2, 3, 3, 3, 3, 3, 4, 4, 0, 0]])
+ target_res = torch.tensor([0, 4, 7, 12, 14, 16], dtype=torch.int32)
+ self.assertTrue(torch.allclose(get_cu_seqlens(attn_mask)[0], target_res))
+
+ def test_get_cu_seqlens_from_pos_ids_1d(self):
+ position_ids = torch.tensor([[0, 1, 2, 3, 0, 1, 2, 0, 1, 2, 3, 4, 0, 1, 0, 0]])
+ target_res = torch.tensor([0, 4, 7, 12, 14, 16], dtype=torch.int32)
+ self.assertTrue(
+ torch.allclose(get_cu_seqlens_from_pos_ids(position_ids)[0], target_res)
+ )
+
+ def test_get_cu_seqlens_from_pos_ids_2d(self):
+ position_ids = torch.tensor(
+ [
+ [0, 1, 2, 3, 0, 1, 2, 0, 1, 2, 3, 4, 0, 1, 0, 0],
+ [0, 1, 2, 3, 4, 0, 1, 2, 0, 1, 2, 3, 4, 5, 6, 0],
+ ]
+ )
+ target_res = torch.tensor(
+ [[0, 4, 7, 12, 14, 16], [0, 5, 8, 15, 16, 16]], dtype=torch.int32
+ )
+ self.assertTrue(
+ torch.allclose(get_cu_seqlens_from_pos_ids(position_ids)[0], target_res)
+ )
+
+ def test_get_max_seqlen_in_batch(self):
+ attn_mask = torch.tensor([[1, 1, 1, 1, 2, 2, 2, 3, 3, 3, 3, 3, 4, 4, 0, 0]])
+ target_res = torch.tensor([4, 3, 5, 2], dtype=torch.int32)
+ self.assertTrue(torch.allclose(get_max_seqlen_in_batch(attn_mask), target_res))
+
+ def test_get_unpad_data(self):
+ attn_mask = torch.tensor([[1, 1, 1, 1, 2, 2, 2, 3, 3, 3, 3, 3, 4, 4, 0, 0]])
+ target_indices = torch.tensor([0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13])
+ target_cu_seqlen = torch.tensor([0, 4, 7, 12, 14], dtype=torch.int32)
+ target_max_seqlen_in_batch = 5
+ indices, cu_seqlen, max_seqlen_in_batch = get_unpad_data(attn_mask)
+ self.assertTrue(torch.allclose(target_indices, indices))
+ self.assertTrue(torch.allclose(target_cu_seqlen, cu_seqlen))
+ self.assertEqual(target_max_seqlen_in_batch, max_seqlen_in_batch)
+
+ attn_mask = torch.tensor(
+ [
+ [1, 1, 1, 1, 2, 2, 2, 3, 3, 3, 3, 3, 4, 4, 0, 0],
+ [1, 1, 1, 2, 2, 2, 2, 2, 3, 3, 4, 4, 4, 5, 5, 5],
+ ]
+ )
+ target_indices = torch.tensor(
+ [
+ 0,
+ 1,
+ 2,
+ 3,
+ 4,
+ 5,
+ 6,
+ 7,
+ 8,
+ 9,
+ 10,
+ 11,
+ 12,
+ 13,
+ 16,
+ 17,
+ 18,
+ 19,
+ 20,
+ 21,
+ 22,
+ 23,
+ 24,
+ 25,
+ 26,
+ 27,
+ 28,
+ 29,
+ 30,
+ 31,
+ ]
+ )
+ target_cu_seqlen = torch.tensor(
+ [0, 4, 7, 12, 14, 17, 22, 24, 27, 30], dtype=torch.int32
+ )
+ target_max_seqlen_in_batch = 5
+ indices, cu_seqlen, max_seqlen_in_batch = get_unpad_data(attn_mask)
+ self.assertTrue(torch.allclose(target_indices, indices))
+ self.assertTrue(torch.allclose(target_cu_seqlen, cu_seqlen))
+ self.assertEqual(target_max_seqlen_in_batch, max_seqlen_in_batch)
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/prompt_strategies/test_alpaca.py b/tests/prompt_strategies/test_alpaca.py
new file mode 100644
index 0000000000000000000000000000000000000000..51dd5900bb14b0ed43beb10ea27a287f6b302f9c
--- /dev/null
+++ b/tests/prompt_strategies/test_alpaca.py
@@ -0,0 +1,118 @@
+"""
+Test module for alpaca integration w chatml
+"""
+import pytest
+from datasets import Dataset
+from tokenizers import AddedToken
+from transformers import AutoTokenizer
+
+from axolotl.datasets import TokenizedPromptDataset
+from axolotl.prompt_tokenizers import AlpacaPromptTokenizingStrategy
+from axolotl.prompters import AlpacaPrompter, PromptStyle
+
+
+@pytest.fixture(name="alpaca_dataset")
+def fixture_alpaca_dataset():
+ return Dataset.from_list(
+ [
+ {
+ "instruction": "Evaluate this sentence for spelling and grammar mistakes",
+ "input": "He finnished his meal and left the resturant",
+ "output": "He finished his meal and left the restaurant.",
+ }
+ ]
+ )
+
+
+@pytest.fixture(name="tokenizer")
+def fixture_tokenizer():
+ # pylint: disable=all
+ tokenizer = AutoTokenizer.from_pretrained(
+ "casperhansen/mistral-7b-instruct-v0.1-awq"
+ )
+ tokenizer.add_special_tokens(
+ {
+ "eos_token": AddedToken(
+ "<|im_end|>", rstrip=False, lstrip=False, normalized=False
+ )
+ }
+ )
+ tokenizer.add_tokens(
+ [
+ AddedToken("<|im_start|>", rstrip=False, lstrip=False, normalized=False),
+ ]
+ )
+
+ return tokenizer
+
+
+class TestAlpacaChatml:
+ """
+ Test class for alpaca prompter
+ """
+
+ def test_no_double_im_end(self, alpaca_dataset, tokenizer):
+ strategy = AlpacaPromptTokenizingStrategy(
+ AlpacaPrompter(prompt_style=PromptStyle.CHATML.value),
+ tokenizer,
+ False, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, alpaca_dataset, process_count=1
+ )
+
+ input_ids = dataset_wrapper[0]["input_ids"]
+ # fmt: off
+ assert input_ids == [
+ 1, # Bos
+ 32001, 1587, 13, 20548, 336, 349, 396, 13126, 369, 13966, 264, 3638, 28725, 5881, 1360, 395, 396, 2787, 369, 5312, 3629, 2758, 28723, 12018, 264, 2899, 369, 6582, 1999, 2691, 274, 272, 2159, 28723, 32000, 28705, 13, # instruction
+ 32001, 2188, 13, 16627, 11931, 456, 12271, 354, 668, 3572, 304, 18756, 3479, 17179, 13, 2428, 854, 28711, 1497, 516, 11314, 304, 1749, 272, 1846, 324, 440, 32000, 28705, 13, # input
+ 32001, 13892, 13, 650, 5967, 516, 11314, 304, 1749, 272, 9926, 28723, 32000, # output
+ ]
+ # fmt: on
+
+ def test_no_train_on_input(self, alpaca_dataset, tokenizer):
+ strategy = AlpacaPromptTokenizingStrategy(
+ AlpacaPrompter(prompt_style=PromptStyle.CHATML.value),
+ tokenizer,
+ False, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, alpaca_dataset, process_count=1
+ )
+
+ labels = dataset_wrapper[0]["labels"]
+ # fmt: off
+ assert labels == [
+ -100, # bos
+ -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, # instruction
+ -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, # input
+ -100, -100, -100, 650, 5967, 516, 11314, 304, 1749, 272, 9926, 28723, 32000, # Output
+ ]
+ # fmt: on
+
+ def test_w_train_on_input(self, alpaca_dataset, tokenizer):
+ strategy = AlpacaPromptTokenizingStrategy(
+ AlpacaPrompter(prompt_style=PromptStyle.CHATML.value),
+ tokenizer,
+ True, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, alpaca_dataset, process_count=1
+ )
+
+ labels = dataset_wrapper[0]["labels"]
+ # fmt: off
+ assert labels == [
+ 1, # Bos
+ 32001, 1587, 13, 20548, 336, 349, 396, 13126, 369, 13966, 264, 3638, 28725, 5881, 1360, 395, 396, 2787, 369, 5312, 3629, 2758, 28723, 12018, 264, 2899, 369, 6582, 1999, 2691, 274, 272, 2159, 28723, 32000, 28705, 13, # instruction
+ 32001, 2188, 13, 16627, 11931, 456, 12271, 354, 668, 3572, 304, 18756, 3479, 17179, 13, 2428, 854, 28711, 1497, 516, 11314, 304, 1749, 272, 1846, 324, 440, 32000, 28705, 13, # input
+ 32001, 13892, 13, 650, 5967, 516, 11314, 304, 1749, 272, 9926, 28723, 32000, # output
+ ]
+ # fmt: on
diff --git a/tests/prompt_strategies/test_raw_io.py b/tests/prompt_strategies/test_raw_io.py
new file mode 100644
index 0000000000000000000000000000000000000000..967de169f32a0069617977833ea30ad54139d7bc
--- /dev/null
+++ b/tests/prompt_strategies/test_raw_io.py
@@ -0,0 +1,118 @@
+"""
+Test module for raw i/o data for prompts
+"""
+import pytest
+from datasets import Dataset
+from tokenizers import AddedToken
+from transformers import AutoTokenizer
+
+from axolotl.datasets import TokenizedPromptDataset
+from axolotl.prompt_strategies.input_output import (
+ RawInputOutputPrompter,
+ RawInputOutputStrategy,
+)
+
+
+@pytest.fixture(name="segments_dataset")
+def fixture_sharegpt_dataset():
+ return Dataset.from_list(
+ [
+ {
+ "segments": [
+ {
+ "label": False,
+ "text": "hello ",
+ },
+ {
+ "label": True,
+ "text": "hi there.",
+ },
+ {
+ "label": False,
+ "text": "goodbye ",
+ },
+ {
+ "label": True,
+ "text": "farewell",
+ },
+ ]
+ }
+ ]
+ )
+
+
+@pytest.fixture(name="tokenizer")
+def fixture_tokenizer():
+ tokenizer = AutoTokenizer.from_pretrained(
+ "casperhansen/mistral-7b-instruct-v0.1-awq"
+ )
+ tokenizer.add_tokens(
+ [
+ AddedToken("", rstrip=False, lstrip=False, normalized=False),
+ ]
+ )
+
+ return tokenizer
+
+
+class TestRawInputOutputPrompts:
+ """
+ Test class for raw i/o prompter
+ """
+
+ def test_segment_prompts(self, segments_dataset, tokenizer):
+ strategy = RawInputOutputStrategy(
+ RawInputOutputPrompter(),
+ tokenizer,
+ False, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, segments_dataset, process_count=1
+ )
+
+ input_ids = dataset_wrapper[0]["input_ids"]
+ labels = dataset_wrapper[0]["labels"]
+
+ assert (
+ tokenizer.decode(input_ids)
+ == " hello hi there. goodbye farewell"
+ )
+ # fmt: off
+ assert input_ids == [
+ 1, #
+ 6312, # hell
+ 28709, # o
+ 28705, #
+ 12014, # hi
+ 736, # there
+ 28723, # .
+ 32000, #
+ 1179, # good
+ 17664, # bye
+ 28705, #
+ 19111, # fare
+ 5458, # well
+ 32000, #
+ ]
+ # fmt: on
+
+ # fmt: off
+ assert labels == [
+ -100, #
+ -100, # hell
+ -100, # o
+ -100, #
+ 12014, # hi
+ 736, # there
+ 28723, # .
+ 32000, #
+ -100, # good
+ -100, # bye
+ -100, #
+ 19111, # fare
+ 5458, # well
+ 32000, #
+ ]
+ # fmt: on
diff --git a/tests/prompt_strategies/test_sharegpt.py b/tests/prompt_strategies/test_sharegpt.py
new file mode 100644
index 0000000000000000000000000000000000000000..6e690983403af5c000d20cd40ed916e825ef6564
--- /dev/null
+++ b/tests/prompt_strategies/test_sharegpt.py
@@ -0,0 +1,316 @@
+"""
+Test module for sharegpt integration w chatml
+"""
+
+import pytest
+from datasets import Dataset
+from tokenizers import AddedToken
+from transformers import AutoTokenizer
+
+from axolotl.datasets import TokenizedPromptDataset
+from axolotl.prompt_strategies.sharegpt import (
+ GlaiveShareGPTPromptTokenizingStrategy,
+ SimpleShareGPTPromptTokenizingStrategy,
+ register_chatml_template,
+ register_llama3_template,
+)
+from axolotl.prompters import ShareGPTPrompterV2
+
+register_chatml_template()
+register_llama3_template()
+
+
+@pytest.fixture(name="sharegpt_dataset")
+def fixture_sharegpt_dataset():
+ return Dataset.from_list(
+ [
+ {
+ "conversations": [
+ {
+ "from": "system",
+ "value": "repeat",
+ },
+ {
+ "from": "human",
+ "value": "hello",
+ },
+ {
+ "from": "gpt",
+ "value": "hello",
+ },
+ {
+ "from": "human",
+ "value": "goodbye",
+ },
+ {
+ "from": "gpt",
+ "value": "goodbye",
+ },
+ ]
+ }
+ ]
+ )
+
+
+@pytest.fixture(name="glaive_dataset")
+def fixture_sharegpt_glaive_dataset():
+ return Dataset.from_list(
+ [
+ {
+ "system": "SYSTEM: This is a system prompt",
+ "chat": "USER: Can you book a flight for me from New York to London? ASSISTANT: I'm sorry, but I don't have the capability to book flights. <|endoftext|>",
+ }
+ ]
+ )
+
+
+@pytest.fixture(name="multi_role_dataset")
+def fixture_multi_role_dataset():
+ return Dataset.from_list(
+ [
+ {
+ "conversations": [
+ {
+ "from": "system",
+ "value": "use get_weather(city) to get the weather for a city",
+ },
+ {
+ "from": "human",
+ "value": "hello, what's the weather in New York?",
+ },
+ {
+ "from": "gpt",
+ "value": "let me get that for you",
+ },
+ {
+ "from": "tool",
+ "value": "get_weather(New York)",
+ },
+ {
+ "from": "gpt",
+ "value": "the weather in New York is 70 degrees and sunny",
+ },
+ ]
+ }
+ ]
+ )
+
+
+@pytest.fixture(name="tokenizer")
+def fixture_tokenizer():
+ tokenizer = AutoTokenizer.from_pretrained(
+ "casperhansen/mistral-7b-instruct-v0.1-awq"
+ )
+ tokenizer.add_special_tokens(
+ {
+ "eos_token": AddedToken(
+ "<|im_end|>", rstrip=False, lstrip=False, normalized=False
+ )
+ }
+ )
+ tokenizer.add_tokens(
+ [
+ AddedToken("<|im_start|>", rstrip=False, lstrip=False, normalized=False),
+ ]
+ )
+
+ return tokenizer
+
+
+@pytest.fixture(name="llama3_tokenizer")
+def fixture_llama3_tokenizer():
+ tokenizer = AutoTokenizer.from_pretrained("NousResearch/Meta-Llama-3-8B")
+ tokenizer.eos_token = "<|eot_id|>"
+
+ return tokenizer
+
+
+class TestSharegptLlama3:
+ """Test class for ShareGPT style datasets with llama-3 prompts"""
+
+ def test_tokenization(self, sharegpt_dataset, llama3_tokenizer):
+ strategy = SimpleShareGPTPromptTokenizingStrategy(
+ ShareGPTPrompterV2(
+ conversation="llama3",
+ role_key_model=None,
+ role_key_human=None,
+ ),
+ llama3_tokenizer,
+ False, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, sharegpt_dataset, process_count=1
+ )
+
+ input_ids = dataset_wrapper[0]["input_ids"]
+
+ # fmt: off
+ assert input_ids == [
+ 128000, # bos
+ 128006, 9125, 128007, # system header
+ 271, 31724, 128009, # sys prompt, eot
+ 128006, 882, 128007, # user header
+ 271, 15339, 128009, # user prompt eot
+ 128006, 78191, 128007, # assistant header
+ 271, 15339, 128009, # assistant response eot
+ 128006, 882, 128007,
+ 271, 19045, 29474, 128009,
+ 128006, 78191, 128007,
+ 271, 19045, 29474, 128009,
+ ]
+ # fmt: on
+
+
+class TestSharegptChatML:
+ """
+ Test class for sharegpt prompter
+ """
+
+ def test_no_double_im_end(self, sharegpt_dataset, tokenizer):
+ strategy = SimpleShareGPTPromptTokenizingStrategy(
+ ShareGPTPrompterV2(
+ conversation="chatml",
+ role_key_model=None,
+ role_key_human=None,
+ ),
+ tokenizer,
+ False, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, sharegpt_dataset, process_count=1
+ )
+
+ input_ids = dataset_wrapper[0]["input_ids"]
+ # fmt: off
+ assert input_ids == [
+ # 28705, 13, is " \n"
+ 1, # bos
+ 32001, 1587, 13, 25997, 32000, 28705, 13, # system
+ 32001, 2188, 13, 21558, 32000, 28705, 13, # human
+ 32001, 13892, 13, 21558, 32000, 28705, 13, # gpt
+ 32001, 2188, 13, 12684, 17664, 32000, 28705, 13, # human
+ 32001, 13892, 13, 12684, 17664, 32000, 28705, 13, # gpt
+ ]
+ # fmt: on
+
+ def test_w_train_on_input(self, sharegpt_dataset, tokenizer):
+ strategy = SimpleShareGPTPromptTokenizingStrategy(
+ ShareGPTPrompterV2(
+ conversation="chatml",
+ role_key_model=None,
+ role_key_human=None,
+ ),
+ tokenizer,
+ False, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, sharegpt_dataset, process_count=1
+ )
+
+ labels = dataset_wrapper[0]["labels"]
+ # fmt: off
+ assert labels == [
+ -100, # bos
+ -100, -100, -100, -100, -100, -100, -100, # system
+ -100, -100, -100, -100, -100, -100, -100, # human
+ -100, -100, 13, 21558, 32000, 28705, 13, # gpt
+ -100, -100, -100, -100, -100, -100, -100, -100, # human
+ -100, -100, 13, 12684, 17664, 32000, 28705, 13, # gpt
+ ]
+ # fmt: on
+
+ def test_no_train_on_input(self, sharegpt_dataset, tokenizer):
+ strategy = SimpleShareGPTPromptTokenizingStrategy(
+ ShareGPTPrompterV2(
+ conversation="chatml",
+ role_key_model=None,
+ role_key_human=None,
+ ),
+ tokenizer,
+ True, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, sharegpt_dataset, process_count=1
+ )
+
+ labels = dataset_wrapper[0]["labels"]
+ # fmt: off
+ assert labels == [
+ 1, # bos
+ 32001, 1587, 13, 25997, 32000, 28705, 13, # system
+ 32001, 2188, 13, 21558, 32000, 28705, 13, # human
+ 32001, 13892, 13, 21558, 32000, 28705, 13, # gpt
+ 32001, 2188, 13, 12684, 17664, 32000, 28705, 13, # human
+ 32001, 13892, 13, 12684, 17664, 32000, 28705, 13, # gpt
+ ]
+ # fmt: on
+
+ def test_chatml_glaive(self, glaive_dataset, tokenizer):
+ strategy = GlaiveShareGPTPromptTokenizingStrategy(
+ ShareGPTPrompterV2(
+ conversation="chatml",
+ role_key_model=None,
+ role_key_human=None,
+ ),
+ tokenizer,
+ True, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, glaive_dataset, process_count=1
+ )
+
+ labels = dataset_wrapper[0]["labels"]
+ # fmt: off
+ assert labels == [
+ 1, # bos
+ 32001, 1587, 13, 3260, 349, 264, 1587, 11510, 32000, 28705, 13, # system
+ 32001, 2188, 13, 6325, 368, 1820, 264, 9314, 354, 528, 477, 1450, 2726, 298, 4222, 28804, 32000, 28705, 13, # human
+ 32001, 13892, 13, 28737, 28742, 28719, 7371, 28725, 562, 315, 949, 28742, 28707, 506, 272, 21368, 298, 1820, 22447, 28723, 28705, 523, 28766, 416, 1009, 772, 28766, 28767, 32000, 28705, 13 # gpt
+ ]
+ # fmt: on
+
+ def test_multi_role_dataset(self, multi_role_dataset, tokenizer):
+ strategy = SimpleShareGPTPromptTokenizingStrategy(
+ ShareGPTPrompterV2(conversation="chatml", roles={"input": ["tool"]}),
+ tokenizer,
+ False, # train_on_inputs
+ 2048, # sequence_len
+ )
+
+ dataset_wrapper = TokenizedPromptDataset(
+ strategy, multi_role_dataset, process_count=1
+ )
+
+ input_ids = dataset_wrapper[0]["input_ids"]
+ # fmt: off
+ assert input_ids == [
+ 1, # bos
+ 32001, 1587, 13, 1730, 625, 28730, 769, 1223, 28732, 18373, 28731, 298, 625, 272, 8086, 354, 264, 2990, 32000, 28705, 13, # system
+ 32001, 2188, 13, 21558, 28725, 767, 28742, 28713, 272, 8086, 297, 1450, 2726, 28804, 32000, 28705, 13, # human
+ 32001, 13892, 13, 895, 528, 625, 369, 354, 368, 32000, 28705, 13, # gpt
+ 32001, 3921, 13, 527, 28730, 769, 1223, 28732, 2972, 2726, 28731, 32000, 28705, 13, # tool
+ 32001, 13892, 13, 1237, 8086, 297, 1450, 2726, 349, 28705, 28787, 28734, 11182, 304, 4376, 1780, 32000, 28705, 13 # gpt
+ ]
+ # fmt: on
+
+ labels = dataset_wrapper[0]["labels"]
+ # fmt: off
+ assert labels == [
+ -100, # bos
+ -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, # system
+ -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, # human
+ -100, -100, 13, 895, 528, 625, 369, 354, 368, 32000, 28705, 13, # gpt
+ -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, -100, # tool
+ -100, -100, 13, 1237, 8086, 297, 1450, 2726, 349, 28705, 28787, 28734, 11182, 304, 4376, 1780, 32000, 28705, 13 # gpt
+ ]
+ # fmt: on
diff --git a/tests/test_data.py b/tests/test_data.py
new file mode 100644
index 0000000000000000000000000000000000000000..16af089a063880b75b8ad1451de1349b9ffdee63
--- /dev/null
+++ b/tests/test_data.py
@@ -0,0 +1,64 @@
+"""
+test module for the axolotl.utis.data module
+"""
+import unittest
+
+from transformers import LlamaTokenizer
+
+from axolotl.utils.data import encode_pretraining, md5
+
+
+class TestEncodePretraining(unittest.TestCase):
+ """
+ test class for encode pretraining and md5 helper
+ """
+
+ def setUp(self):
+ self.tokenizer = LlamaTokenizer.from_pretrained("huggyllama/llama-7b")
+ self.tokenizer.add_special_tokens(
+ {
+ "eos_token": " ",
+ "bos_token": "",
+ "unk_token": "",
+ "pad_token": "",
+ }
+ )
+ self.max_tokens = 15 # set a small number for easy inspection
+
+ def test_encode_pretraining(self):
+ examples = {
+ "text": [
+ "Hello, world!",
+ "Nice to meet you.",
+ "lorem ipsum dolor sit amet.",
+ "Nice to meet you again!.",
+ "hello, hello",
+ ]
+ }
+ result = encode_pretraining(self.tokenizer, self.max_tokens, examples["text"])
+
+ self.assertEqual(len(result["input_ids"]), 3)
+
+ # Assert the length of input_ids and attention_mask is correct
+ self.assertEqual(len(result["input_ids"][0]), self.max_tokens)
+ self.assertEqual(len(result["attention_mask"][0]), self.max_tokens)
+
+ # Assert EOS and PAD tokens are correctly added
+ # hello world! is 4 tokens
+ self.assertEqual(result["input_ids"][0][0], self.tokenizer.bos_token_id)
+ self.assertEqual(result["input_ids"][0][5], self.tokenizer.eos_token_id)
+ self.assertEqual(result["input_ids"][0][6], self.tokenizer.pad_token_id)
+ # second part, 5 tokens
+ self.assertEqual(result["input_ids"][0][7], self.tokenizer.bos_token_id)
+ self.assertEqual(result["input_ids"][0][13], self.tokenizer.eos_token_id)
+ self.assertEqual(result["input_ids"][0][14], self.tokenizer.pad_token_id)
+
+ def test_md5(self):
+ self.assertEqual(md5("hello world"), "5eb63bbbe01eeed093cb22bb8f5acdc3")
+ self.assertEqual(
+ md5("hello world", "utf-8"), "5eb63bbbe01eeed093cb22bb8f5acdc3"
+ )
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/test_datasets.py b/tests/test_datasets.py
new file mode 100644
index 0000000000000000000000000000000000000000..a274b7b894af9dc8eaf880d854ba90c529b613ad
--- /dev/null
+++ b/tests/test_datasets.py
@@ -0,0 +1,272 @@
+"""
+Test dataset loading under various conditions.
+"""
+
+import shutil
+import tempfile
+import unittest
+from pathlib import Path
+
+from datasets import Dataset
+from huggingface_hub import snapshot_download
+from transformers import AutoTokenizer
+
+from axolotl.utils.data import load_tokenized_prepared_datasets
+from axolotl.utils.dict import DictDefault
+
+
+class TestDatasetPreparation(unittest.TestCase):
+ """Test a configured dataloader."""
+
+ def setUp(self) -> None:
+ self.tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b")
+ self.tokenizer.add_special_tokens(
+ {
+ "bos_token": "",
+ "eos_token": " ",
+ "unk_token": "",
+ }
+ )
+ # Alpaca dataset.
+ self.dataset = Dataset.from_list(
+ [
+ {
+ "instruction": "Evaluate this sentence for spelling and grammar mistakes",
+ "input": "He finnished his meal and left the resturant",
+ "output": "He finished his meal and left the restaurant.",
+ }
+ ]
+ )
+
+ def test_load_hub(self):
+ """Core use case. Verify that processing data from the hub works"""
+ with tempfile.TemporaryDirectory() as tmp_dir:
+ prepared_path = Path(tmp_dir) / "prepared"
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "sequence_len": 1024,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ },
+ ],
+ }
+ )
+
+ dataset, _ = load_tokenized_prepared_datasets(
+ self.tokenizer, cfg, prepared_path
+ )
+
+ assert len(dataset) == 2000
+ assert "input_ids" in dataset.features
+ assert "attention_mask" in dataset.features
+ assert "labels" in dataset.features
+
+ def test_load_local_hub(self):
+ """Niche use case. Verify that a local copy of a hub dataset can be loaded"""
+ with tempfile.TemporaryDirectory() as tmp_dir:
+ tmp_ds_path = Path("mhenrichsen/alpaca_2k_test")
+ tmp_ds_path.mkdir(parents=True, exist_ok=True)
+ snapshot_download(
+ repo_id="mhenrichsen/alpaca_2k_test",
+ repo_type="dataset",
+ local_dir=tmp_ds_path,
+ )
+
+ prepared_path = Path(tmp_dir) / "prepared"
+ # Right now a local copy that doesn't fully conform to a dataset
+ # must list data_files and ds_type otherwise the loader won't know
+ # how to load it.
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "sequence_len": 1024,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "ds_type": "parquet",
+ "type": "alpaca",
+ "data_files": [
+ "mhenrichsen/alpaca_2k_test/alpaca_2000.parquet",
+ ],
+ },
+ ],
+ }
+ )
+
+ dataset, _ = load_tokenized_prepared_datasets(
+ self.tokenizer, cfg, prepared_path
+ )
+
+ assert len(dataset) == 2000
+ assert "input_ids" in dataset.features
+ assert "attention_mask" in dataset.features
+ assert "labels" in dataset.features
+ shutil.rmtree(tmp_ds_path)
+
+ def test_load_from_save_to_disk(self):
+ """Usual use case. Verify datasets saved via `save_to_disk` can be loaded."""
+ with tempfile.TemporaryDirectory() as tmp_dir:
+ tmp_ds_name = Path(tmp_dir) / "tmp_dataset"
+ self.dataset.save_to_disk(str(tmp_ds_name))
+
+ prepared_path = Path(tmp_dir) / "prepared"
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "sequence_len": 256,
+ "datasets": [
+ {
+ "path": str(tmp_ds_name),
+ "type": "alpaca",
+ },
+ ],
+ }
+ )
+
+ dataset, _ = load_tokenized_prepared_datasets(
+ self.tokenizer, cfg, prepared_path
+ )
+
+ assert len(dataset) == 1
+ assert "input_ids" in dataset.features
+ assert "attention_mask" in dataset.features
+ assert "labels" in dataset.features
+
+ def test_load_from_dir_of_parquet(self):
+ """Usual use case. Verify a directory of parquet files can be loaded."""
+ with tempfile.TemporaryDirectory() as tmp_dir:
+ tmp_ds_dir = Path(tmp_dir) / "tmp_dataset"
+ tmp_ds_dir.mkdir()
+ tmp_ds_path = tmp_ds_dir / "shard1.parquet"
+ self.dataset.to_parquet(tmp_ds_path)
+
+ prepared_path: Path = Path(tmp_dir) / "prepared"
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "sequence_len": 256,
+ "datasets": [
+ {
+ "path": str(tmp_ds_dir),
+ "ds_type": "parquet",
+ "name": "test_data",
+ "data_files": [
+ str(tmp_ds_path),
+ ],
+ "type": "alpaca",
+ },
+ ],
+ }
+ )
+
+ dataset, _ = load_tokenized_prepared_datasets(
+ self.tokenizer, cfg, prepared_path
+ )
+
+ assert len(dataset) == 1
+ assert "input_ids" in dataset.features
+ assert "attention_mask" in dataset.features
+ assert "labels" in dataset.features
+
+ def test_load_from_dir_of_json(self):
+ """Standard use case. Verify a directory of json files can be loaded."""
+ with tempfile.TemporaryDirectory() as tmp_dir:
+ tmp_ds_dir = Path(tmp_dir) / "tmp_dataset"
+ tmp_ds_dir.mkdir()
+ tmp_ds_path = tmp_ds_dir / "shard1.json"
+ self.dataset.to_json(tmp_ds_path)
+
+ prepared_path: Path = Path(tmp_dir) / "prepared"
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "sequence_len": 256,
+ "datasets": [
+ {
+ "path": str(tmp_ds_dir),
+ "ds_type": "json",
+ "name": "test_data",
+ "data_files": [
+ str(tmp_ds_path),
+ ],
+ "type": "alpaca",
+ },
+ ],
+ }
+ )
+
+ dataset, _ = load_tokenized_prepared_datasets(
+ self.tokenizer, cfg, prepared_path
+ )
+
+ assert len(dataset) == 1
+ assert "input_ids" in dataset.features
+ assert "attention_mask" in dataset.features
+ assert "labels" in dataset.features
+
+ def test_load_from_single_parquet(self):
+ """Standard use case. Verify a single parquet file can be loaded."""
+ with tempfile.TemporaryDirectory() as tmp_dir:
+ tmp_ds_path = Path(tmp_dir) / "tmp_dataset.parquet"
+ self.dataset.to_parquet(tmp_ds_path)
+
+ prepared_path: Path = Path(tmp_dir) / "prepared"
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "sequence_len": 256,
+ "datasets": [
+ {
+ "path": str(tmp_ds_path),
+ "name": "test_data",
+ "type": "alpaca",
+ },
+ ],
+ }
+ )
+
+ dataset, _ = load_tokenized_prepared_datasets(
+ self.tokenizer, cfg, prepared_path
+ )
+
+ assert len(dataset) == 1
+ assert "input_ids" in dataset.features
+ assert "attention_mask" in dataset.features
+ assert "labels" in dataset.features
+
+ def test_load_from_single_json(self):
+ """Standard use case. Verify a single json file can be loaded."""
+ with tempfile.TemporaryDirectory() as tmp_dir:
+ tmp_ds_path = Path(tmp_dir) / "tmp_dataset.json"
+ self.dataset.to_json(tmp_ds_path)
+
+ prepared_path: Path = Path(tmp_dir) / "prepared"
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "sequence_len": 256,
+ "datasets": [
+ {
+ "path": str(tmp_ds_path),
+ "name": "test_data",
+ "type": "alpaca",
+ },
+ ],
+ }
+ )
+
+ dataset, _ = load_tokenized_prepared_datasets(
+ self.tokenizer, cfg, prepared_path
+ )
+
+ assert len(dataset) == 1
+ assert "input_ids" in dataset.features
+ assert "attention_mask" in dataset.features
+ assert "labels" in dataset.features
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/test_dict.py b/tests/test_dict.py
new file mode 100644
index 0000000000000000000000000000000000000000..2007cb085e17226249e60c6994f147b3cbc767ca
--- /dev/null
+++ b/tests/test_dict.py
@@ -0,0 +1,104 @@
+"""Module for testing DictDefault class"""
+
+
+import unittest
+
+import pytest
+
+from axolotl.utils.dict import DictDefault
+
+
+class DictDefaultTest(unittest.TestCase):
+ """
+ Test DictDefault class
+ """
+
+ def test_dict_default(self):
+ cfg = DictDefault(
+ {
+ "key_a": {"key_b": "value_a"},
+ "key_c": "value_c",
+ "key_d": ["value_d", "value_e"],
+ }
+ )
+
+ assert (
+ cfg.key_a.key_b == "value_a"
+ ), "DictDefault should return value for existing nested keys"
+
+ assert (
+ cfg.key_c == "value_c"
+ ), "DictDefault should return value for existing keys"
+
+ assert (
+ cfg.key_d[0] == "value_d"
+ ), "DictDefault should return value for existing keys in list"
+
+ assert (
+ "value_e" in cfg.key_d
+ ), "DictDefault should support in operator for existing keys in list"
+
+ def test_dict_or_operator(self):
+ cfg = DictDefault({"key_a": {"key_b": "value_b"}, "key_f": "value_g"})
+
+ cfg = cfg | DictDefault( # pylint: disable=unsupported-binary-operation
+ {
+ "key_a": {"key_b": "value_a"},
+ "key_c": "value_c",
+ "key_d": ["value_d", "value_e"],
+ "key_f": "value_f",
+ }
+ )
+
+ assert (
+ cfg.key_a.key_b == "value_b"
+ ), "DictDefault should support OR operator for existing nested keys"
+
+ assert cfg.key_c == "value_c", "DictDefault should not delete existing key"
+
+ assert cfg.key_d == [
+ "value_d",
+ "value_e",
+ ], "DictDefault should not overwrite existing keys in list"
+
+ assert (
+ cfg.key_f == "value_g"
+ ), "DictDefault should support OR operator for existing key"
+
+ def test_dict_missingkey(self):
+ cfg = DictDefault({})
+
+ assert cfg.random_key is None, "DictDefault should return None for missing keys"
+
+ def test_dict_or(self):
+ cfg = DictDefault({}) | DictDefault({})
+
+ assert (
+ cfg.random_key is None
+ ), "DictDefault should return None for missing keys after | operation"
+
+ def test_dict_nested_missingparentkey(self):
+ """
+ Due to subclassing Dict, DictDefault will error if we try to access a nested key whose parent key does not exist.
+ """
+ cfg = DictDefault({})
+
+ with pytest.raises(
+ AttributeError,
+ match=r"'NoneType' object has no attribute 'another_random_key'",
+ ):
+ cfg.random_key.another_random_key = "value"
+
+ def test_dict_shorthand_assignment(self):
+ """
+ Shorthand assignment is said to not be supported if subclassed. However, their example raises error instead of None.
+ This test ensures that it is supported for current implementation.
+
+ Ref: https://github.com/mewwts/addict#default-values
+ """
+
+ cfg = DictDefault({"key_a": {"key_b": "value_a"}})
+
+ cfg.key_a.key_b = "value_b"
+
+ assert cfg.key_a.key_b == "value_b", "Shorthand assignment should be supported"
diff --git a/tests/test_expand_mask.py b/tests/test_expand_mask.py
new file mode 100644
index 0000000000000000000000000000000000000000..01241c2958b14460cb5cd3ba91671fc05759dda3
--- /dev/null
+++ b/tests/test_expand_mask.py
@@ -0,0 +1,44 @@
+"""
+Unit tests for the monkey patch for expand mask to handle packed sequences
+"""
+import unittest
+
+import torch
+
+from axolotl.monkeypatch.llama_expand_mask import _expand_mask
+
+
+class TestExpandMask(unittest.TestCase):
+ """
+ Test class for attention mask expansion for packed sequences
+ """
+
+ def test_output(self):
+ mask = torch.tensor([[1, 1, 1, 2], [2, 3, 3, 0]])
+ dtype = torch.float32
+ expected_output = torch.tensor(
+ [
+ [
+ [
+ [0.0000e00, -3.4028e38, -3.4028e38, -3.4028e38],
+ [0.0000e00, 0.0000e00, -3.4028e38, -3.4028e38],
+ [0.0000e00, 0.0000e00, 0.0000e00, -3.4028e38],
+ [-3.4028e38, -3.4028e38, -3.4028e38, 0.0000e00],
+ ]
+ ],
+ [
+ [
+ [0.0000e00, -3.4028e38, -3.4028e38, -3.4028e38],
+ [-3.4028e38, 0.0000e00, -3.4028e38, -3.4028e38],
+ [-3.4028e38, 0.0000e00, 0.0000e00, -3.4028e38],
+ [-3.4028e38, -3.4028e38, -3.4028e38, -3.4028e38],
+ ]
+ ],
+ ]
+ )
+ # Check that the output matches the expected output
+ self.assertTrue(torch.allclose(_expand_mask(mask, dtype), expected_output))
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/test_freeze.py b/tests/test_freeze.py
new file mode 100644
index 0000000000000000000000000000000000000000..49d30ba5face906bb152667a85e2b4a40e687d0a
--- /dev/null
+++ b/tests/test_freeze.py
@@ -0,0 +1,285 @@
+"""
+This module contains unit tests for the `freeze_layers_except` function.
+
+The `freeze_layers_except` function is used to freeze layers in a model, except for the specified layers.
+The unit tests in this module verify the behavior of the `freeze_layers_except` function in different scenarios.
+"""
+
+import unittest
+
+import torch
+from torch import nn
+
+from axolotl.utils.freeze import freeze_layers_except
+
+ZERO = [0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0]
+ONE_TO_TEN = [1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0, 9.0, 10.0]
+
+
+class TestFreezeLayersExcept(unittest.TestCase):
+ """
+ A test case class for the `freeze_layers_except` function.
+ """
+
+ def setUp(self):
+ self.model = _TestModel()
+
+ def test_freeze_layers_with_dots_in_name(self):
+ freeze_layers_except(self.model, ["features.layer"])
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ def test_freeze_layers_without_dots_in_name(self):
+ freeze_layers_except(self.model, ["classifier"])
+ self.assertFalse(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertTrue(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ def test_freeze_layers_regex_patterns(self):
+ # The second pattern cannot match because only characters 'a' to 'c' are allowed after the word 'class', whereas it should be matching the character 'i'.
+ freeze_layers_except(self.model, [r"^features.[a-z]+.weight$", r"class[a-c]+"])
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ def test_all_layers_frozen(self):
+ freeze_layers_except(self.model, [])
+ self.assertFalse(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be frozen.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ def test_all_layers_unfrozen(self):
+ freeze_layers_except(self.model, ["features.layer", "classifier"])
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertTrue(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be trainable.",
+ )
+
+ def test_freeze_layers_with_range_pattern_start_end(self):
+ freeze_layers_except(self.model, ["features.layer[1:5]"])
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ self._assert_gradient_output(
+ [
+ ZERO,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ZERO,
+ ZERO,
+ ZERO,
+ ZERO,
+ ZERO,
+ ]
+ )
+
+ def test_freeze_layers_with_range_pattern_single_index(self):
+ freeze_layers_except(self.model, ["features.layer[5]"])
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ self._assert_gradient_output(
+ [ZERO, ZERO, ZERO, ZERO, ZERO, ONE_TO_TEN, ZERO, ZERO, ZERO, ZERO]
+ )
+
+ def test_freeze_layers_with_range_pattern_start_omitted(self):
+ freeze_layers_except(self.model, ["features.layer[:5]"])
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ self._assert_gradient_output(
+ [
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ZERO,
+ ZERO,
+ ZERO,
+ ZERO,
+ ZERO,
+ ]
+ )
+
+ def test_freeze_layers_with_range_pattern_end_omitted(self):
+ freeze_layers_except(self.model, ["features.layer[4:]"])
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ self._assert_gradient_output(
+ [
+ ZERO,
+ ZERO,
+ ZERO,
+ ZERO,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ]
+ )
+
+ def test_freeze_layers_with_range_pattern_merge_included(self):
+ freeze_layers_except(self.model, ["features.layer[4:]", "features.layer[5:6]"])
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ self._assert_gradient_output(
+ [
+ ZERO,
+ ZERO,
+ ZERO,
+ ZERO,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ]
+ )
+
+ def test_freeze_layers_with_range_pattern_merge_intersect(self):
+ freeze_layers_except(self.model, ["features.layer[4:7]", "features.layer[6:8]"])
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ self._assert_gradient_output(
+ [
+ ZERO,
+ ZERO,
+ ZERO,
+ ZERO,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ONE_TO_TEN,
+ ZERO,
+ ZERO,
+ ]
+ )
+
+ def test_freeze_layers_with_range_pattern_merge_separate(self):
+ freeze_layers_except(
+ self.model,
+ ["features.layer[1:2]", "features.layer[3:4]", "features.layer[5:6]"],
+ )
+ self.assertTrue(
+ self.model.features.layer.weight.requires_grad,
+ "model.features.layer should be trainable.",
+ )
+ self.assertFalse(
+ self.model.classifier.weight.requires_grad,
+ "model.classifier should be frozen.",
+ )
+
+ self._assert_gradient_output(
+ [
+ ZERO,
+ ONE_TO_TEN,
+ ZERO,
+ ONE_TO_TEN,
+ ZERO,
+ ONE_TO_TEN,
+ ZERO,
+ ZERO,
+ ZERO,
+ ZERO,
+ ]
+ )
+
+ def _assert_gradient_output(self, expected):
+ input_tensor = torch.tensor([ONE_TO_TEN], dtype=torch.float32)
+
+ self.model.features.layer.weight.grad = None # Reset gradients
+ output = self.model.features.layer(input_tensor)
+ loss = output.sum()
+ loss.backward()
+
+ expected_grads = torch.tensor(expected)
+ torch.testing.assert_close(
+ self.model.features.layer.weight.grad, expected_grads
+ )
+
+
+class _SubLayerModule(nn.Module):
+ def __init__(self):
+ super().__init__()
+ self.layer = nn.Linear(10, 10)
+
+
+class _TestModel(nn.Module):
+ def __init__(self):
+ super().__init__()
+ self.features = _SubLayerModule()
+ self.classifier = nn.Linear(10, 2)
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/test_normalize_config.py b/tests/test_normalize_config.py
new file mode 100644
index 0000000000000000000000000000000000000000..2e76ceb45d336c961b2e886c77b7b14df53a3394
--- /dev/null
+++ b/tests/test_normalize_config.py
@@ -0,0 +1,91 @@
+"""
+Test classes for checking functionality of the cfg normalization
+"""
+import unittest
+from unittest.mock import patch
+
+from axolotl.utils.config import normalize_cfg_datasets, normalize_config
+from axolotl.utils.dict import DictDefault
+
+
+class NormalizeConfigTestCase(unittest.TestCase):
+ """
+ test class for normalize_config checks
+ """
+
+ def _get_base_cfg(self):
+ return DictDefault(
+ {
+ "base_model": "JackFram/llama-68m",
+ "base_model_config": "JackFram/llama-68m",
+ "tokenizer_type": "LlamaTokenizer",
+ "num_epochs": 1,
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ }
+ )
+
+ def test_base_model_config_set_when_empty(self):
+ cfg = self._get_base_cfg()
+ del cfg.base_model_config
+ normalize_config(cfg)
+
+ assert cfg.base_model_config == cfg.base_model
+
+ def test_chat_template_chatml(self):
+ cfg = DictDefault(
+ {
+ "chat_template": "chatml",
+ "datasets": [
+ {
+ "path": "lorem/ipsum",
+ "type": "sharegpt",
+ "conversation": "vicuna_v1.1",
+ },
+ {
+ "path": "sit/amet",
+ "type": "sharegpt",
+ },
+ ],
+ }
+ )
+
+ normalize_cfg_datasets(cfg)
+
+ assert cfg.datasets[0].conversation == "vicuna_v1.1"
+ assert cfg.datasets[1].conversation == "chatml"
+
+ @patch("axolotl.utils.config.is_torch_bf16_gpu_available")
+ def test_bf16_auto_setter_available(self, mock_bf16_avail):
+ cfg = self._get_base_cfg()
+ cfg.bf16 = "auto"
+ mock_bf16_avail.return_value = True
+
+ normalize_config(cfg)
+
+ self.assertTrue(cfg.bf16)
+ self.assertFalse(cfg.fp16)
+
+ @patch("axolotl.utils.config.is_torch_bf16_gpu_available")
+ def test_bf16_auto_setter_not_available(self, mock_bf16_avail):
+ cfg = self._get_base_cfg()
+ cfg.bf16 = "auto"
+ cfg.fp16 = None
+ mock_bf16_avail.return_value = False
+
+ normalize_config(cfg)
+
+ self.assertFalse(cfg.bf16)
+ self.assertTrue(cfg.fp16)
+
+ @patch("axolotl.utils.config.is_torch_bf16_gpu_available")
+ def test_bf16_disables_fp16(self, mock_bf16_avail):
+ cfg = self._get_base_cfg()
+ cfg.bf16 = True
+ cfg.fp16 = False
+ mock_bf16_avail.return_value = True
+
+ normalize_config(cfg)
+
+ self.assertTrue(cfg.bf16)
+ self.assertFalse(cfg.fp16)
diff --git a/tests/test_packed_batch_sampler.py b/tests/test_packed_batch_sampler.py
new file mode 100644
index 0000000000000000000000000000000000000000..50f39d60f5f0a09a05796ba87592dc5f426b4446
--- /dev/null
+++ b/tests/test_packed_batch_sampler.py
@@ -0,0 +1,99 @@
+"""Module for testing streaming dataset sequence packing"""
+import pytest
+from datasets import concatenate_datasets, load_dataset
+from torch.utils.data import DataLoader, RandomSampler
+from transformers import AutoTokenizer
+
+from axolotl.datasets import TokenizedPromptDataset
+from axolotl.prompt_strategies.completion import load
+from axolotl.utils.collators import V2BatchSamplerDataCollatorForSeq2Seq
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.samplers import MultipackBatchSampler, get_dataset_lengths
+
+
+@pytest.fixture(name="tokenizer")
+def fixture_tokenizer():
+ tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b")
+ tokenizer.pad_token = " "
+ return tokenizer
+
+
+@pytest.fixture(name="max_seq_length")
+def fixture_max_seq_length():
+ return 4096
+
+
+class TestBatchedSamplerPacking:
+ """
+ Test class for packing streaming dataset sequences
+ """
+
+ @pytest.mark.parametrize(
+ "batch_size, num_workers",
+ [
+ (1, 0),
+ (2, 0),
+ (1, 2),
+ (2, 2),
+ ],
+ )
+ def test_packing(self, batch_size, num_workers, tokenizer, max_seq_length):
+ import axolotl.monkeypatch.data.batch_dataset_fetcher # pylint: disable=unused-import # noqa: F401
+
+ dataset = load_dataset(
+ "Trelis/tiny-shakespeare",
+ split="train",
+ )
+
+ cfg = DictDefault(
+ {
+ "train_on_inputs": True,
+ "sequence_len": max_seq_length,
+ }
+ )
+ ds_cfg = DictDefault(
+ {
+ "field": "Text",
+ }
+ )
+ completion_strategy = load(tokenizer, cfg, ds_cfg)
+ dataset_wrapper = TokenizedPromptDataset(
+ completion_strategy,
+ dataset,
+ )
+ train_dataset = concatenate_datasets([dataset_wrapper])
+ batch_sampler = MultipackBatchSampler(
+ sampler=RandomSampler(train_dataset),
+ batch_size=batch_size,
+ drop_last=True,
+ batch_max_len=max_seq_length,
+ lengths=get_dataset_lengths(train_dataset),
+ )
+
+ loader = DataLoader(
+ train_dataset,
+ batch_sampler=batch_sampler,
+ collate_fn=V2BatchSamplerDataCollatorForSeq2Seq( # pylint: disable=unexpected-keyword-arg
+ tokenizer=tokenizer,
+ padding=True,
+ pad_to_multiple_of=max_seq_length,
+ return_tensors="pt",
+ ),
+ num_workers=num_workers,
+ )
+ inputs = next(iter(loader))
+
+ assert inputs["input_ids"].shape == (batch_size, max_seq_length)
+ assert inputs["labels"].shape == (batch_size, max_seq_length)
+ assert inputs["attention_mask"].shape == (batch_size, max_seq_length)
+
+ assert inputs["input_ids"].tolist()[0][0] == 2
+ assert inputs["labels"].tolist()[0][0] == -100
+ assert inputs["attention_mask"].tolist()[0][0] == 0
+ assert inputs["attention_mask"].tolist()[0][-1] > 1
+
+ if batch_size >= 2:
+ assert inputs["input_ids"].tolist()[1][0] == 2
+ assert inputs["labels"].tolist()[1][0] == -100
+ assert inputs["attention_mask"].tolist()[1][0] == 0
+ assert inputs["attention_mask"].tolist()[1][-1] > 1
diff --git a/tests/test_packed_dataset.py b/tests/test_packed_dataset.py
new file mode 100644
index 0000000000000000000000000000000000000000..da8fb7a937c37b77ac4acf034397cf49ef4bb0dc
--- /dev/null
+++ b/tests/test_packed_dataset.py
@@ -0,0 +1,69 @@
+"""Module for testing dataset sequence packing"""
+
+import unittest
+from pathlib import Path
+
+from datasets import Dataset, load_dataset
+from transformers import AutoTokenizer
+
+from axolotl.datasets import ConstantLengthDataset, TokenizedPromptDataset
+from axolotl.prompt_tokenizers import AlpacaPromptTokenizingStrategy
+from axolotl.prompters import AlpacaPrompter
+
+
+class TestPacking(unittest.TestCase):
+ """
+ Test class for packing dataset sequences
+ """
+
+ def setUp(self) -> None:
+ # pylint: disable=duplicate-code
+ self.tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b")
+ self.tokenizer.add_special_tokens(
+ {
+ "bos_token": "",
+ "eos_token": " ",
+ "unk_token": "",
+ }
+ )
+
+ def test_increments_attention(self):
+ prompter = AlpacaPrompter("chat")
+ strat = AlpacaPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ dateset = load_dataset(
+ "json",
+ data_files=str(Path(__file__).parent / "fixtures/alpaca/alpaca.json"),
+ )["train"]
+ dataset = Dataset.from_list(list(TokenizedPromptDataset(strat, dateset)))
+
+ constant_len_dataset = ConstantLengthDataset(
+ self.tokenizer,
+ [dataset],
+ seq_length=2048,
+ )
+ packed_dataset = Dataset.from_list(list(constant_len_dataset))
+ example = packed_dataset[0]
+ next_bos_index = (
+ example["input_ids"][1:].index(self.tokenizer.bos_token_id) + 1
+ ) # add one since we sliced
+
+ # first example doesn't have mask reset
+ assert example["input_ids"][0] == self.tokenizer.bos_token_id
+ assert example["attention_mask"][0] == 1
+ assert example["position_ids"][0] == 0
+ assert example["position_ids"][1] == 1
+
+ # but subsequent one does
+ assert example["input_ids"][next_bos_index] == self.tokenizer.bos_token_id
+ assert example["attention_mask"][next_bos_index] == 2
+ assert example["position_ids"][next_bos_index] == 0
+ assert example["position_ids"][next_bos_index + 1] == 1
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/test_packed_pretraining.py b/tests/test_packed_pretraining.py
new file mode 100644
index 0000000000000000000000000000000000000000..528f9c8074842a31126e446bb8ec642b3033a9e1
--- /dev/null
+++ b/tests/test_packed_pretraining.py
@@ -0,0 +1,93 @@
+"""Module for testing streaming dataset sequence packing"""
+import functools
+import unittest
+
+import torch
+from datasets import load_dataset
+from torch.utils.data import DataLoader
+from transformers import AutoTokenizer
+
+from axolotl.utils.data import get_dataset_wrapper, wrap_pretraining_dataset
+from axolotl.utils.dict import DictDefault
+
+
+class TestPretrainingPacking(unittest.TestCase):
+ """
+ Test class for packing streaming dataset sequences
+ """
+
+ def setUp(self) -> None:
+ # pylint: disable=duplicate-code
+ self.tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b")
+ self.tokenizer.pad_token = " "
+
+ def test_packing_stream_dataset(self):
+ # pylint: disable=duplicate-code
+ dataset = load_dataset(
+ "c4",
+ "en",
+ streaming=True,
+ )["train"]
+
+ cfg = DictDefault(
+ {
+ "pretraining_dataset": [
+ {
+ "path": "c4",
+ "name": "en",
+ "type": "pretrain",
+ }
+ ],
+ "sample_packing": True,
+ "pad_to_sequence_len": True,
+ "sequence_len": 2048,
+ "micro_batch_size": 2,
+ }
+ )
+
+ ds_wrapper_partial = functools.partial(
+ get_dataset_wrapper,
+ cfg.pretraining_dataset[0],
+ self.tokenizer,
+ cfg,
+ cfg.pretraining_dataset[0]["type"] or "pretrain",
+ )
+
+ original_bsz = cfg.micro_batch_size
+ train_dataset = wrap_pretraining_dataset(
+ dataset,
+ self.tokenizer,
+ cfg,
+ ds_wrapper_partial,
+ max_tokens=cfg.sequence_len,
+ batch_size=cfg.micro_batch_size,
+ seed=cfg.seed or 42,
+ )
+
+ trainer_loader = DataLoader(
+ train_dataset,
+ batch_size=1,
+ collate_fn=None,
+ drop_last=True,
+ )
+ idx = 0
+ for data in trainer_loader:
+ if idx > 10:
+ break
+ assert data["input_ids"].shape == torch.Size(
+ [1, original_bsz * cfg.sequence_len]
+ )
+ assert data["position_ids"].shape == torch.Size(
+ [1, original_bsz * cfg.sequence_len]
+ )
+ assert data["labels"].shape == torch.Size(
+ [1, original_bsz * cfg.sequence_len]
+ )
+ assert data["attention_mask"].shape == torch.Size(
+ [1, original_bsz * cfg.sequence_len]
+ )
+ idx += 1
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/test_prompt_tokenizers.py b/tests/test_prompt_tokenizers.py
new file mode 100644
index 0000000000000000000000000000000000000000..63e9a621bf96eac6756ec6802014f0d85939f092
--- /dev/null
+++ b/tests/test_prompt_tokenizers.py
@@ -0,0 +1,507 @@
+"""Module for testing prompt tokenizers."""
+
+import json
+import logging
+import unittest
+from copy import deepcopy
+from pathlib import Path
+from typing import Optional
+
+import pytest
+from datasets import load_dataset
+from transformers import AddedToken, AutoTokenizer, LlamaTokenizer
+
+from axolotl.prompt_strategies.alpaca_chat import NoSystemPrompter
+from axolotl.prompt_strategies.alpaca_w_system import (
+ InstructionWSystemPromptTokenizingStrategy,
+ SystemDataPrompter,
+)
+from axolotl.prompt_strategies.llama2_chat import (
+ Llama2ChatPrompter,
+ LLama2ChatTokenizingStrategy,
+)
+from axolotl.prompt_strategies.orpo.chat_template import load
+from axolotl.prompt_strategies.sharegpt import GlaiveShareGPTPromptTokenizingStrategy
+from axolotl.prompt_tokenizers import (
+ AlpacaPromptTokenizingStrategy,
+ ShareGPTPromptTokenizingStrategy,
+)
+from axolotl.prompters import AlpacaPrompter, PromptStyle, ShareGPTPrompterV2
+from axolotl.utils.dict import DictDefault
+
+LOG = logging.getLogger("axolotl")
+
+test_data = {
+ "multi_turn_sys": {
+ "conversations": [
+ {"from": "system", "value": "lorem"},
+ {"from": "human", "value": "abc"},
+ {"from": "gpt", "value": "ipsum"},
+ {"from": "human", "value": "123"},
+ {"from": "gpt", "value": "sit"},
+ ]
+ },
+ "single_turn_sys": {
+ "conversations": [
+ {"from": "system", "value": "lorem"},
+ {"from": "human", "value": "abc"},
+ {"from": "gpt", "value": "ipsum"},
+ ]
+ },
+ "single_turn_no_sys": {
+ "conversations": [
+ {"from": "human", "value": "abc"},
+ {"from": "gpt", "value": "ipsum"},
+ ]
+ },
+ "multi_turn_no_sys": {
+ "conversations": [
+ {"from": "human", "value": "abc"},
+ {"from": "gpt", "value": "ipsum"},
+ {"from": "human", "value": "123"},
+ {"from": "gpt", "value": "sit"},
+ ]
+ },
+}
+
+
+def prompt_strat(conversation, tokenizer):
+ "Helper function to create a prompt strategy for testing."
+ prompter = ShareGPTPrompterV2(conversation=conversation)
+ return ShareGPTPromptTokenizingStrategy(
+ prompter,
+ tokenizer,
+ False,
+ 2048,
+ )
+
+
+class TestPromptTokenizationStrategies(unittest.TestCase):
+ """
+ Test class for prompt tokenization strategies.
+ """
+
+ _caplog: Optional[pytest.LogCaptureFixture] = None
+
+ @pytest.fixture(autouse=True)
+ def inject_fixtures(self, caplog):
+ self._caplog = caplog
+
+ def setUp(self) -> None:
+ # pylint: disable=duplicate-code
+ self.tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b")
+ self.tokenizer.add_special_tokens(
+ {
+ "bos_token": "",
+ "eos_token": " ",
+ "unk_token": "",
+ }
+ )
+
+ def test_sharegpt_integration(self):
+ with open(
+ Path(__file__).parent / "fixtures/conversation.json", encoding="utf-8"
+ ) as fin:
+ data = fin.read()
+ conversation = json.loads(data)
+ with open(
+ Path(__file__).parent / "fixtures/conversation.tokenized.json",
+ encoding="utf-8",
+ ) as fin:
+ data = fin.read()
+ tokenized_conversation = json.loads(data)
+ prompter = ShareGPTPrompterV2()
+ strat = ShareGPTPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ example = strat.tokenize_prompt(conversation)
+ for fields in ["input_ids", "attention_mask", "labels"]:
+ self.assertEqual(len(example[fields]), len(tokenized_conversation[fields]))
+ self.assertEqual(example[fields], tokenized_conversation[fields])
+
+ def test_sharegpt_warnings_integration(self):
+ with open(
+ Path(__file__).parent / "fixtures/conversation.missingturns.json",
+ encoding="utf-8",
+ ) as fin:
+ data = fin.read()
+ conversation = json.loads(data)
+ prompter = ShareGPTPrompterV2()
+ strat = ShareGPTPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ with self._caplog.at_level(logging.WARNING):
+ strat.tokenize_prompt(conversation)
+ assert "assistant turn has empty text" in self._caplog.records[1].message
+
+ def test_sharegpt_warnings_turns(self):
+ conversation = {
+ "conversations": [
+ {"from": "system", "value": "lorem"},
+ {"from": "gpt", "value": "ipsum"},
+ {"from": "human", "value": "dolor"},
+ {"from": "human", "value": "dolor"},
+ {"from": "gpt", "value": "sit"},
+ ]
+ }
+ prompter = ShareGPTPrompterV2()
+ strat = ShareGPTPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ with self._caplog.at_level(logging.WARNING):
+ strat.tokenize_prompt(conversation)
+ assert (
+ "Role did not alternate between turns (gpt and human)"
+ in self._caplog.records[0].message
+ )
+
+ def test_sharegpt_llama(self):
+ "Make sure the sharegpt/llama is tokenized and formatted correctly."
+ strat = prompt_strat("llama-2", self.tokenizer)
+
+ def tokenize(conv):
+ return strat.tokenize_prompt(deepcopy(conv))["input_ids"]
+
+ def decode(ids):
+ return strat.tokenizer.decode(ids)
+
+ # fmt: off
+ # System message, multi-turn conversations
+ mt_ids = tokenize(test_data['multi_turn_sys'])
+ assert decode(mt_ids) == ' [INST] <>\nlorem\n< >\n\nabc [/INST] ipsum [INST] 123 [/INST] sit '
+ assert mt_ids == [1, 518, 25580, 29962, 3532, 14816, 29903, 6778, 13, 29880, 3668, 13, 29966, 829, 14816, 29903, 6778, 13, 13, 10736, 518, 29914, 25580, 29962, 23421, 2, 1, 518, 25580, 29962, 29871, 29896, 29906, 29941, 518, 29914, 25580, 29962, 7845, 2]
+
+ # System message, single-turn conversations
+ st_ids = tokenize(test_data['single_turn_sys'])
+ assert decode(st_ids) == ' [INST] <>\nlorem\n< >\n\nabc [/INST] ipsum '
+ assert st_ids == [1, 518, 25580, 29962, 3532, 14816, 29903, 6778, 13, 29880, 3668, 13, 29966, 829, 14816, 29903, 6778, 13, 13, 10736, 518, 29914, 25580, 29962, 23421, 2]
+
+ # No system message, single-turn
+ ns_ids = tokenize(test_data['single_turn_no_sys'])
+ assert decode(ns_ids) == ' [INST] abc [/INST] ipsum '
+ assert ns_ids == [1, 518, 25580, 29962, 25638, 518, 29914, 25580, 29962, 23421, 2]
+
+ # No system message, multi-turn
+ ns_mt_ids = tokenize(test_data['multi_turn_no_sys'])
+ assert decode(ns_mt_ids) == ' [INST] abc [/INST] ipsum [INST] 123 [/INST] sit '
+ assert ns_mt_ids == [1, 518, 25580, 29962, 25638, 518, 29914, 25580, 29962, 23421, 2, 1, 518, 25580, 29962, 29871, 29896, 29906, 29941, 518, 29914, 25580, 29962, 7845, 2]
+ # fmt: on
+
+ def test_sharegpt_mistral(self):
+ "Make sure the sharegpt/mistral is tokenized and formatted correctly."
+ strat = prompt_strat("mistral", self.tokenizer)
+
+ def tokenize(conv):
+ return strat.tokenize_prompt(deepcopy(conv))["input_ids"]
+
+ def decode(ids):
+ return strat.tokenizer.decode(ids)
+
+ # fmt: off
+ # System message, multi-turn conversations
+ mt_ids = tokenize(test_data['multi_turn_sys'])
+ assert decode(mt_ids) == ' [INST] lorem\nabc [/INST] ipsum [INST] 123 [/INST] sit '
+ assert mt_ids == [1, 518, 25580, 29962, 29871, 301, 3668, 13, 10736, 518, 29914, 25580, 29962, 23421, 2, 518, 25580, 29962, 29871, 29896, 29906, 29941, 518, 29914, 25580, 29962, 7845, 2]
+
+ # System message, single-turn conversations
+ st_ids = tokenize(test_data['single_turn_sys'])
+ assert decode(st_ids) == ' [INST] lorem\nabc [/INST] ipsum '
+ assert st_ids == [1, 518, 25580, 29962, 29871, 301, 3668, 13, 10736, 518, 29914, 25580, 29962, 23421, 2]
+
+ # No system message, single-turn
+ ns_ids = tokenize(test_data['single_turn_no_sys'])
+ assert decode(ns_ids) == ' [INST] abc [/INST] ipsum '
+ assert ns_ids == [1, 518, 25580, 29962, 25638, 518, 29914, 25580, 29962, 23421, 2]
+
+ # No system message, multi-turn
+ ns_mt_ids = tokenize(test_data['multi_turn_no_sys'])
+ assert decode(ns_mt_ids) == ' [INST] abc [/INST] ipsum [INST] 123 [/INST] sit '
+ assert ns_mt_ids == [1, 518, 25580, 29962, 25638, 518, 29914, 25580, 29962, 23421, 2, 518, 25580, 29962, 29871, 29896, 29906, 29941, 518, 29914, 25580, 29962, 7845, 2]
+ # fmt: on
+
+ def test_sharegpt_changes_roles(self):
+ conversation = {
+ "roles": ["USER", "CHARACTER"],
+ "conversations": [
+ {"from": "system", "value": "lorem"},
+ {"from": "gpt", "value": "ipsum"},
+ {"from": "human", "value": "dolor"},
+ {"from": "gpt", "value": "sit"},
+ ],
+ }
+ prompter = ShareGPTPrompterV2()
+ strat = ShareGPTPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ with self._caplog.at_level(logging.WARNING):
+ res = strat.tokenize_prompt(conversation)
+ assert "CHARACTER" in self.tokenizer.decode(res["input_ids"])
+
+ def test_sharegpt_assistant_label_ignore(self):
+ conversation = {
+ "roles": ["user", "assistant"],
+ "conversations": [
+ {"from": "system", "value": "lorem"},
+ {"from": "gpt", "value": "ipsum"},
+ {"from": "human", "value": "dolor"},
+ {"from": "gpt", "value": "sit"},
+ ],
+ }
+ prompter = ShareGPTPrompterV2()
+ strat = ShareGPTPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ with self._caplog.at_level(logging.WARNING):
+ res = strat.tokenize_prompt(conversation)
+ idx = res["input_ids"].index(20255) # assistant token
+ assert res["labels"][idx] == -100
+
+ def test_glaive_tool_label_ignore(self):
+ conversation = {
+ "system": "SYSTEM: This is a system prompt",
+ "chat": "USER: Can you book a flight for me from New York to London? ASSISTANT: I'm sorry, but I don't have the capability to book flights. <|endoftext|>",
+ }
+ prompter = ShareGPTPrompterV2()
+ strat = GlaiveShareGPTPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ with self._caplog.at_level(logging.WARNING):
+ res = strat.tokenize_prompt(conversation)
+ idx = res["input_ids"].index(13566) # assistant token
+ assert res["labels"][idx] == -100
+
+ def test_no_sys_prompt(self):
+ """
+ tests the interface between the user and assistant parts
+ """
+ prompter = NoSystemPrompter()
+ # pylint: disable=duplicate-code
+ strat = AlpacaPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ sample = {
+ "instruction": "hello cruel. lorem ipsum dolor sit amet.",
+ "output": "world!",
+ }
+ example = strat.tokenize_prompt(sample)
+ world_idx = example["input_ids"].index(3186)
+ assert example["labels"][world_idx] == 3186
+ assert example["labels"][world_idx - 1] == -100
+
+ def test_alpaca(self):
+ """
+ tests the interface between the user and assistant parts
+ """
+ # pylint: disable=duplicate-code
+ prompter = AlpacaPrompter()
+ strat = AlpacaPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ sample = {"instruction": "hello!", "output": "Hi! How can I help?"}
+ example = strat.tokenize_prompt(sample)
+ world_idx = example["input_ids"].index(6324)
+ assert example["labels"][world_idx] == 6324
+ assert example["labels"][world_idx - 1] == -100
+
+
+class InstructionWSystemPromptTokenizingStrategyTest(unittest.TestCase):
+ """
+ Test class for prompt tokenization strategies with sys prompt from the dataset
+ """
+
+ def setUp(self) -> None:
+ # pylint: disable=duplicate-code
+ self.tokenizer = AutoTokenizer.from_pretrained("huggyllama/llama-7b")
+ self.tokenizer.add_special_tokens(
+ {
+ "bos_token": "",
+ "eos_token": " ",
+ "unk_token": "",
+ }
+ )
+
+ def test_system_alpaca(self):
+ prompter = SystemDataPrompter(PromptStyle.CHAT.value)
+ strat = InstructionWSystemPromptTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 2048,
+ )
+ sample = {
+ "system": "use cot",
+ "instruction": "hello!",
+ "output": "Hi! How can I help?",
+ }
+ example = strat.tokenize_prompt(sample)
+ assert example["input_ids"][0:5] == [
+ 1,
+ 28962,
+ 1254,
+ 12665,
+ 29901,
+ ] # "SYSTEM:"
+ assert example["input_ids"][5:7] == [671, 20118] # " use cot"
+ assert example["input_ids"][8] == 11889 # USER
+
+
+class Llama2ChatTokenizationTest(unittest.TestCase):
+ """
+ Test class for prompt tokenization strategies with sys prompt from the dataset
+ """
+
+ def setUp(self) -> None:
+ # pylint: disable=duplicate-code
+ self.tokenizer = LlamaTokenizer.from_pretrained("NousResearch/Llama-2-7b-hf")
+ # woraround because official Meta repos are not open
+
+ def test_llama2_chat_integration(self):
+ with open(
+ Path(__file__).parent / "fixtures/conversation.json", encoding="utf-8"
+ ) as fin:
+ data = fin.read()
+ conversation = json.loads(data)
+ with open(
+ Path(__file__).parent / "fixtures/conversation.tokenized_llama2chat.json",
+ encoding="utf-8",
+ ) as fin:
+ data = fin.read()
+ tokenized_conversation = json.loads(data)
+ prompter = Llama2ChatPrompter()
+ strat = LLama2ChatTokenizingStrategy(
+ prompter,
+ self.tokenizer,
+ False,
+ 4096,
+ )
+ example = strat.tokenize_prompt(conversation)
+ for fields in ["input_ids", "attention_mask", "labels"]:
+ self.assertEqual(len(example[fields]), len(tokenized_conversation[fields]))
+ self.assertEqual(example[fields], tokenized_conversation[fields])
+
+ def compare_with_transformers_integration(self):
+ # this needs transformers >= v4.31.0
+ from transformers.models.llama.tokenization_llama import B_SYS, E_SYS
+ from transformers.pipelines.conversational import Conversation
+
+ # from transformers.models.llama.tokenization_llama import DEFAULT_SYSTEM_PROMPT
+ # broken as of 23/7/20
+ # see https://github.com/huggingface/transformers/pull/24935
+ # pylint: disable=C0103
+ DEFAULT_SYSTEM_PROMPT = """\
+You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.
+
+If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information."""
+ with open(
+ Path(__file__).parent / "fixtures/conversation.json", encoding="utf-8"
+ ) as fin:
+ data = fin.read()
+ conversation = json.loads(data)
+ with open(
+ Path(__file__).parent / "fixtures/conversation.tokenized_llama2chat.json",
+ encoding="utf-8",
+ ) as fin:
+ data = fin.read()
+ tokenized_conversation = json.loads(data)
+
+ user_input = []
+ answers = []
+ for msg in conversation["conversations"]:
+ if msg["from"] == "human":
+ user_input.append(msg["value"])
+ else:
+ answers.append(msg["value"])
+ hf_conf = Conversation(
+ text=user_input[-1],
+ past_user_inputs=[B_SYS + DEFAULT_SYSTEM_PROMPT + E_SYS + user_input[0]]
+ + user_input[1:-1],
+ generated_responses=answers,
+ )
+ # pylint: disable=W0212
+ hf_tokens = self.tokenizer._build_conversation_input_ids(hf_conf)
+
+ self.assertEqual(
+ hf_tokens, tokenized_conversation["input_ids"][: len(hf_tokens)]
+ )
+
+
+class OrpoTokenizationTest(unittest.TestCase):
+ """test case for the ORPO tokenization"""
+
+ def setUp(self) -> None:
+ # pylint: disable=duplicate-code
+ tokenizer = LlamaTokenizer.from_pretrained(
+ "casperhansen/mistral-7b-instruct-v0.1-awq"
+ )
+ tokenizer.add_special_tokens(
+ {
+ "eos_token": AddedToken(
+ "<|im_end|>", rstrip=False, lstrip=False, normalized=False
+ )
+ }
+ )
+ tokenizer.add_tokens(
+ [
+ AddedToken(
+ "<|im_start|>", rstrip=False, lstrip=False, normalized=False
+ ),
+ ]
+ )
+ self.tokenizer = tokenizer
+ self.dataset = load_dataset(
+ "argilla/ultrafeedback-binarized-preferences-cleaned", split="train"
+ ).select([0])
+
+ def test_orpo_integration(self):
+ strat = load(
+ self.tokenizer,
+ DictDefault({"train_on_inputs": False}),
+ DictDefault({"chat_template": "chatml"}),
+ )
+ res = strat.tokenize_prompt(self.dataset[0])
+ assert "rejected_input_ids" in res
+ assert "rejected_labels" in res
+ assert "input_ids" in res
+ assert "labels" in res
+ assert "prompt_attention_mask" in res
+
+ assert len(res["rejected_input_ids"]) == len(res["rejected_labels"])
+ assert len(res["input_ids"]) == len(res["labels"])
+ assert len(res["input_ids"]) == len(res["prompt_attention_mask"])
+
+ assert res["rejected_labels"][0] == -100
+ assert res["rejected_input_ids"][-1] == res["rejected_labels"][-1]
+
+ assert res["labels"][0] == -100
+ assert res["input_ids"][-1] == res["labels"][-1]
+
+ assert res["prompt_attention_mask"][0] == 1
+ assert res["prompt_attention_mask"][-1] == 0
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/test_prompters.py b/tests/test_prompters.py
new file mode 100644
index 0000000000000000000000000000000000000000..6c5b8f27c2e9c0582adbe152bfca1e4b581f35b3
--- /dev/null
+++ b/tests/test_prompters.py
@@ -0,0 +1,124 @@
+"""Module testing prompters"""
+
+import unittest
+
+from axolotl.prompt_strategies.alpaca_w_system import SystemDataPrompter
+from axolotl.prompters import (
+ AlpacaPrompter,
+ MultipleChoiceExplainPrompter,
+ PromptStyle,
+ UnpromptedPrompter,
+)
+
+
+class AlpacaPrompterTest(unittest.TestCase):
+ """
+ Test AlpacaPrompter
+ """
+
+ def test_prompt_style_w_none(self):
+ prompter = AlpacaPrompter(prompt_style=None)
+ res = next(prompter.build_prompt("tell me a joke"))
+ # just testing that it uses instruct style
+ assert "### Instruction:" in res
+
+ def test_prompt_style_w_instruct(self):
+ prompter = AlpacaPrompter(prompt_style=PromptStyle.INSTRUCT.value)
+ res = next(
+ prompter.build_prompt("tell me a joke about the following", "alpacas")
+ )
+ assert "Below is an instruction" in res
+ assert "### Instruction:" in res
+ assert "### Input:" in res
+ assert "alpacas" in res
+ assert "### Response:" in res
+ assert "USER:" not in res
+ assert "ASSISTANT:" not in res
+ res = next(prompter.build_prompt("tell me a joke about the following"))
+ assert "Below is an instruction" in res
+ assert "### Instruction:" in res
+ assert "### Input:" not in res
+ assert "### Response:" in res
+ assert "USER:" not in res
+ assert "ASSISTANT:" not in res
+
+ def test_prompt_style_w_chat(self):
+ prompter = AlpacaPrompter(prompt_style=PromptStyle.CHAT.value)
+ res = next(
+ prompter.build_prompt("tell me a joke about the following", "alpacas")
+ )
+ assert "Below is an instruction" in res
+ assert "### Instruction:" not in res
+ assert "### Input:" not in res
+ assert "alpacas" in res
+ assert "### Response:" not in res
+ assert "USER:" in res
+ assert "ASSISTANT:" in res
+ res = next(prompter.build_prompt("tell me a joke about the following"))
+ assert "Below is an instruction" in res
+ assert "### Instruction:" not in res
+ assert "### Input:" not in res
+ assert "### Response:" not in res
+ assert "USER:" in res
+ assert "ASSISTANT:" in res
+
+ def test_system_prompt(self):
+ prompter = SystemDataPrompter(prompt_style=PromptStyle.CHAT.value)
+ res = next(
+ prompter.build_prompt_w_system(
+ "use cot", "tell me a joke about the following", "alpacas"
+ )
+ )
+ assert "use cot" in res
+ assert res.startswith("SYSTEM:")
+ assert "### Instruction:" not in res
+ assert "### Input:" not in res
+ assert "alpacas" in res
+ assert "### Response:" not in res
+ assert "USER:" in res
+ assert "ASSISTANT:" in res
+
+
+class UnpromptedPrompterTest(unittest.TestCase):
+ """
+ Test class for UnpromptedPrompter with no system prompts
+ """
+
+ def test_prompt_style_w_none(self):
+ prompter = UnpromptedPrompter(prompt_style=None)
+ res = next(prompter.build_prompt("tell me a joke"))
+ assert "### Instruction:" in res
+ assert "tell me a joke" in res
+ assert res.startswith("###")
+
+ def test_prompt_style_w_instruct(self):
+ prompter = UnpromptedPrompter(prompt_style=PromptStyle.INSTRUCT.value)
+ res = next(
+ prompter.build_prompt("tell me a joke about the following", "alpacas")
+ )
+ assert "### Instruction:" in res
+ assert "tell me a joke" in res
+ assert res.startswith("###")
+
+ def test_prompt_style_w_chat(self):
+ prompter = UnpromptedPrompter(prompt_style=PromptStyle.CHAT.value)
+ res = next(
+ prompter.build_prompt("tell me a joke about the following", "alpacas")
+ )
+ assert "USER:" in res
+ assert "tell me a joke" in res
+ assert res.startswith("USER:")
+
+
+class MultipleChoiceExplainPrompterTest(unittest.TestCase):
+ """
+ Test class for MultipleChoiceExplainPrompter
+ """
+
+ def test_prompt_style_w_chat(self):
+ prompter = MultipleChoiceExplainPrompter(prompt_style=PromptStyle.CHAT.value)
+ res = next(prompter.build_prompt("choose one", "- A\n- B\n- C", "C"))
+ assert "USER:" in res
+ assert "choose one" in res
+ assert "Choose the answer that best answers the question." in res
+ assert "- A\n- B\n- C" in res
diff --git a/tests/test_schedulers.py b/tests/test_schedulers.py
new file mode 100644
index 0000000000000000000000000000000000000000..9402d7af7fd59f8aff67db13cb3bf6ddbfce3328
--- /dev/null
+++ b/tests/test_schedulers.py
@@ -0,0 +1,52 @@
+"""
+test module for the axolotl.utis.data module
+"""
+import unittest
+
+import torch
+from torch.optim import SGD
+
+from axolotl.utils.schedulers import get_cosine_schedule_with_warmup_decay_constant
+
+
+class TestCosineConstantLr(unittest.TestCase):
+ """
+ test class for encode pretraining and md5 helper
+ """
+
+ def setUp(self):
+ self.train_steps = 1000
+ self.warmup_steps = 10
+ self.min_lr_ratio = 0.1
+ self.constant_lr_ratio = 0.8
+ self._lr = 0.01
+ self.optimizer = SGD([torch.tensor(1)], lr=self._lr)
+ self.lr_scheduler = get_cosine_schedule_with_warmup_decay_constant( # pylint: disable=attribute-defined-outside-init
+ self.optimizer,
+ num_warmup_steps=self.warmup_steps,
+ num_training_steps=self.train_steps,
+ min_lr_ratio=self.min_lr_ratio,
+ constant_lr_ratio=self.constant_lr_ratio,
+ )
+
+ def test_schedulers(self):
+ self.assertEqual(self.lr_scheduler.get_last_lr()[0], 0)
+ for _ in range(self.warmup_steps):
+ self.lr_scheduler.step()
+ self.assertEqual(self.lr_scheduler.get_last_lr()[0], self._lr)
+ constant_step = int(self.train_steps * self.constant_lr_ratio)
+ remaining_step = self.train_steps - constant_step
+ for _ in range(constant_step):
+ self.lr_scheduler.step()
+ self.assertEqual(
+ self.lr_scheduler.get_last_lr()[0], self._lr * self.min_lr_ratio
+ )
+ for _ in range(remaining_step):
+ self.lr_scheduler.step()
+ self.assertEqual(
+ self.lr_scheduler.get_last_lr()[0], self._lr * self.min_lr_ratio
+ )
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/test_tokenizers.py b/tests/test_tokenizers.py
new file mode 100644
index 0000000000000000000000000000000000000000..69c441f8c646e410e01e9040708dbe6c0db50af6
--- /dev/null
+++ b/tests/test_tokenizers.py
@@ -0,0 +1,87 @@
+"""
+Test cases for the tokenizer loading
+"""
+import unittest
+
+import pytest
+
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.models import load_tokenizer
+
+
+class TestTokenizers(unittest.TestCase):
+ """
+ test class for the load_tokenizer fn
+ """
+
+ def test_default_use_fast(self):
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ }
+ )
+ tokenizer = load_tokenizer(cfg)
+ assert "Fast" in tokenizer.__class__.__name__
+
+ def test_dont_use_fast(self):
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "tokenizer_use_fast": False,
+ }
+ )
+ tokenizer = load_tokenizer(cfg)
+ assert "Fast" not in tokenizer.__class__.__name__
+
+ def test_special_tokens_modules_to_save(self):
+ # setting special_tokens to new token
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "adapter": "lora",
+ "special_tokens": {"bos_token": "[INST]"},
+ }
+ )
+ with pytest.raises(
+ ValueError,
+ match=r".*Please set lora_modules_to_save*",
+ ):
+ load_tokenizer(cfg)
+
+ # setting special_tokens but not changing from default
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "adapter": "lora",
+ "special_tokens": {"bos_token": ""},
+ }
+ )
+ load_tokenizer(cfg)
+
+ # non-adapter setting special_tokens
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "special_tokens": {"bos_token": "[INST]"},
+ }
+ )
+ load_tokenizer(cfg)
+
+ def test_add_additional_special_tokens(self):
+ cfg = DictDefault(
+ {
+ "tokenizer_config": "huggyllama/llama-7b",
+ "special_tokens": {"additional_special_tokens": ["<|im_start|>"]},
+ }
+ )
+ tokenizer = load_tokenizer(cfg)
+ self.assertEqual(tokenizer("<|im_start|>user")["input_ids"], [1, 32000, 1404])
+ self.assertEqual(len(tokenizer), 32001)
+
+ # ensure reloading the tokenizer again from cfg results in same vocab length
+ tokenizer = load_tokenizer(cfg)
+ self.assertEqual(len(tokenizer), 32001)
+
+
+if __name__ == "__main__":
+ unittest.main()
diff --git a/tests/test_validation.py b/tests/test_validation.py
new file mode 100644
index 0000000000000000000000000000000000000000..27824f28870fcab07c1003eadb84162fefc2bb47
--- /dev/null
+++ b/tests/test_validation.py
@@ -0,0 +1,1322 @@
+# pylint: disable=too-many-lines
+"""Module for testing the validation module"""
+
+import logging
+import os
+import warnings
+from typing import Optional
+
+import pytest
+from pydantic import ValidationError
+
+from axolotl.utils.config import validate_config
+from axolotl.utils.config.models.input.v0_4_1 import AxolotlConfigWCapabilities
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.models import check_model_config
+from axolotl.utils.wandb_ import setup_wandb_env_vars
+
+warnings.filterwarnings("error")
+
+
+@pytest.fixture(name="minimal_cfg")
+def fixture_cfg():
+ return DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "learning_rate": 0.000001,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ }
+ ],
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ }
+ )
+
+
+class BaseValidation:
+ """
+ Base validation module to setup the log capture
+ """
+
+ _caplog: Optional[pytest.LogCaptureFixture] = None
+
+ @pytest.fixture(autouse=True)
+ def inject_fixtures(self, caplog):
+ self._caplog = caplog
+
+
+# pylint: disable=too-many-public-methods
+class TestValidation(BaseValidation):
+ """
+ Test the validation module
+ """
+
+ def test_defaults(self, minimal_cfg):
+ test_cfg = DictDefault(
+ {
+ "weight_decay": None,
+ }
+ | minimal_cfg
+ )
+ cfg = validate_config(test_cfg)
+
+ assert cfg.train_on_inputs is False
+ assert cfg.weight_decay is None
+
+ def test_datasets_min_length(self):
+ cfg = DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "learning_rate": 0.000001,
+ "datasets": [],
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ }
+ )
+
+ with pytest.raises(
+ ValidationError,
+ match=r".*List should have at least 1 item after validation*",
+ ):
+ validate_config(cfg)
+
+ def test_datasets_min_length_empty(self):
+ cfg = DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "learning_rate": 0.000001,
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ }
+ )
+
+ with pytest.raises(
+ ValueError, match=r".*either datasets or pretraining_dataset is required*"
+ ):
+ validate_config(cfg)
+
+ def test_pretrain_dataset_min_length(self):
+ cfg = DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "learning_rate": 0.000001,
+ "pretraining_dataset": [],
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "max_steps": 100,
+ }
+ )
+
+ with pytest.raises(
+ ValidationError,
+ match=r".*List should have at least 1 item after validation*",
+ ):
+ validate_config(cfg)
+
+ def test_valid_pretrain_dataset(self):
+ cfg = DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "learning_rate": 0.000001,
+ "pretraining_dataset": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ }
+ ],
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ "max_steps": 100,
+ }
+ )
+
+ validate_config(cfg)
+
+ def test_valid_sft_dataset(self):
+ cfg = DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "learning_rate": 0.000001,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ }
+ ],
+ "micro_batch_size": 1,
+ "gradient_accumulation_steps": 1,
+ }
+ )
+
+ validate_config(cfg)
+
+ def test_batch_size_unused_warning(self):
+ cfg = DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "learning_rate": 0.000001,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ }
+ ],
+ "micro_batch_size": 4,
+ "batch_size": 32,
+ }
+ )
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert "batch_size is not recommended" in self._caplog.records[0].message
+
+ def test_batch_size_more_params(self):
+ cfg = DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "learning_rate": 0.000001,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ }
+ ],
+ "batch_size": 32,
+ }
+ )
+
+ with pytest.raises(ValueError, match=r".*At least two of*"):
+ validate_config(cfg)
+
+ def test_lr_as_float(self, minimal_cfg):
+ cfg = (
+ DictDefault( # pylint: disable=unsupported-binary-operation
+ {
+ "learning_rate": "5e-5",
+ }
+ )
+ | minimal_cfg
+ )
+
+ new_cfg = validate_config(cfg)
+
+ assert new_cfg.learning_rate == 0.00005
+
+ def test_model_config_remap(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "model_config": {"model_type": "mistral"},
+ }
+ )
+ | minimal_cfg
+ )
+
+ new_cfg = validate_config(cfg)
+ assert new_cfg.overrides_of_model_config["model_type"] == "mistral"
+
+ def test_model_type_remap(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "model_type": "AutoModelForCausalLM",
+ }
+ )
+ | minimal_cfg
+ )
+
+ new_cfg = validate_config(cfg)
+ assert new_cfg.type_of_model == "AutoModelForCausalLM"
+
+ def test_model_revision_remap(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "model_revision": "main",
+ }
+ )
+ | minimal_cfg
+ )
+
+ new_cfg = validate_config(cfg)
+ assert new_cfg.revision_of_model == "main"
+
+ def test_qlora(self, minimal_cfg):
+ base_cfg = (
+ DictDefault(
+ {
+ "adapter": "qlora",
+ }
+ )
+ | minimal_cfg
+ )
+
+ cfg = (
+ DictDefault( # pylint: disable=unsupported-binary-operation
+ {
+ "load_in_8bit": True,
+ }
+ )
+ | base_cfg
+ )
+
+ with pytest.raises(ValueError, match=r".*8bit.*"):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault( # pylint: disable=unsupported-binary-operation
+ {
+ "gptq": True,
+ }
+ )
+ | base_cfg
+ )
+
+ with pytest.raises(ValueError, match=r".*gptq.*"):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault( # pylint: disable=unsupported-binary-operation
+ {
+ "load_in_4bit": False,
+ }
+ )
+ | base_cfg
+ )
+
+ with pytest.raises(ValueError, match=r".*4bit.*"):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault( # pylint: disable=unsupported-binary-operation
+ {
+ "load_in_4bit": True,
+ }
+ )
+ | base_cfg
+ )
+
+ validate_config(cfg)
+
+ def test_qlora_merge(self, minimal_cfg):
+ base_cfg = (
+ DictDefault(
+ {
+ "adapter": "qlora",
+ "merge_lora": True,
+ }
+ )
+ | minimal_cfg
+ )
+
+ cfg = (
+ DictDefault( # pylint: disable=unsupported-binary-operation
+ {
+ "load_in_8bit": True,
+ }
+ )
+ | base_cfg
+ )
+
+ with pytest.raises(ValueError, match=r".*8bit.*"):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault( # pylint: disable=unsupported-binary-operation
+ {
+ "gptq": True,
+ }
+ )
+ | base_cfg
+ )
+
+ with pytest.raises(ValueError, match=r".*gptq.*"):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault( # pylint: disable=unsupported-binary-operation
+ {
+ "load_in_4bit": True,
+ }
+ )
+ | base_cfg
+ )
+
+ with pytest.raises(ValueError, match=r".*4bit.*"):
+ validate_config(cfg)
+
+ def test_hf_use_auth_token(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "push_dataset_to_hub": "namespace/repo",
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(ValueError, match=r".*hf_use_auth_token.*"):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "push_dataset_to_hub": "namespace/repo",
+ "hf_use_auth_token": True,
+ }
+ )
+ | minimal_cfg
+ )
+ validate_config(cfg)
+
+ def test_gradient_accumulations_or_batch_size(self):
+ cfg = DictDefault(
+ {
+ "base_model": "TinyLlama/TinyLlama-1.1B-Chat-v0.6",
+ "learning_rate": 0.000001,
+ "datasets": [
+ {
+ "path": "mhenrichsen/alpaca_2k_test",
+ "type": "alpaca",
+ }
+ ],
+ "gradient_accumulation_steps": 1,
+ "batch_size": 1,
+ }
+ )
+
+ with pytest.raises(
+ ValueError, match=r".*gradient_accumulation_steps or batch_size.*"
+ ):
+ validate_config(cfg)
+
+ def test_falcon_fsdp(self, minimal_cfg):
+ regex_exp = r".*FSDP is not supported for falcon models.*"
+
+ # Check for lower-case
+ cfg = (
+ DictDefault(
+ {
+ "base_model": "tiiuae/falcon-7b",
+ "fsdp": ["full_shard", "auto_wrap"],
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(ValueError, match=regex_exp):
+ validate_config(cfg)
+
+ # Check for upper-case
+ cfg = (
+ DictDefault(
+ {
+ "base_model": "Falcon-7b",
+ "fsdp": ["full_shard", "auto_wrap"],
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(ValueError, match=regex_exp):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "base_model": "tiiuae/falcon-7b",
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ def test_mpt_gradient_checkpointing(self, minimal_cfg):
+ regex_exp = r".*gradient_checkpointing is not supported for MPT models*"
+
+ # Check for lower-case
+ cfg = (
+ DictDefault(
+ {
+ "base_model": "mosaicml/mpt-7b",
+ "gradient_checkpointing": True,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(ValueError, match=regex_exp):
+ validate_config(cfg)
+
+ def test_flash_optimum(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "flash_optimum": True,
+ "adapter": "lora",
+ "bf16": False,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert any(
+ "BetterTransformers probably doesn't work with PEFT adapters"
+ in record.message
+ for record in self._caplog.records
+ )
+
+ cfg = (
+ DictDefault(
+ {
+ "flash_optimum": True,
+ "bf16": False,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert any(
+ "probably set bfloat16 or float16" in record.message
+ for record in self._caplog.records
+ )
+
+ cfg = (
+ DictDefault(
+ {
+ "flash_optimum": True,
+ "fp16": True,
+ }
+ )
+ | minimal_cfg
+ )
+ regex_exp = r".*AMP is not supported.*"
+
+ with pytest.raises(ValueError, match=regex_exp):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "flash_optimum": True,
+ "bf16": True,
+ }
+ )
+ | minimal_cfg
+ )
+ regex_exp = r".*AMP is not supported.*"
+
+ with pytest.raises(ValueError, match=regex_exp):
+ validate_config(cfg)
+
+ def test_adamw_hyperparams(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "optimizer": None,
+ "adam_epsilon": 0.0001,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert any(
+ "adamw hyperparameters found, but no adamw optimizer set"
+ in record.message
+ for record in self._caplog.records
+ )
+
+ cfg = (
+ DictDefault(
+ {
+ "optimizer": "adafactor",
+ "adam_beta1": 0.0001,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert any(
+ "adamw hyperparameters found, but no adamw optimizer set"
+ in record.message
+ for record in self._caplog.records
+ )
+
+ cfg = (
+ DictDefault(
+ {
+ "optimizer": "adamw_bnb_8bit",
+ "adam_beta1": 0.9,
+ "adam_beta2": 0.99,
+ "adam_epsilon": 0.0001,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "optimizer": "adafactor",
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ def test_deprecated_packing(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "max_packed_sequence_len": 1024,
+ }
+ )
+ | minimal_cfg
+ )
+ with pytest.raises(
+ DeprecationWarning,
+ match=r"`max_packed_sequence_len` is no longer supported",
+ ):
+ validate_config(cfg)
+
+ def test_packing(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "sample_packing": True,
+ "pad_to_sequence_len": None,
+ "flash_attention": True,
+ }
+ )
+ | minimal_cfg
+ )
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert any(
+ "`pad_to_sequence_len: true` is recommended when using sample_packing"
+ in record.message
+ for record in self._caplog.records
+ )
+
+ def test_merge_lora_no_bf16_fail(self, minimal_cfg):
+ """
+ This is assumed to be run on a CPU machine, so bf16 is not supported.
+ """
+
+ cfg = (
+ DictDefault(
+ {
+ "bf16": True,
+ "capabilities": {"bf16": False},
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(ValueError, match=r".*AMP is not supported on this GPU*"):
+ AxolotlConfigWCapabilities(**cfg.to_dict())
+
+ cfg = (
+ DictDefault(
+ {
+ "bf16": True,
+ "merge_lora": True,
+ "capabilities": {"bf16": False},
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ def test_sharegpt_deprecation(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {"datasets": [{"path": "lorem/ipsum", "type": "sharegpt:chat"}]}
+ )
+ | minimal_cfg
+ )
+ with self._caplog.at_level(logging.WARNING):
+ new_cfg = validate_config(cfg)
+ assert any(
+ "`type: sharegpt:chat` will soon be deprecated." in record.message
+ for record in self._caplog.records
+ )
+ assert new_cfg.datasets[0].type == "sharegpt"
+
+ cfg = (
+ DictDefault(
+ {
+ "datasets": [
+ {"path": "lorem/ipsum", "type": "sharegpt_simple:load_role"}
+ ]
+ }
+ )
+ | minimal_cfg
+ )
+ with self._caplog.at_level(logging.WARNING):
+ new_cfg = validate_config(cfg)
+ assert any(
+ "`type: sharegpt_simple` will soon be deprecated." in record.message
+ for record in self._caplog.records
+ )
+ assert new_cfg.datasets[0].type == "sharegpt:load_role"
+
+ def test_no_conflict_save_strategy(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "save_strategy": "epoch",
+ "save_steps": 10,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError, match=r".*save_strategy and save_steps mismatch.*"
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "save_strategy": "no",
+ "save_steps": 10,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError, match=r".*save_strategy and save_steps mismatch.*"
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "save_strategy": "steps",
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "save_strategy": "steps",
+ "save_steps": 10,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "save_steps": 10,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "save_strategy": "no",
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ def test_no_conflict_eval_strategy(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "evaluation_strategy": "epoch",
+ "eval_steps": 10,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError, match=r".*evaluation_strategy and eval_steps mismatch.*"
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "evaluation_strategy": "no",
+ "eval_steps": 10,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError, match=r".*evaluation_strategy and eval_steps mismatch.*"
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "evaluation_strategy": "steps",
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "evaluation_strategy": "steps",
+ "eval_steps": 10,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "eval_steps": 10,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "evaluation_strategy": "no",
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "evaluation_strategy": "epoch",
+ "val_set_size": 0,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError,
+ match=r".*eval_steps and evaluation_strategy are not supported with val_set_size == 0.*",
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "eval_steps": 10,
+ "val_set_size": 0,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError,
+ match=r".*eval_steps and evaluation_strategy are not supported with val_set_size == 0.*",
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "val_set_size": 0,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "eval_steps": 10,
+ "val_set_size": 0.01,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "evaluation_strategy": "epoch",
+ "val_set_size": 0.01,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ def test_eval_table_size_conflict_eval_packing(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "sample_packing": True,
+ "eval_table_size": 100,
+ "flash_attention": True,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError, match=r".*Please set 'eval_sample_packing' to false.*"
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "sample_packing": True,
+ "eval_sample_packing": False,
+ "flash_attention": True,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "sample_packing": False,
+ "eval_table_size": 100,
+ "flash_attention": True,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "sample_packing": True,
+ "eval_table_size": 100,
+ "eval_sample_packing": False,
+ "flash_attention": True,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ def test_load_in_x_bit_without_adapter(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "load_in_4bit": True,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError,
+ match=r".*load_in_8bit and load_in_4bit are not supported without setting an adapter.*",
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "load_in_8bit": True,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError,
+ match=r".*load_in_8bit and load_in_4bit are not supported without setting an adapter.*",
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "load_in_4bit": True,
+ "adapter": "qlora",
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "load_in_8bit": True,
+ "adapter": "lora",
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ def test_warmup_step_no_conflict(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "warmup_steps": 10,
+ "warmup_ratio": 0.1,
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError,
+ match=r".*warmup_steps and warmup_ratio are mutually exclusive*",
+ ):
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "warmup_steps": 10,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ cfg = (
+ DictDefault(
+ {
+ "warmup_ratio": 0.1,
+ }
+ )
+ | minimal_cfg
+ )
+
+ validate_config(cfg)
+
+ def test_unfrozen_parameters_w_peft_layers_to_transform(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "adapter": "lora",
+ "unfrozen_parameters": [
+ "model.layers.2[0-9]+.block_sparse_moe.gate.*"
+ ],
+ "peft_layers_to_transform": [0, 1],
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError,
+ match=r".*can have unexpected behavior*",
+ ):
+ validate_config(cfg)
+
+ def test_hub_model_id_save_value_warns_save_stragey_no(self, minimal_cfg):
+ cfg = DictDefault({"hub_model_id": "test", "save_strategy": "no"}) | minimal_cfg
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert len(self._caplog.records) == 1
+
+ def test_hub_model_id_save_value_warns_random_value(self, minimal_cfg):
+ cfg = (
+ DictDefault({"hub_model_id": "test", "save_strategy": "test"}) | minimal_cfg
+ )
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert len(self._caplog.records) == 1
+
+ def test_hub_model_id_save_value_steps(self, minimal_cfg):
+ cfg = (
+ DictDefault({"hub_model_id": "test", "save_strategy": "steps"})
+ | minimal_cfg
+ )
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert len(self._caplog.records) == 0
+
+ def test_hub_model_id_save_value_epochs(self, minimal_cfg):
+ cfg = (
+ DictDefault({"hub_model_id": "test", "save_strategy": "epoch"})
+ | minimal_cfg
+ )
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert len(self._caplog.records) == 0
+
+ def test_hub_model_id_save_value_none(self, minimal_cfg):
+ cfg = DictDefault({"hub_model_id": "test", "save_strategy": None}) | minimal_cfg
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert len(self._caplog.records) == 0
+
+ def test_hub_model_id_save_value_no_set_save_strategy(self, minimal_cfg):
+ cfg = DictDefault({"hub_model_id": "test"}) | minimal_cfg
+
+ with self._caplog.at_level(logging.WARNING):
+ validate_config(cfg)
+ assert len(self._caplog.records) == 0
+
+
+class TestValidationCheckModelConfig(BaseValidation):
+ """
+ Test the validation for the config when the model config is available
+ """
+
+ def test_llama_add_tokens_adapter(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {"adapter": "qlora", "load_in_4bit": True, "tokens": ["<|imstart|>"]}
+ )
+ | minimal_cfg
+ )
+ model_config = DictDefault({"model_type": "llama"})
+
+ with pytest.raises(
+ ValueError,
+ match=r".*`lora_modules_to_save` not properly set when adding new tokens*",
+ ):
+ check_model_config(cfg, model_config)
+
+ cfg = (
+ DictDefault(
+ {
+ "adapter": "qlora",
+ "load_in_4bit": True,
+ "tokens": ["<|imstart|>"],
+ "lora_modules_to_save": ["embed_tokens"],
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError,
+ match=r".*`lora_modules_to_save` not properly set when adding new tokens*",
+ ):
+ check_model_config(cfg, model_config)
+
+ cfg = (
+ DictDefault(
+ {
+ "adapter": "qlora",
+ "load_in_4bit": True,
+ "tokens": ["<|imstart|>"],
+ "lora_modules_to_save": ["embed_tokens", "lm_head"],
+ }
+ )
+ | minimal_cfg
+ )
+
+ check_model_config(cfg, model_config)
+
+ def test_phi_add_tokens_adapter(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {"adapter": "qlora", "load_in_4bit": True, "tokens": ["<|imstart|>"]}
+ )
+ | minimal_cfg
+ )
+ model_config = DictDefault({"model_type": "phi"})
+
+ with pytest.raises(
+ ValueError,
+ match=r".*`lora_modules_to_save` not properly set when adding new tokens*",
+ ):
+ check_model_config(cfg, model_config)
+
+ cfg = (
+ DictDefault(
+ {
+ "adapter": "qlora",
+ "load_in_4bit": True,
+ "tokens": ["<|imstart|>"],
+ "lora_modules_to_save": ["embd.wte", "lm_head.linear"],
+ }
+ )
+ | minimal_cfg
+ )
+
+ with pytest.raises(
+ ValueError,
+ match=r".*`lora_modules_to_save` not properly set when adding new tokens*",
+ ):
+ check_model_config(cfg, model_config)
+
+ cfg = (
+ DictDefault(
+ {
+ "adapter": "qlora",
+ "load_in_4bit": True,
+ "tokens": ["<|imstart|>"],
+ "lora_modules_to_save": ["embed_tokens", "lm_head"],
+ }
+ )
+ | minimal_cfg
+ )
+
+ check_model_config(cfg, model_config)
+
+
+class TestValidationWandb(BaseValidation):
+ """
+ Validation test for wandb
+ """
+
+ def test_wandb_set_run_id_to_name(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "wandb_run_id": "foo",
+ }
+ )
+ | minimal_cfg
+ )
+
+ with self._caplog.at_level(logging.WARNING):
+ new_cfg = validate_config(cfg)
+ assert any(
+ "wandb_run_id sets the ID of the run. If you would like to set the name, please use wandb_name instead."
+ in record.message
+ for record in self._caplog.records
+ )
+
+ assert new_cfg.wandb_name == "foo" and new_cfg.wandb_run_id == "foo"
+
+ cfg = (
+ DictDefault(
+ {
+ "wandb_name": "foo",
+ }
+ )
+ | minimal_cfg
+ )
+
+ new_cfg = validate_config(cfg)
+
+ assert new_cfg.wandb_name == "foo" and new_cfg.wandb_run_id is None
+
+ def test_wandb_sets_env(self, minimal_cfg):
+ cfg = (
+ DictDefault(
+ {
+ "wandb_project": "foo",
+ "wandb_name": "bar",
+ "wandb_run_id": "bat",
+ "wandb_entity": "baz",
+ "wandb_mode": "online",
+ "wandb_watch": "false",
+ "wandb_log_model": "checkpoint",
+ }
+ )
+ | minimal_cfg
+ )
+
+ new_cfg = validate_config(cfg)
+
+ setup_wandb_env_vars(new_cfg)
+
+ assert os.environ.get("WANDB_PROJECT", "") == "foo"
+ assert os.environ.get("WANDB_NAME", "") == "bar"
+ assert os.environ.get("WANDB_RUN_ID", "") == "bat"
+ assert os.environ.get("WANDB_ENTITY", "") == "baz"
+ assert os.environ.get("WANDB_MODE", "") == "online"
+ assert os.environ.get("WANDB_WATCH", "") == "false"
+ assert os.environ.get("WANDB_LOG_MODEL", "") == "checkpoint"
+ assert os.environ.get("WANDB_DISABLED", "") != "true"
+
+ os.environ.pop("WANDB_PROJECT", None)
+ os.environ.pop("WANDB_NAME", None)
+ os.environ.pop("WANDB_RUN_ID", None)
+ os.environ.pop("WANDB_ENTITY", None)
+ os.environ.pop("WANDB_MODE", None)
+ os.environ.pop("WANDB_WATCH", None)
+ os.environ.pop("WANDB_LOG_MODEL", None)
+ os.environ.pop("WANDB_DISABLED", None)
+
+ def test_wandb_set_disabled(self, minimal_cfg):
+ cfg = DictDefault({}) | minimal_cfg
+
+ new_cfg = validate_config(cfg)
+
+ setup_wandb_env_vars(new_cfg)
+
+ assert os.environ.get("WANDB_DISABLED", "") == "true"
+
+ cfg = (
+ DictDefault(
+ {
+ "wandb_project": "foo",
+ }
+ )
+ | minimal_cfg
+ )
+
+ new_cfg = validate_config(cfg)
+
+ setup_wandb_env_vars(new_cfg)
+
+ assert os.environ.get("WANDB_DISABLED", "") != "true"
+
+ os.environ.pop("WANDB_PROJECT", None)
+ os.environ.pop("WANDB_DISABLED", None)
diff --git a/tests/utils/test_models.py b/tests/utils/test_models.py
new file mode 100644
index 0000000000000000000000000000000000000000..e06bb6c250d3ab73edddc65cb3e83c713e81339a
--- /dev/null
+++ b/tests/utils/test_models.py
@@ -0,0 +1,37 @@
+"""Module for testing models utils file."""
+
+
+import unittest
+from unittest.mock import patch
+
+import pytest
+
+from axolotl.utils.dict import DictDefault
+from axolotl.utils.models import load_model
+
+
+class ModelsUtilsTest(unittest.TestCase):
+ """Testing module for models utils."""
+
+ def test_cfg_throws_error_with_s2_attention_and_sample_packing(self):
+ cfg = DictDefault(
+ {
+ "s2_attention": True,
+ "sample_packing": True,
+ "base_model": "",
+ "model_type": "LlamaForCausalLM",
+ }
+ )
+
+ # Mock out call to HF hub
+ with patch(
+ "axolotl.utils.models.load_model_config"
+ ) as mocked_load_model_config:
+ mocked_load_model_config.return_value = {}
+ with pytest.raises(ValueError) as exc:
+ # Should error before hitting tokenizer, so we pass in an empty str
+ load_model(cfg, tokenizer="")
+ assert (
+ "shifted-sparse attention does not currently support sample packing"
+ in str(exc.value)
+ )
diff --git a/unsloth/LICENSE b/unsloth/LICENSE
new file mode 100644
index 0000000000000000000000000000000000000000..8894f17a3bc6c0b55fa3d18846f1f24dfdd5598b
--- /dev/null
+++ b/unsloth/LICENSE
@@ -0,0 +1,201 @@
+ Apache License
+ Version 2.0, January 2004
+ http://www.apache.org/licenses/
+
+ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
+
+ 1. Definitions.
+
+ "License" shall mean the terms and conditions for use, reproduction,
+ and distribution as defined by Sections 1 through 9 of this document.
+
+ "Licensor" shall mean the copyright owner or entity authorized by
+ the copyright owner that is granting the License.
+
+ "Legal Entity" shall mean the union of the acting entity and all
+ other entities that control, are controlled by, or are under common
+ control with that entity. For the purposes of this definition,
+ "control" means (i) the power, direct or indirect, to cause the
+ direction or management of such entity, whether by contract or
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
+ outstanding shares, or (iii) beneficial ownership of such entity.
+
+ "You" (or "Your") shall mean an individual or Legal Entity
+ exercising permissions granted by this License.
+
+ "Source" form shall mean the preferred form for making modifications,
+ including but not limited to software source code, documentation
+ source, and configuration files.
+
+ "Object" form shall mean any form resulting from mechanical
+ transformation or translation of a Source form, including but
+ not limited to compiled object code, generated documentation,
+ and conversions to other media types.
+
+ "Work" shall mean the work of authorship, whether in Source or
+ Object form, made available under the License, as indicated by a
+ copyright notice that is included in or attached to the work
+ (an example is provided in the Appendix below).
+
+ "Derivative Works" shall mean any work, whether in Source or Object
+ form, that is based on (or derived from) the Work and for which the
+ editorial revisions, annotations, elaborations, or other modifications
+ represent, as a whole, an original work of authorship. For the purposes
+ of this License, Derivative Works shall not include works that remain
+ separable from, or merely link (or bind by name) to the interfaces of,
+ the Work and Derivative Works thereof.
+
+ "Contribution" shall mean any work of authorship, including
+ the original version of the Work and any modifications or additions
+ to that Work or Derivative Works thereof, that is intentionally
+ submitted to Licensor for inclusion in the Work by the copyright owner
+ or by an individual or Legal Entity authorized to submit on behalf of
+ the copyright owner. For the purposes of this definition, "submitted"
+ means any form of electronic, verbal, or written communication sent
+ to the Licensor or its representatives, including but not limited to
+ communication on electronic mailing lists, source code control systems,
+ and issue tracking systems that are managed by, or on behalf of, the
+ Licensor for the purpose of discussing and improving the Work, but
+ excluding communication that is conspicuously marked or otherwise
+ designated in writing by the copyright owner as "Not a Contribution."
+
+ "Contributor" shall mean Licensor and any individual or Legal Entity
+ on behalf of whom a Contribution has been received by Licensor and
+ subsequently incorporated within the Work.
+
+ 2. Grant of Copyright License. Subject to the terms and conditions of
+ this License, each Contributor hereby grants to You a perpetual,
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
+ copyright license to reproduce, prepare Derivative Works of,
+ publicly display, publicly perform, sublicense, and distribute the
+ Work and such Derivative Works in Source or Object form.
+
+ 3. Grant of Patent License. Subject to the terms and conditions of
+ this License, each Contributor hereby grants to You a perpetual,
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
+ (except as stated in this section) patent license to make, have made,
+ use, offer to sell, sell, import, and otherwise transfer the Work,
+ where such license applies only to those patent claims licensable
+ by such Contributor that are necessarily infringed by their
+ Contribution(s) alone or by combination of their Contribution(s)
+ with the Work to which such Contribution(s) was submitted. If You
+ institute patent litigation against any entity (including a
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
+ or a Contribution incorporated within the Work constitutes direct
+ or contributory patent infringement, then any patent licenses
+ granted to You under this License for that Work shall terminate
+ as of the date such litigation is filed.
+
+ 4. Redistribution. You may reproduce and distribute copies of the
+ Work or Derivative Works thereof in any medium, with or without
+ modifications, and in Source or Object form, provided that You
+ meet the following conditions:
+
+ (a) You must give any other recipients of the Work or
+ Derivative Works a copy of this License; and
+
+ (b) You must cause any modified files to carry prominent notices
+ stating that You changed the files; and
+
+ (c) You must retain, in the Source form of any Derivative Works
+ that You distribute, all copyright, patent, trademark, and
+ attribution notices from the Source form of the Work,
+ excluding those notices that do not pertain to any part of
+ the Derivative Works; and
+
+ (d) If the Work includes a "NOTICE" text file as part of its
+ distribution, then any Derivative Works that You distribute must
+ include a readable copy of the attribution notices contained
+ within such NOTICE file, excluding those notices that do not
+ pertain to any part of the Derivative Works, in at least one
+ of the following places: within a NOTICE text file distributed
+ as part of the Derivative Works; within the Source form or
+ documentation, if provided along with the Derivative Works; or,
+ within a display generated by the Derivative Works, if and
+ wherever such third-party notices normally appear. The contents
+ of the NOTICE file are for informational purposes only and
+ do not modify the License. You may add Your own attribution
+ notices within Derivative Works that You distribute, alongside
+ or as an addendum to the NOTICE text from the Work, provided
+ that such additional attribution notices cannot be construed
+ as modifying the License.
+
+ You may add Your own copyright statement to Your modifications and
+ may provide additional or different license terms and conditions
+ for use, reproduction, or distribution of Your modifications, or
+ for any such Derivative Works as a whole, provided Your use,
+ reproduction, and distribution of the Work otherwise complies with
+ the conditions stated in this License.
+
+ 5. Submission of Contributions. Unless You explicitly state otherwise,
+ any Contribution intentionally submitted for inclusion in the Work
+ by You to the Licensor shall be under the terms and conditions of
+ this License, without any additional terms or conditions.
+ Notwithstanding the above, nothing herein shall supersede or modify
+ the terms of any separate license agreement you may have executed
+ with Licensor regarding such Contributions.
+
+ 6. Trademarks. This License does not grant permission to use the trade
+ names, trademarks, service marks, or product names of the Licensor,
+ except as required for reasonable and customary use in describing the
+ origin of the Work and reproducing the content of the NOTICE file.
+
+ 7. Disclaimer of Warranty. Unless required by applicable law or
+ agreed to in writing, Licensor provides the Work (and each
+ Contributor provides its Contributions) on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
+ implied, including, without limitation, any warranties or conditions
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
+ PARTICULAR PURPOSE. You are solely responsible for determining the
+ appropriateness of using or redistributing the Work and assume any
+ risks associated with Your exercise of permissions under this License.
+
+ 8. Limitation of Liability. In no event and under no legal theory,
+ whether in tort (including negligence), contract, or otherwise,
+ unless required by applicable law (such as deliberate and grossly
+ negligent acts) or agreed to in writing, shall any Contributor be
+ liable to You for damages, including any direct, indirect, special,
+ incidental, or consequential damages of any character arising as a
+ result of this License or out of the use or inability to use the
+ Work (including but not limited to damages for loss of goodwill,
+ work stoppage, computer failure or malfunction, or any and all
+ other commercial damages or losses), even if such Contributor
+ has been advised of the possibility of such damages.
+
+ 9. Accepting Warranty or Additional Liability. While redistributing
+ the Work or Derivative Works thereof, You may choose to offer,
+ and charge a fee for, acceptance of support, warranty, indemnity,
+ or other liability obligations and/or rights consistent with this
+ License. However, in accepting such obligations, You may act only
+ on Your own behalf and on Your sole responsibility, not on behalf
+ of any other Contributor, and only if You agree to indemnify,
+ defend, and hold each Contributor harmless for any liability
+ incurred by, or claims asserted against, such Contributor by reason
+ of your accepting any such warranty or additional liability.
+
+ END OF TERMS AND CONDITIONS
+
+ APPENDIX: How to apply the Apache License to your work.
+
+ To apply the Apache License to your work, attach the following
+ boilerplate notice, with the fields enclosed by brackets "[]"
+ replaced with your own identifying information. (Don't include
+ the brackets!) The text should be enclosed in the appropriate
+ comment syntax for the file format. We also recommend that a
+ file or class name and description of purpose be included on the
+ same "printed page" as the copyright notice for easier
+ identification within third-party archives.
+
+ Copyright [2024-] [Unsloth AI, Daniel Han-Chen & Michael Han-Chen]
+
+ Licensed under the Apache License, Version 2.0 (the "License");
+ you may not use this file except in compliance with the License.
+ You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
diff --git a/unsloth/README.md b/unsloth/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..ca5b6533b54bfcf0d87357ee436d612a6230d9e6
--- /dev/null
+++ b/unsloth/README.md
@@ -0,0 +1,440 @@
+
+
+
+
+
+
+
+
+
+
+
+
+### Finetune Llama 3, Mistral & Gemma 2-5x faster with 80% less memory!
+
+
+
+
+
+## ✨ Finetune for Free
+
+All notebooks are **beginner friendly**! Add your dataset, click "Run All", and you'll get a 2x faster finetuned model which can be exported to GGUF, vLLM or uploaded to Hugging Face.
+
+| Unsloth supports | Free Notebooks | Performance | Memory use |
+|-----------|---------|--------|----------|
+| **Llama 3 (8B)** | [▶️ Start for free](https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing) | 2x faster | 60% less |
+| **Mistral (7B)** | [▶️ Start for free](https://colab.research.google.com/drive/1Dyauq4kTZoLewQ1cApceUQVNcnnNTzg_?usp=sharing) | 2.2x faster | 73% less |
+| **Gemma (7B)** | [▶️ Start for free](https://colab.research.google.com/drive/10NbwlsRChbma1v55m8LAPYG15uQv6HLo?usp=sharing) | 2.4x faster | 71% less |
+| **ORPO** | [▶️ Start for free](https://colab.research.google.com/drive/11t4njE3c4Lxl-07OD8lJSMKkfyJml3Tn?usp=sharing) | 1.9x faster | 43% less |
+| **DPO Zephyr** | [▶️ Start for free](https://colab.research.google.com/drive/15vttTpzzVXv_tJwEk-hIcQ0S9FcEWvwP?usp=sharing) | 1.9x faster | 43% less |
+| **Phi-3 (3.8B)** | [▶️ Start for free](https://colab.research.google.com/drive/1NvkBmkHfucGO3Ve9s1NKZvMNlw5p83ym?usp=sharing) | 2x faster | 50% less |
+| **TinyLlama** | [▶️ Start for free](https://colab.research.google.com/drive/1AZghoNBQaMDgWJpi4RbffGM1h6raLUj9?usp=sharing) | 3.9x faster | 74% less |
+
+- Benchmarking compared to FA2 + Hugging Face combined.
+- **Kaggle Notebooks** for [Llama-3 8b](https://www.kaggle.com/code/danielhanchen/kaggle-llama-3-8b-unsloth-notebook), [Gemma 7b](https://www.kaggle.com/code/danielhanchen/kaggle-gemma-7b-unsloth-notebook/), [Mistral 7b](https://www.kaggle.com/code/danielhanchen/kaggle-mistral-7b-unsloth-notebook)
+- This [conversational notebook](https://colab.research.google.com/drive/1XamvWYinY6FOSX9GLvnqSjjsNflxdhNc?usp=sharing) is useful for Llama-3. And ChatML for [Mistral 7b](https://colab.research.google.com/drive/1Aau3lgPzeZKQ-98h69CCu1UJcvIBLmy2?usp=sharing).
+- This [text completion notebook](https://colab.research.google.com/drive/1ef-tab5bhkvWmBOObepl1WgJvfvSzn5Q?usp=sharing) is for continued pretraining / raw text.
+
+## 🦥 Unsloth.ai News
+- 📣 NEW! Qwen1.5-7B, Qwen1.5-14B, Qwen1.5-32B, Qwen1.5-72B now work, courtesy of Firefly's PR [#428](https://github.com/unslothai/unsloth/pull/428)
+- 📣 NEW! [Llama-3 8b](https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing) now works! Llama-3 70b also works (change the model name in the notebook).
+- 📣 NEW! [ORPO support](https://colab.research.google.com/drive/11t4njE3c4Lxl-07OD8lJSMKkfyJml3Tn?usp=sharing) is here!
+- 📣 NEW! [Phi-3 3.8b support](https://colab.research.google.com/drive/1NvkBmkHfucGO3Ve9s1NKZvMNlw5p83ym?usp=sharing) is here!
+- 📣 NEW! We cut memory usage by a [further 30%](https://unsloth.ai/blog/long-context) and now support fine-tuning of LLMs with [4x longer context windows](https://unsloth.ai/blog/long-context)! No change required if you're using our notebooks. To enable, simply change 1 line:
+```python
+model = FastLanguageModel.get_peft_model(
+ model,
+ use_gradient_checkpointing = "unsloth", # <<<<<<<
+)
+```
+- 📣 [CodeGemma](https://colab.research.google.com/drive/19lwcRk_ZQ_ZtX-qzFP3qZBBHZNcMD1hh?usp=sharing) now works along with [Gemma 7b](https://colab.research.google.com/drive/10NbwlsRChbma1v55m8LAPYG15uQv6HLo?usp=sharing) and [Gemma 2b](https://colab.research.google.com/drive/15gGm7x_jTm017_Ic8e317tdIpDG53Mtu?usp=sharing)
+- 📣 [2x faster inference](https://colab.research.google.com/drive/1aqlNQi7MMJbynFDyOQteD2t0yVfjb9Zh?usp=sharing) added for all our models
+
+## 🔗 Links and Resources
+| Type | Links |
+| ------------------------------- | --------------------------------------- |
+| 📚 **Wiki & FAQ** | [Read Our Wiki](https://github.com/unslothai/unsloth/wiki) |
+| **Twitter (aka X)** | [Follow us on X](https://twitter.com/unslothai)|
+| 📜 **Documentation** | [Read The Doc](https://github.com/unslothai/unsloth/tree/main#-documentation) |
+| 💾 **Installation** | [unsloth/README.md](https://github.com/unslothai/unsloth/tree/main#installation-instructions)|
+| 🥇 **Benchmarking** | [Performance Tables](https://github.com/unslothai/unsloth/tree/main#-performance-benchmarking)
+| 🌐 **Released Models** | [Unsloth Releases](https://huggingface.co/unsloth)|
+| ✍️ **Blog** | [Read our Blogs](https://unsloth.ai/blog)|
+
+## ⭐ Key Features
+- All kernels written in [OpenAI's Triton](https://openai.com/research/triton) language. **Manual backprop engine**.
+- **0% loss in accuracy** - no approximation methods - all exact.
+- No change of hardware. Supports NVIDIA GPUs since 2018+. Minimum CUDA Capability 7.0 (V100, T4, Titan V, RTX 20, 30, 40x, A100, H100, L40 etc) [Check your GPU!](https://developer.nvidia.com/cuda-gpus) GTX 1070, 1080 works, but is slow.
+- Works on **Linux** and **Windows** via WSL.
+- Supports 4bit and 16bit QLoRA / LoRA finetuning via [bitsandbytes](https://github.com/TimDettmers/bitsandbytes).
+- Open source trains 5x faster - see [Unsloth Pro](https://unsloth.ai/) for up to **30x faster training**!
+- If you trained a model with 🦥Unsloth, you can use this cool sticker!
+
+
+## 🥇 Performance Benchmarking
+- For the full list of **reproducable** benchmarking tables, [go to our website](https://unsloth.ai/blog/mistral-benchmark#Benchmark%20tables)
+
+| 1 A100 40GB | 🤗Hugging Face | Flash Attention | 🦥Unsloth Open Source | 🦥[Unsloth Pro](https://unsloth.ai/pricing) |
+|--------------|--------------|-----------------|---------------------|-----------------|
+| Alpaca | 1x | 1.04x | 1.98x | **15.64x** |
+| LAION Chip2 | 1x | 0.92x | 1.61x | **20.73x** |
+| OASST | 1x | 1.19x | 2.17x | **14.83x** |
+| Slim Orca | 1x | 1.18x | 2.22x | **14.82x** |
+
+- Benchmarking table below was conducted by [🤗Hugging Face](https://huggingface.co/blog/unsloth-trl).
+
+| Free Colab T4 | Dataset | 🤗Hugging Face | Pytorch 2.1.1 | 🦥Unsloth | 🦥 VRAM reduction |
+| --- | --- | --- | --- | --- | --- |
+| Llama-2 7b | OASST | 1x | 1.19x | 1.95x | -43.3% |
+| Mistral 7b | Alpaca | 1x | 1.07x | 1.56x | -13.7% |
+| Tiny Llama 1.1b | Alpaca | 1x | 2.06x | 3.87x | -73.8% |
+| DPO with Zephyr | Ultra Chat | 1x | 1.09x | 1.55x | -18.6% |
+
+
+
+## 💾 Installation Instructions
+### Conda Installation
+Select either `pytorch-cuda=11.8` for CUDA 11.8 or `pytorch-cuda=12.1` for CUDA 12.1. If you have `mamba`, use `mamba` instead of `conda` for faster solving. See this [Github issue](https://github.com/unslothai/unsloth/issues/73) for help on debugging Conda installs.
+```bash
+conda create --name unsloth_env python=3.10
+conda activate unsloth_env
+
+conda install pytorch-cuda=<12.1/11.8> pytorch cudatoolkit xformers -c pytorch -c nvidia -c xformers
+
+pip install "unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git"
+
+pip install --no-deps trl peft accelerate bitsandbytes
+```
+
+### Pip Installation
+Do **NOT** use this if you have Anaconda. You must use the Conda install method, or else stuff will BREAK.
+
+1. Find your CUDA version via
+```python
+import torch; torch.version.cuda
+```
+2. For Pytorch 2.1.0: You can update Pytorch via Pip (interchange `cu121` / `cu118`). Go to https://pytorch.org/ to learn more. Select either `cu118` for CUDA 11.8 or `cu121` for CUDA 12.1. If you have a RTX 3060 or higher (A100, H100 etc), use the `"ampere"` path. For Pytorch 2.1.1: go to step 3. For Pytorch 2.2.0: go to step 4.
+```bash
+pip install --upgrade --force-reinstall --no-cache-dir torch==2.1.0 triton \
+ --index-url https://download.pytorch.org/whl/cu121
+```
+```bash
+pip install "unsloth[cu118] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu118-ampere] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-ampere] @ git+https://github.com/unslothai/unsloth.git"
+```
+3. For Pytorch 2.1.1: Use the `"ampere"` path for newer RTX 30xx GPUs or higher.
+```bash
+pip install --upgrade --force-reinstall --no-cache-dir torch==2.1.1 triton \
+ --index-url https://download.pytorch.org/whl/cu121
+```
+```bash
+pip install "unsloth[cu118-torch211] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-torch211] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu118-ampere-torch211] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-ampere-torch211] @ git+https://github.com/unslothai/unsloth.git"
+```
+4. For Pytorch 2.2.0: Use the `"ampere"` path for newer RTX 30xx GPUs or higher.
+```bash
+pip install --upgrade --force-reinstall --no-cache-dir torch==2.2.0 triton \
+ --index-url https://download.pytorch.org/whl/cu121
+```
+```bash
+pip install "unsloth[cu118-torch220] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-torch220] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu118-ampere-torch220] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-ampere-torch220] @ git+https://github.com/unslothai/unsloth.git"
+```
+5. If you get errors, try the below first, then go back to step 1:
+```bash
+pip install --upgrade pip
+```
+6. For Pytorch 2.2.1:
+```bash
+# RTX 3090, 4090 Ampere GPUs:
+pip install "unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git"
+pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes
+
+# Pre Ampere RTX 2080, T4, GTX 1080 GPUs:
+pip install "unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git"
+pip install --no-deps xformers trl peft accelerate bitsandbytes
+```
+7. For Pytorch 2.3.0: Use the `"ampere"` path for newer RTX 30xx GPUs or higher.
+```bash
+pip install "unsloth[cu118-torch230] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-torch230] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu118-ampere-torch230] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-ampere-torch230] @ git+https://github.com/unslothai/unsloth.git"
+```
+8. To troubleshoot installs try the below (all must succeed). Xformers should mostly all be available.
+```bash
+nvcc
+python -m xformers.info
+python -m bitsandbytes
+```
+
+## 📜 Documentation
+- Go to our [Wiki page](https://github.com/unslothai/unsloth/wiki) for saving to GGUF, checkpointing, evaluation and more!
+- We support Huggingface's TRL, Trainer, Seq2SeqTrainer or even Pytorch code!
+- We're in 🤗Hugging Face's official docs! Check out the [SFT docs](https://huggingface.co/docs/trl/main/en/sft_trainer#accelerate-fine-tuning-2x-using-unsloth) and [DPO docs](https://huggingface.co/docs/trl/main/en/dpo_trainer#accelerate-dpo-fine-tuning-using-unsloth)!
+
+```python
+from unsloth import FastLanguageModel
+import torch
+from trl import SFTTrainer
+from transformers import TrainingArguments
+from datasets import load_dataset
+max_seq_length = 2048 # Supports RoPE Scaling interally, so choose any!
+# Get LAION dataset
+url = "https://huggingface.co/datasets/laion/OIG/resolve/main/unified_chip2.jsonl"
+dataset = load_dataset("json", data_files = {"train" : url}, split = "train")
+
+# 4bit pre quantized models we support for 4x faster downloading + no OOMs.
+fourbit_models = [
+ "unsloth/mistral-7b-bnb-4bit",
+ "unsloth/mistral-7b-instruct-v0.2-bnb-4bit",
+ "unsloth/llama-2-7b-bnb-4bit",
+ "unsloth/gemma-7b-bnb-4bit",
+ "unsloth/gemma-7b-it-bnb-4bit", # Instruct version of Gemma 7b
+ "unsloth/gemma-2b-bnb-4bit",
+ "unsloth/gemma-2b-it-bnb-4bit", # Instruct version of Gemma 2b
+ "unsloth/llama-3-8b-bnb-4bit", # [NEW] 15 Trillion token Llama-3
+ "unsloth/Phi-3-mini-4k-instruct-bnb-4bit",
+] # More models at https://huggingface.co/unsloth
+
+model, tokenizer = FastLanguageModel.from_pretrained(
+ model_name = "unsloth/llama-3-8b-bnb-4bit",
+ max_seq_length = max_seq_length,
+ dtype = None,
+ load_in_4bit = True,
+)
+
+# Do model patching and add fast LoRA weights
+model = FastLanguageModel.get_peft_model(
+ model,
+ r = 16,
+ target_modules = ["q_proj", "k_proj", "v_proj", "o_proj",
+ "gate_proj", "up_proj", "down_proj",],
+ lora_alpha = 16,
+ lora_dropout = 0, # Supports any, but = 0 is optimized
+ bias = "none", # Supports any, but = "none" is optimized
+ # [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
+ use_gradient_checkpointing = "unsloth", # True or "unsloth" for very long context
+ random_state = 3407,
+ max_seq_length = max_seq_length,
+ use_rslora = False, # We support rank stabilized LoRA
+ loftq_config = None, # And LoftQ
+)
+
+trainer = SFTTrainer(
+ model = model,
+ train_dataset = dataset,
+ dataset_text_field = "text",
+ max_seq_length = max_seq_length,
+ tokenizer = tokenizer,
+ args = TrainingArguments(
+ per_device_train_batch_size = 2,
+ gradient_accumulation_steps = 4,
+ warmup_steps = 10,
+ max_steps = 60,
+ fp16 = not torch.cuda.is_bf16_supported(),
+ bf16 = torch.cuda.is_bf16_supported(),
+ logging_steps = 1,
+ output_dir = "outputs",
+ optim = "adamw_8bit",
+ seed = 3407,
+ ),
+)
+trainer.train()
+
+# Go to https://github.com/unslothai/unsloth/wiki for advanced tips like
+# (1) Saving to GGUF / merging to 16bit for vLLM
+# (2) Continued training from a saved LoRA adapter
+# (3) Adding an evaluation loop / OOMs
+# (4) Cutomized chat templates
+```
+
+
+## DPO Support
+DPO (Direct Preference Optimization), PPO, Reward Modelling all seem to work as per 3rd party independent testing from [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory). We have a preliminary Google Colab notebook for reproducing Zephyr on Tesla T4 here: [notebook](https://colab.research.google.com/drive/15vttTpzzVXv_tJwEk-hIcQ0S9FcEWvwP?usp=sharing).
+
+We're in 🤗Hugging Face's official docs! We're on the [SFT docs](https://huggingface.co/docs/trl/main/en/sft_trainer#accelerate-fine-tuning-2x-using-unsloth) and the [DPO docs](https://huggingface.co/docs/trl/main/en/dpo_trainer#accelerate-dpo-fine-tuning-using-unsloth)!
+
+```python
+from unsloth import FastLanguageModel, PatchDPOTrainer
+PatchDPOTrainer()
+import torch
+from transformers import TrainingArguments
+from trl import DPOTrainer
+
+model, tokenizer = FastLanguageModel.from_pretrained(
+ model_name = "unsloth/zephyr-sft-bnb-4bit",
+ max_seq_length = max_seq_length,
+ dtype = None,
+ load_in_4bit = True,
+)
+
+# Do model patching and add fast LoRA weights
+model = FastLanguageModel.get_peft_model(
+ model,
+ r = 64,
+ target_modules = ["q_proj", "k_proj", "v_proj", "o_proj",
+ "gate_proj", "up_proj", "down_proj",],
+ lora_alpha = 64,
+ lora_dropout = 0, # Supports any, but = 0 is optimized
+ bias = "none", # Supports any, but = "none" is optimized
+ # [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
+ use_gradient_checkpointing = "unsloth", # True or "unsloth" for very long context
+ random_state = 3407,
+ max_seq_length = max_seq_length,
+)
+
+dpo_trainer = DPOTrainer(
+ model = model,
+ ref_model = None,
+ args = TrainingArguments(
+ per_device_train_batch_size = 4,
+ gradient_accumulation_steps = 8,
+ warmup_ratio = 0.1,
+ num_train_epochs = 3,
+ fp16 = not torch.cuda.is_bf16_supported(),
+ bf16 = torch.cuda.is_bf16_supported(),
+ logging_steps = 1,
+ optim = "adamw_8bit",
+ seed = 42,
+ output_dir = "outputs",
+ ),
+ beta = 0.1,
+ train_dataset = YOUR_DATASET_HERE,
+ # eval_dataset = YOUR_DATASET_HERE,
+ tokenizer = tokenizer,
+ max_length = 1024,
+ max_prompt_length = 512,
+)
+dpo_trainer.train()
+```
+
+## 🥇 Detailed Benchmarking Tables
+- Click "Code" for fully reproducible examples
+- "Unsloth Equal" is a preview of our PRO version, with code stripped out. All settings and the loss curve remains identical.
+- For the full list of benchmarking tables, [go to our website](https://unsloth.ai/blog/mistral-benchmark#Benchmark%20tables)
+
+| 1 A100 40GB | 🤗Hugging Face | Flash Attention 2 | 🦥Unsloth Open | Unsloth Equal | Unsloth Pro | Unsloth Max |
+|--------------|-------------|-------------|-----------------|--------------|---------------|-------------|
+| Alpaca | 1x | 1.04x | 1.98x | 2.48x | 5.32x | **15.64x** |
+| code | [Code](https://colab.research.google.com/drive/1u4dBeM-0vGNVmmO6X7cScAut-Hyt4KDF?usp=sharing) | [Code](https://colab.research.google.com/drive/1fgTOxpMbVjloQBvZyz4lF4BacKSZOB2A?usp=sharing) | [Code](https://colab.research.google.com/drive/1YIPY_18xm-K0iJDgvNkRoJsgkPMPAO3G?usp=sharing) | [Code](https://colab.research.google.com/drive/1ANW8EFL3LVyTD7Gq4TkheC1Z7Rxw-rHp?usp=sharing) | | |
+| seconds| 1040 | 1001 | 525 | 419 | 196 | 67 |
+| memory MB| 18235 | 15365 | 9631 | 8525 | | |
+| % saved| | 15.74 | 47.18 | 53.25 | | | |
+
+### Llama-Factory 3rd party benchmarking
+- [Link to performance table.](https://github.com/hiyouga/LLaMA-Factory/wiki/Performance-Comparison) TGS: tokens per GPU per second. Model: LLaMA2-7B. GPU: NVIDIA A100 * 1. Batch size: 4. Gradient accumulation: 2. LoRA rank: 8. Max length: 1024.
+
+| Method | Bits | TGS | GRAM | Speed |
+| --- | --- | --- | --- | --- |
+| HF | 16 | 2392 | 18GB | 100% |
+| HF+FA2 | 16 | 2954 | 17GB | 123% |
+| Unsloth+FA2 | 16 | 4007 | 16GB | **168%** |
+| HF | 4 | 2415 | 9GB | 101% |
+| Unsloth+FA2 | 4 | 3726 | 7GB | **160%** |
+
+### Performance comparisons between popular models
+
+ Click for specific model benchmarking tables (Mistral 7b, CodeLlama 34b etc.)
+
+### Mistral 7b
+| 1 A100 40GB | Hugging Face | Flash Attention 2 | Unsloth Open | Unsloth Equal | Unsloth Pro | Unsloth Max |
+|--------------|-------------|-------------|-----------------|--------------|---------------|-------------|
+| Mistral 7B Slim Orca | 1x | 1.15x | 2.15x | 2.53x | 4.61x | **13.69x** |
+| code | [Code](https://colab.research.google.com/drive/1mePk3KzwTD81hr5mcNcs_AX3Kbg_Ha0x?usp=sharing) | [Code](https://colab.research.google.com/drive/1dgHxjvTmX6hb0bPcLp26RXSE6_n9DKj7?usp=sharing) | [Code](https://colab.research.google.com/drive/1SKrKGV-BZoU4kv5q3g0jtE_OhRgPtrrQ?usp=sharing) | [Code](https://colab.research.google.com/drive/18yOiyX0T81mTwZqOALFSCX_tSAqju6aD?usp=sharing) | |
+| seconds | 1813 | 1571 | 842 | 718 | 393 | 132 |
+| memory MB | 32853 | 19385 | 12465 | 10271 | | |
+| % saved| | 40.99 | 62.06 | 68.74 | | |
+
+### CodeLlama 34b
+| 1 A100 40GB | Hugging Face | Flash Attention 2 | Unsloth Open | Unsloth Equal | Unsloth Pro | Unsloth Max |
+|--------------|-------------|-------------|-----------------|--------------|---------------|-------------|
+| Code Llama 34B | OOM ❌ | 0.99x | 1.87x | 2.61x | 4.27x | 12.82x |
+| code | [▶️ Code](https://colab.research.google.com/drive/1ykfz3BqrtC_AUFegCzUQjjfUNlxp6Otc?usp=sharing) | [Code](https://colab.research.google.com/drive/12ZypxQh7OC6kBXvWZI-5d05I4m-B_hoR?usp=sharing) | [Code](https://colab.research.google.com/drive/1gdHyAx8XJsz2yNV-DHvbHjR1iCef5Qmh?usp=sharing) | [Code](https://colab.research.google.com/drive/1fm7wqx9MJ0kRrwKOfmLkK1Rmw-pySahB?usp=sharing) | |
+| seconds | 1953 | 1982 | 1043 | 748 | 458 | 152 |
+| memory MB | 40000 | 33217 | 27413 | 22161 | | |
+| % saved| | 16.96| 31.47 | 44.60 | | | |
+
+### 1 Tesla T4
+
+| 1 T4 16GB | Hugging Face | Flash Attention | Unsloth Open | Unsloth Pro Equal | Unsloth Pro | Unsloth Max |
+|--------------|-------------|-----------------|-----------------|---------------|---------------|-------------|
+| Alpaca | 1x | 1.09x | 1.69x | 1.79x | 2.93x | **8.3x** |
+| code | [▶️ Code](https://colab.research.google.com/drive/1XpLIV4s8Bj5uryB-X2gqM88oRGHEGdaB?usp=sharing) | [Code](https://colab.research.google.com/drive/1LyXu6CjuymQg6ddHX8g1dpUvrMa1nn4L?usp=sharing) | [Code](https://colab.research.google.com/drive/1gsv4LpY7C32otl1rgRo5wXTk4HIitXoM?usp=sharing) | [Code](https://colab.research.google.com/drive/1VtULwRQwhEnVdNryjm27zXfdSM1tNfFK?usp=sharing) | | |
+| seconds | 1599 | 1468 | 942 | 894 | 545 | 193 |
+| memory MB | 7199 | 7059 | 6459 | 5443 | | |
+| % saved | | 1.94 | 10.28 | 24.39 | | |
+
+### 2 Tesla T4s via DDP
+
+ | 2 T4 DDP | Hugging Face | Flash Attention | Unsloth Open | Unsloth Equal | Unsloth Pro | Unsloth Max |
+|--------------|----------|-------------|-----------------|--------------|---------------|-------------|
+| Alpaca | 1x | 0.99x | 4.95x | 4.44x | 7.28x | **20.61x** |
+| code | [▶️ Code](https://www.kaggle.com/danielhanchen/hf-original-alpaca-t4-ddp) | [Code](https://www.kaggle.com/danielhanchen/hf-sdpa-alpaca-t4-ddp) | [Code](https://www.kaggle.com/danielhanchen/unsloth-alpaca-t4-ddp) | | |
+| seconds | 9882 | 9946 | 1996 | 2227 | 1357 | 480 |
+| memory MB| 9176 | 9128 | 6904 | 6782 | | |
+| % saved | | 0.52 | 24.76 | 26.09 | | | |
+
+
+### Performance comparisons on 1 Tesla T4 GPU:
+
+ Click for Time taken for 1 epoch
+
+One Tesla T4 on Google Colab
+`bsz = 2, ga = 4, max_grad_norm = 0.3, num_train_epochs = 1, seed = 3047, lr = 2e-4, wd = 0.01, optim = "adamw_8bit", schedule = "linear", schedule_steps = 10`
+
+| System | GPU | Alpaca (52K) | LAION OIG (210K) | Open Assistant (10K) | SlimOrca (518K) |
+| --- | --- | --- | --- | --- | --- |
+| Huggingface | 1 T4 | 23h 15m | 56h 28m | 8h 38m | 391h 41m |
+| Unsloth Open | 1 T4 | 13h 7m (1.8x) | 31h 47m (1.8x) | 4h 27m (1.9x) | 240h 4m (1.6x) |
+| Unsloth Pro | 1 T4 | 3h 6m (7.5x) | 5h 17m (10.7x) | 1h 7m (7.7x) | 59h 53m (6.5x) |
+| Unsloth Max | 1 T4 | 2h 39m (8.8x) | 4h 31m (12.5x) | 0h 58m (8.9x) | 51h 30m (7.6x) |
+
+**Peak Memory Usage**
+
+| System | GPU | Alpaca (52K) | LAION OIG (210K) | Open Assistant (10K) | SlimOrca (518K) |
+| --- | --- | --- | --- | --- | --- |
+| Huggingface | 1 T4 | 7.3GB | 5.9GB | 14.0GB | 13.3GB |
+| Unsloth Open | 1 T4 | 6.8GB | 5.7GB | 7.8GB | 7.7GB |
+| Unsloth Pro | 1 T4 | 6.4GB | 6.4GB | 6.4GB | 6.4GB |
+| Unsloth Max | 1 T4 | 11.4GB | 12.4GB | 11.9GB | 14.4GB |
+
+
+
+ Click for Performance Comparisons on 2 Tesla T4 GPUs via DDP:
+**Time taken for 1 epoch**
+
+Two Tesla T4s on Kaggle
+`bsz = 2, ga = 4, max_grad_norm = 0.3, num_train_epochs = 1, seed = 3047, lr = 2e-4, wd = 0.01, optim = "adamw_8bit", schedule = "linear", schedule_steps = 10`
+
+| System | GPU | Alpaca (52K) | LAION OIG (210K) | Open Assistant (10K) | SlimOrca (518K) * |
+| --- | --- | --- | --- | --- | --- |
+| Huggingface | 2 T4 | 84h 47m | 163h 48m | 30h 51m | 1301h 24m * |
+| Unsloth Pro | 2 T4 | 3h 20m (25.4x) | 5h 43m (28.7x) | 1h 12m (25.7x) | 71h 40m (18.1x) * |
+| Unsloth Max | 2 T4 | 3h 4m (27.6x) | 5h 14m (31.3x) | 1h 6m (28.1x) | 54h 20m (23.9x) * |
+
+**Peak Memory Usage on a Multi GPU System (2 GPUs)**
+
+| System | GPU | Alpaca (52K) | LAION OIG (210K) | Open Assistant (10K) | SlimOrca (518K) * |
+| --- | --- | --- | --- | --- | --- |
+| Huggingface | 2 T4 | 8.4GB \| 6GB | 7.2GB \| 5.3GB | 14.3GB \| 6.6GB | 10.9GB \| 5.9GB * |
+| Unsloth Pro | 2 T4 | 7.7GB \| 4.9GB | 7.5GB \| 4.9GB | 8.5GB \| 4.9GB | 6.2GB \| 4.7GB * |
+| Unsloth Max | 2 T4 | 10.5GB \| 5GB | 10.6GB \| 5GB | 10.6GB \| 5GB | 10.5GB \| 5GB * |
+
+* Slim Orca `bsz=1` for all benchmarks since `bsz=2` OOMs. We can handle `bsz=2`, but we benchmark it with `bsz=1` for consistency.
+
+
+
+
+
+### Thank You to
+- [HuyNguyen-hust](https://github.com/HuyNguyen-hust) for making [RoPE Embeddings 28% faster](https://github.com/unslothai/unsloth/pull/238)
+- [RandomInternetPreson](https://github.com/RandomInternetPreson) for confirming WSL support
+- [152334H](https://github.com/152334H) for experimental DPO support
+- [atgctg](https://github.com/atgctg) for syntax highlighting
diff --git a/unsloth/images/Colab.png b/unsloth/images/Colab.png
new file mode 100644
index 0000000000000000000000000000000000000000..b394add439e4e8ff47f626dbf8e73f9c297d77ad
--- /dev/null
+++ b/unsloth/images/Colab.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:81ff231ac422053de057094a5e7d15c2ed2f33acd2330ff57dfabfa7bc2bc1fc
+size 11637
diff --git a/unsloth/images/Discord button.png b/unsloth/images/Discord button.png
new file mode 100644
index 0000000000000000000000000000000000000000..f7ebcdd6ca1a5cc6fbd9bc06cfa8e33fb9fe1d37
--- /dev/null
+++ b/unsloth/images/Discord button.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:26d749193097256557675a05fa25e8dc31676566a375202c06ab460bc5bd3005
+size 13767
diff --git a/unsloth/images/Discord.png b/unsloth/images/Discord.png
new file mode 100644
index 0000000000000000000000000000000000000000..f7ebcdd6ca1a5cc6fbd9bc06cfa8e33fb9fe1d37
--- /dev/null
+++ b/unsloth/images/Discord.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:26d749193097256557675a05fa25e8dc31676566a375202c06ab460bc5bd3005
+size 13767
diff --git a/unsloth/images/Free version button.png b/unsloth/images/Free version button.png
new file mode 100644
index 0000000000000000000000000000000000000000..6e7dcdfab9ec5b500aafa96795113e28acdbfc80
--- /dev/null
+++ b/unsloth/images/Free version button.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f692d0636dbb002c25c4149164f5c05656833d9da4ba2b5409227446e6da35cd
+size 9530
diff --git a/unsloth/images/Kaggle.png b/unsloth/images/Kaggle.png
new file mode 100644
index 0000000000000000000000000000000000000000..c22054191d9bb270587a67b3404f193de962d520
--- /dev/null
+++ b/unsloth/images/Kaggle.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c97803714456ed4fac1654c4f51c0de28a974baedb8b77f37042ed8527db27de
+size 9733
diff --git a/unsloth/images/Kofi button.png b/unsloth/images/Kofi button.png
new file mode 100644
index 0000000000000000000000000000000000000000..a6f02935b611093533a972e9a687606121e61820
--- /dev/null
+++ b/unsloth/images/Kofi button.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f64087ba3ea46c6bd150caefbe09da2d1dbea9c6b74aea4f25ec9e066ff90990
+size 18109
diff --git a/unsloth/images/LAION 2GPU.png b/unsloth/images/LAION 2GPU.png
new file mode 100644
index 0000000000000000000000000000000000000000..036d92ae353227cc7def863a2e25c19178137733
--- /dev/null
+++ b/unsloth/images/LAION 2GPU.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:67caf9cb7a14d198439cee12b7897f00a167f087cedf079ab2f2a479d81a5565
+size 51428
diff --git a/unsloth/images/Slim Orca 2GPUs.png b/unsloth/images/Slim Orca 2GPUs.png
new file mode 100644
index 0000000000000000000000000000000000000000..0daa75c551e8bc63d7ac2467d0631493506db60b
--- /dev/null
+++ b/unsloth/images/Slim Orca 2GPUs.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:613baa6ba06a3af4a3793ee1d2ca3b5bd3a623d6ed5a83251f0b8906d83ff656
+size 43135
diff --git a/unsloth/images/buy me a coffee button.png b/unsloth/images/buy me a coffee button.png
new file mode 100644
index 0000000000000000000000000000000000000000..5850402c0e485904405e679dbe7cc5cc1ed51522
--- /dev/null
+++ b/unsloth/images/buy me a coffee button.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0f799d119ccb4bc84f8fba6fecaa1cd58b570896332166744ad8019d400d4661
+size 18966
diff --git a/unsloth/images/made with unsloth.png b/unsloth/images/made with unsloth.png
new file mode 100644
index 0000000000000000000000000000000000000000..e9761f3302e43df70c6a1479ef818e914468e299
--- /dev/null
+++ b/unsloth/images/made with unsloth.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5bd1d93165e18feaaf1f4b4a4d002f48e7760e6a2fc5ffd9dde00612d7ec34ef
+size 70443
diff --git a/unsloth/images/peft x trl button.png b/unsloth/images/peft x trl button.png
new file mode 100644
index 0000000000000000000000000000000000000000..83b6daf75fbf47003a3687061feb62d03013338b
--- /dev/null
+++ b/unsloth/images/peft x trl button.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c2f8813c08d443aa75116f6f240803e592ee0f222c1b0c0896e311b3a090848a
+size 36919
diff --git a/unsloth/images/start free finetune button.png b/unsloth/images/start free finetune button.png
new file mode 100644
index 0000000000000000000000000000000000000000..f38a0ac59180bbd38309e90d2829ff45f784b067
--- /dev/null
+++ b/unsloth/images/start free finetune button.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5dfeec1ee54a066198421e6b14f9a8e9fc8d37e0bd18b60dff94688b48ea88ee
+size 11432
diff --git a/unsloth/images/unsloth end.png b/unsloth/images/unsloth end.png
new file mode 100644
index 0000000000000000000000000000000000000000..3db61cbc53c72ab715f30041412db3e5409561ec
--- /dev/null
+++ b/unsloth/images/unsloth end.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:026eb3fe168c9efcaf3f0e5028e44e1b4d9dab9b4b804da5ee9c7691b727fb03
+size 892307
diff --git a/unsloth/images/unsloth loading page render.png b/unsloth/images/unsloth loading page render.png
new file mode 100644
index 0000000000000000000000000000000000000000..f77b75afe220ba727841cd3575943f7351286cf4
--- /dev/null
+++ b/unsloth/images/unsloth loading page render.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a1401edf8cd408bc4c508a58b4ebcecfa698a2ed829dbb498e1259d1bc09e898
+size 789661
diff --git a/unsloth/images/unsloth logo black text.png b/unsloth/images/unsloth logo black text.png
new file mode 100644
index 0000000000000000000000000000000000000000..e616477503cf84534ac5b0916ed5491d8fc51fcb
--- /dev/null
+++ b/unsloth/images/unsloth logo black text.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9a6485b736f02b899cc8297ffc179f0cb13f03b8f601b9b8b0d606fc39562060
+size 58002
diff --git a/unsloth/images/unsloth logo only.png b/unsloth/images/unsloth logo only.png
new file mode 100644
index 0000000000000000000000000000000000000000..0455392f919c4f97688e08965221dd544eef8ec8
--- /dev/null
+++ b/unsloth/images/unsloth logo only.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:61527e64f05b6ffc4a6fbfa9fc1229d6082a1166f82b88e76740aebc9354d5ee
+size 57176
diff --git a/unsloth/images/unsloth logo white text.png b/unsloth/images/unsloth logo white text.png
new file mode 100644
index 0000000000000000000000000000000000000000..aec908bd7bbef3b4c1815095163a816045dda698
--- /dev/null
+++ b/unsloth/images/unsloth logo white text.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5ee02ca80ad3d0877f7b4efb4366bf568e4c511fe386d6f83f37090aff423eec
+size 58964
diff --git a/unsloth/images/unsloth made with love.png b/unsloth/images/unsloth made with love.png
new file mode 100644
index 0000000000000000000000000000000000000000..b9a6a2cb313b70e3d4a5d8950b0e350a548bf3e6
--- /dev/null
+++ b/unsloth/images/unsloth made with love.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:37a08c5d671c5186512a66d6067c03afa133c4bf633ac5c79ce1077e3da7f57c
+size 63453
diff --git a/unsloth/images/unsloth new logo.png b/unsloth/images/unsloth new logo.png
new file mode 100644
index 0000000000000000000000000000000000000000..75dc59af18ddb02db369a0e02d8831e25619bbbe
--- /dev/null
+++ b/unsloth/images/unsloth new logo.png
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d4eccff7c5811639ee11fb38f6030fb28e480b5900f62198e9d2455033b8a13e
+size 60119
diff --git a/unsloth/pyproject.toml b/unsloth/pyproject.toml
new file mode 100644
index 0000000000000000000000000000000000000000..0398d0df4c411ee631883be935a6c0fa8d7f3fb2
--- /dev/null
+++ b/unsloth/pyproject.toml
@@ -0,0 +1,282 @@
+[build-system]
+requires = ["setuptools", "setuptools-scm"]
+build-backend = "setuptools.build_meta"
+
+[project]
+name = "unsloth"
+dynamic = ["version"]
+description = "2-5X faster LLM finetuning"
+readme = "README.md"
+requires-python = ">=3.9"
+license = {file = "LICENSE"}
+keywords = ["ai", "llm",]
+authors = [
+ {email = "info@unsloth.ai"},
+ {name = "Unsloth AI team"},
+]
+maintainers = [
+ {name = "Daniel Han", email = "danielhanchen@gmail.com"},
+ {name = "Michael Han", email = "info@unsloth.ai"},
+]
+classifiers = [
+ "Programming Language :: Python",
+]
+
+[tool.setuptools.dynamic]
+version = {attr = "unsloth.models._utils.__version__"}
+
+[tool.setuptools]
+include-package-data = false
+
+[tool.setuptools.packages.find]
+exclude = ["images*"]
+
+[project.optional-dependencies]
+huggingface = [
+ "tyro",
+ "transformers>=4.38.2",
+ "datasets>=2.16.0",
+ "sentencepiece>=0.2.0",
+ "tqdm",
+ "psutil",
+ "wheel>=0.42.0",
+ "numpy",
+ "accelerate>=0.26.1",
+ "trl>=0.7.9",
+ "peft>=0.7.1",
+ "protobuf<4.0.0",
+]
+cu118only = [
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.22.post7%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.22.post7%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.22.post7%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+cu121only = [
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.22.post7-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.22.post7-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.22.post7-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+cu118onlytorch211 = [
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.23%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.23%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.23%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+cu121onlytorch211 = [
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.23-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.23-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.23-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+cu118onlytorch212 = [
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.23.post1%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.23.post1%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.23.post1%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+cu121onlytorch212 = [
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.23.post1-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.23.post1-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.23.post1-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+cu118onlytorch220 = [
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.24%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.24%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.24%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+cu121onlytorch220 = [
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+cu118onlytorch230 = [
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.26.post1%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.26.post1%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu118/xformers-0.0.26.post1%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+cu121onlytorch230 = [
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.26.post1-cp39-cp39-manylinux2014_x86_64.whl ; python_version=='3.9'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.26.post1-cp310-cp310-manylinux2014_x86_64.whl ; python_version=='3.10'",
+ "xformers @ https://download.pytorch.org/whl/cu121/xformers-0.0.26.post1-cp311-cp311-manylinux2014_x86_64.whl ; python_version=='3.11'",
+]
+
+cu118 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu118only]",
+]
+cu121 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121only]",
+]
+cu118-torch211 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu118onlytorch211]",
+]
+cu121-torch211 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch211]",
+]
+cu118-torch212 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu118onlytorch212]",
+]
+cu121-torch212 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch212]",
+]
+cu118-torch220 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu118onlytorch220]",
+]
+cu121-torch220 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch220]",
+]
+cu118-torch230 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu118onlytorch230]",
+]
+cu121-torch230 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch230]",
+]
+kaggle = [
+ "unsloth[huggingface]",
+]
+kaggle-new = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+]
+conda = [
+ "unsloth[huggingface]",
+]
+colab-torch211 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch211]",
+]
+colab-ampere-torch211 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch211]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+colab-torch220 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch220]",
+]
+colab-ampere-torch220 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch220]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+colab-new = [
+ "tyro",
+ "transformers>=4.38.2",
+ "datasets>=2.16.0",
+ "sentencepiece",
+ "tqdm",
+ "psutil",
+ "wheel>=0.42.0",
+ "numpy",
+ "protobuf<4.0.0",
+]
+colab-no-deps = [
+ "accelerate>=0.26.1",
+ "trl>=0.7.9",
+ "peft>=0.7.1",
+ "xformers",
+ "bitsandbytes",
+ "protobuf<4.0.0",
+]
+colab = [
+ "unsloth[cu121]",
+]
+colab-ampere = [
+ "unsloth[colab-ampere-torch220]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+cu118-ampere = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu118only]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+cu121-ampere = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121only]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+cu118-ampere-torch211 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu118onlytorch211]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+cu121-ampere-torch211 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch211]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+cu118-ampere-torch220 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu118onlytorch220]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+cu121-ampere-torch220 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch220]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+cu118-ampere-torch230 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu118onlytorch230]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+cu121-ampere-torch230 = [
+ "unsloth[huggingface]",
+ "bitsandbytes",
+ "unsloth[cu121onlytorch230]",
+ "packaging",
+ "ninja",
+ "flash-attn",
+]
+
+[project.urls]
+homepage = "http://www.unsloth.ai"
+documentation = "https://github.com/unslothai/unsloth"
+repository = "https://github.com/unslothai/unsloth"
diff --git a/unsloth/unsloth.egg-info/PKG-INFO b/unsloth/unsloth.egg-info/PKG-INFO
new file mode 100644
index 0000000000000000000000000000000000000000..faf3e6234756d3c43ac73bf5fe7ca5f1b3beb394
--- /dev/null
+++ b/unsloth/unsloth.egg-info/PKG-INFO
@@ -0,0 +1,860 @@
+Metadata-Version: 2.1
+Name: unsloth
+Version: 2024.5
+Summary: 2-5X faster LLM finetuning
+Author: Unsloth AI team
+Author-email: info@unsloth.ai
+Maintainer-email: Daniel Han , Michael Han
+License: Apache License
+ Version 2.0, January 2004
+ http://www.apache.org/licenses/
+
+ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
+
+ 1. Definitions.
+
+ "License" shall mean the terms and conditions for use, reproduction,
+ and distribution as defined by Sections 1 through 9 of this document.
+
+ "Licensor" shall mean the copyright owner or entity authorized by
+ the copyright owner that is granting the License.
+
+ "Legal Entity" shall mean the union of the acting entity and all
+ other entities that control, are controlled by, or are under common
+ control with that entity. For the purposes of this definition,
+ "control" means (i) the power, direct or indirect, to cause the
+ direction or management of such entity, whether by contract or
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
+ outstanding shares, or (iii) beneficial ownership of such entity.
+
+ "You" (or "Your") shall mean an individual or Legal Entity
+ exercising permissions granted by this License.
+
+ "Source" form shall mean the preferred form for making modifications,
+ including but not limited to software source code, documentation
+ source, and configuration files.
+
+ "Object" form shall mean any form resulting from mechanical
+ transformation or translation of a Source form, including but
+ not limited to compiled object code, generated documentation,
+ and conversions to other media types.
+
+ "Work" shall mean the work of authorship, whether in Source or
+ Object form, made available under the License, as indicated by a
+ copyright notice that is included in or attached to the work
+ (an example is provided in the Appendix below).
+
+ "Derivative Works" shall mean any work, whether in Source or Object
+ form, that is based on (or derived from) the Work and for which the
+ editorial revisions, annotations, elaborations, or other modifications
+ represent, as a whole, an original work of authorship. For the purposes
+ of this License, Derivative Works shall not include works that remain
+ separable from, or merely link (or bind by name) to the interfaces of,
+ the Work and Derivative Works thereof.
+
+ "Contribution" shall mean any work of authorship, including
+ the original version of the Work and any modifications or additions
+ to that Work or Derivative Works thereof, that is intentionally
+ submitted to Licensor for inclusion in the Work by the copyright owner
+ or by an individual or Legal Entity authorized to submit on behalf of
+ the copyright owner. For the purposes of this definition, "submitted"
+ means any form of electronic, verbal, or written communication sent
+ to the Licensor or its representatives, including but not limited to
+ communication on electronic mailing lists, source code control systems,
+ and issue tracking systems that are managed by, or on behalf of, the
+ Licensor for the purpose of discussing and improving the Work, but
+ excluding communication that is conspicuously marked or otherwise
+ designated in writing by the copyright owner as "Not a Contribution."
+
+ "Contributor" shall mean Licensor and any individual or Legal Entity
+ on behalf of whom a Contribution has been received by Licensor and
+ subsequently incorporated within the Work.
+
+ 2. Grant of Copyright License. Subject to the terms and conditions of
+ this License, each Contributor hereby grants to You a perpetual,
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
+ copyright license to reproduce, prepare Derivative Works of,
+ publicly display, publicly perform, sublicense, and distribute the
+ Work and such Derivative Works in Source or Object form.
+
+ 3. Grant of Patent License. Subject to the terms and conditions of
+ this License, each Contributor hereby grants to You a perpetual,
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
+ (except as stated in this section) patent license to make, have made,
+ use, offer to sell, sell, import, and otherwise transfer the Work,
+ where such license applies only to those patent claims licensable
+ by such Contributor that are necessarily infringed by their
+ Contribution(s) alone or by combination of their Contribution(s)
+ with the Work to which such Contribution(s) was submitted. If You
+ institute patent litigation against any entity (including a
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
+ or a Contribution incorporated within the Work constitutes direct
+ or contributory patent infringement, then any patent licenses
+ granted to You under this License for that Work shall terminate
+ as of the date such litigation is filed.
+
+ 4. Redistribution. You may reproduce and distribute copies of the
+ Work or Derivative Works thereof in any medium, with or without
+ modifications, and in Source or Object form, provided that You
+ meet the following conditions:
+
+ (a) You must give any other recipients of the Work or
+ Derivative Works a copy of this License; and
+
+ (b) You must cause any modified files to carry prominent notices
+ stating that You changed the files; and
+
+ (c) You must retain, in the Source form of any Derivative Works
+ that You distribute, all copyright, patent, trademark, and
+ attribution notices from the Source form of the Work,
+ excluding those notices that do not pertain to any part of
+ the Derivative Works; and
+
+ (d) If the Work includes a "NOTICE" text file as part of its
+ distribution, then any Derivative Works that You distribute must
+ include a readable copy of the attribution notices contained
+ within such NOTICE file, excluding those notices that do not
+ pertain to any part of the Derivative Works, in at least one
+ of the following places: within a NOTICE text file distributed
+ as part of the Derivative Works; within the Source form or
+ documentation, if provided along with the Derivative Works; or,
+ within a display generated by the Derivative Works, if and
+ wherever such third-party notices normally appear. The contents
+ of the NOTICE file are for informational purposes only and
+ do not modify the License. You may add Your own attribution
+ notices within Derivative Works that You distribute, alongside
+ or as an addendum to the NOTICE text from the Work, provided
+ that such additional attribution notices cannot be construed
+ as modifying the License.
+
+ You may add Your own copyright statement to Your modifications and
+ may provide additional or different license terms and conditions
+ for use, reproduction, or distribution of Your modifications, or
+ for any such Derivative Works as a whole, provided Your use,
+ reproduction, and distribution of the Work otherwise complies with
+ the conditions stated in this License.
+
+ 5. Submission of Contributions. Unless You explicitly state otherwise,
+ any Contribution intentionally submitted for inclusion in the Work
+ by You to the Licensor shall be under the terms and conditions of
+ this License, without any additional terms or conditions.
+ Notwithstanding the above, nothing herein shall supersede or modify
+ the terms of any separate license agreement you may have executed
+ with Licensor regarding such Contributions.
+
+ 6. Trademarks. This License does not grant permission to use the trade
+ names, trademarks, service marks, or product names of the Licensor,
+ except as required for reasonable and customary use in describing the
+ origin of the Work and reproducing the content of the NOTICE file.
+
+ 7. Disclaimer of Warranty. Unless required by applicable law or
+ agreed to in writing, Licensor provides the Work (and each
+ Contributor provides its Contributions) on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
+ implied, including, without limitation, any warranties or conditions
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
+ PARTICULAR PURPOSE. You are solely responsible for determining the
+ appropriateness of using or redistributing the Work and assume any
+ risks associated with Your exercise of permissions under this License.
+
+ 8. Limitation of Liability. In no event and under no legal theory,
+ whether in tort (including negligence), contract, or otherwise,
+ unless required by applicable law (such as deliberate and grossly
+ negligent acts) or agreed to in writing, shall any Contributor be
+ liable to You for damages, including any direct, indirect, special,
+ incidental, or consequential damages of any character arising as a
+ result of this License or out of the use or inability to use the
+ Work (including but not limited to damages for loss of goodwill,
+ work stoppage, computer failure or malfunction, or any and all
+ other commercial damages or losses), even if such Contributor
+ has been advised of the possibility of such damages.
+
+ 9. Accepting Warranty or Additional Liability. While redistributing
+ the Work or Derivative Works thereof, You may choose to offer,
+ and charge a fee for, acceptance of support, warranty, indemnity,
+ or other liability obligations and/or rights consistent with this
+ License. However, in accepting such obligations, You may act only
+ on Your own behalf and on Your sole responsibility, not on behalf
+ of any other Contributor, and only if You agree to indemnify,
+ defend, and hold each Contributor harmless for any liability
+ incurred by, or claims asserted against, such Contributor by reason
+ of your accepting any such warranty or additional liability.
+
+ END OF TERMS AND CONDITIONS
+
+ APPENDIX: How to apply the Apache License to your work.
+
+ To apply the Apache License to your work, attach the following
+ boilerplate notice, with the fields enclosed by brackets "[]"
+ replaced with your own identifying information. (Don't include
+ the brackets!) The text should be enclosed in the appropriate
+ comment syntax for the file format. We also recommend that a
+ file or class name and description of purpose be included on the
+ same "printed page" as the copyright notice for easier
+ identification within third-party archives.
+
+ Copyright [2024-] [Unsloth AI, Daniel Han-Chen & Michael Han-Chen]
+
+ Licensed under the Apache License, Version 2.0 (the "License");
+ you may not use this file except in compliance with the License.
+ You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
+
+Project-URL: homepage, http://www.unsloth.ai
+Project-URL: documentation, https://github.com/unslothai/unsloth
+Project-URL: repository, https://github.com/unslothai/unsloth
+Keywords: ai,llm
+Classifier: Programming Language :: Python
+Requires-Python: >=3.9
+Description-Content-Type: text/markdown
+License-File: LICENSE
+Provides-Extra: huggingface
+Requires-Dist: tyro; extra == "huggingface"
+Requires-Dist: transformers>=4.38.2; extra == "huggingface"
+Requires-Dist: datasets>=2.16.0; extra == "huggingface"
+Requires-Dist: sentencepiece>=0.2.0; extra == "huggingface"
+Requires-Dist: tqdm; extra == "huggingface"
+Requires-Dist: psutil; extra == "huggingface"
+Requires-Dist: wheel>=0.42.0; extra == "huggingface"
+Requires-Dist: numpy; extra == "huggingface"
+Requires-Dist: accelerate>=0.26.1; extra == "huggingface"
+Requires-Dist: trl>=0.7.9; extra == "huggingface"
+Requires-Dist: peft>=0.7.1; extra == "huggingface"
+Requires-Dist: protobuf<4.0.0; extra == "huggingface"
+Provides-Extra: cu118only
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.22.post7%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu118only"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.22.post7%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu118only"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.22.post7%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu118only"
+Provides-Extra: cu121only
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.22.post7-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu121only"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.22.post7-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu121only"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.22.post7-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu121only"
+Provides-Extra: cu118onlytorch211
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu118onlytorch211"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu118onlytorch211"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu118onlytorch211"
+Provides-Extra: cu121onlytorch211
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu121onlytorch211"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu121onlytorch211"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu121onlytorch211"
+Provides-Extra: cu118onlytorch212
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23.post1%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu118onlytorch212"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23.post1%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu118onlytorch212"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23.post1%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu118onlytorch212"
+Provides-Extra: cu121onlytorch212
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23.post1-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu121onlytorch212"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23.post1-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu121onlytorch212"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23.post1-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu121onlytorch212"
+Provides-Extra: cu118onlytorch220
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.24%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu118onlytorch220"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.24%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu118onlytorch220"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.24%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu118onlytorch220"
+Provides-Extra: cu121onlytorch220
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu121onlytorch220"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu121onlytorch220"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu121onlytorch220"
+Provides-Extra: cu118onlytorch230
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.26.post1%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu118onlytorch230"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.26.post1%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu118onlytorch230"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.26.post1%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu118onlytorch230"
+Provides-Extra: cu121onlytorch230
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.26.post1-cp39-cp39-manylinux2014_x86_64.whl ; python_version == "3.9" and extra == "cu121onlytorch230"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.26.post1-cp310-cp310-manylinux2014_x86_64.whl ; python_version == "3.10" and extra == "cu121onlytorch230"
+Requires-Dist: xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.26.post1-cp311-cp311-manylinux2014_x86_64.whl ; python_version == "3.11" and extra == "cu121onlytorch230"
+Provides-Extra: cu118
+Requires-Dist: unsloth[huggingface]; extra == "cu118"
+Requires-Dist: bitsandbytes; extra == "cu118"
+Requires-Dist: unsloth[cu118only]; extra == "cu118"
+Provides-Extra: cu121
+Requires-Dist: unsloth[huggingface]; extra == "cu121"
+Requires-Dist: bitsandbytes; extra == "cu121"
+Requires-Dist: unsloth[cu121only]; extra == "cu121"
+Provides-Extra: cu118-torch211
+Requires-Dist: unsloth[huggingface]; extra == "cu118-torch211"
+Requires-Dist: bitsandbytes; extra == "cu118-torch211"
+Requires-Dist: unsloth[cu118onlytorch211]; extra == "cu118-torch211"
+Provides-Extra: cu121-torch211
+Requires-Dist: unsloth[huggingface]; extra == "cu121-torch211"
+Requires-Dist: bitsandbytes; extra == "cu121-torch211"
+Requires-Dist: unsloth[cu121onlytorch211]; extra == "cu121-torch211"
+Provides-Extra: cu118-torch212
+Requires-Dist: unsloth[huggingface]; extra == "cu118-torch212"
+Requires-Dist: bitsandbytes; extra == "cu118-torch212"
+Requires-Dist: unsloth[cu118onlytorch212]; extra == "cu118-torch212"
+Provides-Extra: cu121-torch212
+Requires-Dist: unsloth[huggingface]; extra == "cu121-torch212"
+Requires-Dist: bitsandbytes; extra == "cu121-torch212"
+Requires-Dist: unsloth[cu121onlytorch212]; extra == "cu121-torch212"
+Provides-Extra: cu118-torch220
+Requires-Dist: unsloth[huggingface]; extra == "cu118-torch220"
+Requires-Dist: bitsandbytes; extra == "cu118-torch220"
+Requires-Dist: unsloth[cu118onlytorch220]; extra == "cu118-torch220"
+Provides-Extra: cu121-torch220
+Requires-Dist: unsloth[huggingface]; extra == "cu121-torch220"
+Requires-Dist: bitsandbytes; extra == "cu121-torch220"
+Requires-Dist: unsloth[cu121onlytorch220]; extra == "cu121-torch220"
+Provides-Extra: cu118-torch230
+Requires-Dist: unsloth[huggingface]; extra == "cu118-torch230"
+Requires-Dist: bitsandbytes; extra == "cu118-torch230"
+Requires-Dist: unsloth[cu118onlytorch230]; extra == "cu118-torch230"
+Provides-Extra: cu121-torch230
+Requires-Dist: unsloth[huggingface]; extra == "cu121-torch230"
+Requires-Dist: bitsandbytes; extra == "cu121-torch230"
+Requires-Dist: unsloth[cu121onlytorch230]; extra == "cu121-torch230"
+Provides-Extra: kaggle
+Requires-Dist: unsloth[huggingface]; extra == "kaggle"
+Provides-Extra: kaggle-new
+Requires-Dist: unsloth[huggingface]; extra == "kaggle-new"
+Requires-Dist: bitsandbytes; extra == "kaggle-new"
+Provides-Extra: conda
+Requires-Dist: unsloth[huggingface]; extra == "conda"
+Provides-Extra: colab-torch211
+Requires-Dist: unsloth[huggingface]; extra == "colab-torch211"
+Requires-Dist: bitsandbytes; extra == "colab-torch211"
+Requires-Dist: unsloth[cu121onlytorch211]; extra == "colab-torch211"
+Provides-Extra: colab-ampere-torch211
+Requires-Dist: unsloth[huggingface]; extra == "colab-ampere-torch211"
+Requires-Dist: bitsandbytes; extra == "colab-ampere-torch211"
+Requires-Dist: unsloth[cu121onlytorch211]; extra == "colab-ampere-torch211"
+Requires-Dist: packaging; extra == "colab-ampere-torch211"
+Requires-Dist: ninja; extra == "colab-ampere-torch211"
+Requires-Dist: flash-attn; extra == "colab-ampere-torch211"
+Provides-Extra: colab-torch220
+Requires-Dist: unsloth[huggingface]; extra == "colab-torch220"
+Requires-Dist: bitsandbytes; extra == "colab-torch220"
+Requires-Dist: unsloth[cu121onlytorch220]; extra == "colab-torch220"
+Provides-Extra: colab-ampere-torch220
+Requires-Dist: unsloth[huggingface]; extra == "colab-ampere-torch220"
+Requires-Dist: bitsandbytes; extra == "colab-ampere-torch220"
+Requires-Dist: unsloth[cu121onlytorch220]; extra == "colab-ampere-torch220"
+Requires-Dist: packaging; extra == "colab-ampere-torch220"
+Requires-Dist: ninja; extra == "colab-ampere-torch220"
+Requires-Dist: flash-attn; extra == "colab-ampere-torch220"
+Provides-Extra: colab-new
+Requires-Dist: tyro; extra == "colab-new"
+Requires-Dist: transformers>=4.38.2; extra == "colab-new"
+Requires-Dist: datasets>=2.16.0; extra == "colab-new"
+Requires-Dist: sentencepiece; extra == "colab-new"
+Requires-Dist: tqdm; extra == "colab-new"
+Requires-Dist: psutil; extra == "colab-new"
+Requires-Dist: wheel>=0.42.0; extra == "colab-new"
+Requires-Dist: numpy; extra == "colab-new"
+Requires-Dist: protobuf<4.0.0; extra == "colab-new"
+Provides-Extra: colab-no-deps
+Requires-Dist: accelerate>=0.26.1; extra == "colab-no-deps"
+Requires-Dist: trl>=0.7.9; extra == "colab-no-deps"
+Requires-Dist: peft>=0.7.1; extra == "colab-no-deps"
+Requires-Dist: xformers; extra == "colab-no-deps"
+Requires-Dist: bitsandbytes; extra == "colab-no-deps"
+Requires-Dist: protobuf<4.0.0; extra == "colab-no-deps"
+Provides-Extra: colab
+Requires-Dist: unsloth[cu121]; extra == "colab"
+Provides-Extra: colab-ampere
+Requires-Dist: unsloth[colab-ampere-torch220]; extra == "colab-ampere"
+Requires-Dist: packaging; extra == "colab-ampere"
+Requires-Dist: ninja; extra == "colab-ampere"
+Requires-Dist: flash-attn; extra == "colab-ampere"
+Provides-Extra: cu118-ampere
+Requires-Dist: unsloth[huggingface]; extra == "cu118-ampere"
+Requires-Dist: bitsandbytes; extra == "cu118-ampere"
+Requires-Dist: unsloth[cu118only]; extra == "cu118-ampere"
+Requires-Dist: packaging; extra == "cu118-ampere"
+Requires-Dist: ninja; extra == "cu118-ampere"
+Requires-Dist: flash-attn; extra == "cu118-ampere"
+Provides-Extra: cu121-ampere
+Requires-Dist: unsloth[huggingface]; extra == "cu121-ampere"
+Requires-Dist: bitsandbytes; extra == "cu121-ampere"
+Requires-Dist: unsloth[cu121only]; extra == "cu121-ampere"
+Requires-Dist: packaging; extra == "cu121-ampere"
+Requires-Dist: ninja; extra == "cu121-ampere"
+Requires-Dist: flash-attn; extra == "cu121-ampere"
+Provides-Extra: cu118-ampere-torch211
+Requires-Dist: unsloth[huggingface]; extra == "cu118-ampere-torch211"
+Requires-Dist: bitsandbytes; extra == "cu118-ampere-torch211"
+Requires-Dist: unsloth[cu118onlytorch211]; extra == "cu118-ampere-torch211"
+Requires-Dist: packaging; extra == "cu118-ampere-torch211"
+Requires-Dist: ninja; extra == "cu118-ampere-torch211"
+Requires-Dist: flash-attn; extra == "cu118-ampere-torch211"
+Provides-Extra: cu121-ampere-torch211
+Requires-Dist: unsloth[huggingface]; extra == "cu121-ampere-torch211"
+Requires-Dist: bitsandbytes; extra == "cu121-ampere-torch211"
+Requires-Dist: unsloth[cu121onlytorch211]; extra == "cu121-ampere-torch211"
+Requires-Dist: packaging; extra == "cu121-ampere-torch211"
+Requires-Dist: ninja; extra == "cu121-ampere-torch211"
+Requires-Dist: flash-attn; extra == "cu121-ampere-torch211"
+Provides-Extra: cu118-ampere-torch220
+Requires-Dist: unsloth[huggingface]; extra == "cu118-ampere-torch220"
+Requires-Dist: bitsandbytes; extra == "cu118-ampere-torch220"
+Requires-Dist: unsloth[cu118onlytorch220]; extra == "cu118-ampere-torch220"
+Requires-Dist: packaging; extra == "cu118-ampere-torch220"
+Requires-Dist: ninja; extra == "cu118-ampere-torch220"
+Requires-Dist: flash-attn; extra == "cu118-ampere-torch220"
+Provides-Extra: cu121-ampere-torch220
+Requires-Dist: unsloth[huggingface]; extra == "cu121-ampere-torch220"
+Requires-Dist: bitsandbytes; extra == "cu121-ampere-torch220"
+Requires-Dist: unsloth[cu121onlytorch220]; extra == "cu121-ampere-torch220"
+Requires-Dist: packaging; extra == "cu121-ampere-torch220"
+Requires-Dist: ninja; extra == "cu121-ampere-torch220"
+Requires-Dist: flash-attn; extra == "cu121-ampere-torch220"
+Provides-Extra: cu118-ampere-torch230
+Requires-Dist: unsloth[huggingface]; extra == "cu118-ampere-torch230"
+Requires-Dist: bitsandbytes; extra == "cu118-ampere-torch230"
+Requires-Dist: unsloth[cu118onlytorch230]; extra == "cu118-ampere-torch230"
+Requires-Dist: packaging; extra == "cu118-ampere-torch230"
+Requires-Dist: ninja; extra == "cu118-ampere-torch230"
+Requires-Dist: flash-attn; extra == "cu118-ampere-torch230"
+Provides-Extra: cu121-ampere-torch230
+Requires-Dist: unsloth[huggingface]; extra == "cu121-ampere-torch230"
+Requires-Dist: bitsandbytes; extra == "cu121-ampere-torch230"
+Requires-Dist: unsloth[cu121onlytorch230]; extra == "cu121-ampere-torch230"
+Requires-Dist: packaging; extra == "cu121-ampere-torch230"
+Requires-Dist: ninja; extra == "cu121-ampere-torch230"
+Requires-Dist: flash-attn; extra == "cu121-ampere-torch230"
+
+
+
+
+
+
+
+
+
+
+
+
+
+### Finetune Llama 3, Mistral & Gemma 2-5x faster with 80% less memory!
+
+
+
+
+
+## ✨ Finetune for Free
+
+All notebooks are **beginner friendly**! Add your dataset, click "Run All", and you'll get a 2x faster finetuned model which can be exported to GGUF, vLLM or uploaded to Hugging Face.
+
+| Unsloth supports | Free Notebooks | Performance | Memory use |
+|-----------|---------|--------|----------|
+| **Llama 3 (8B)** | [▶️ Start for free](https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing) | 2x faster | 60% less |
+| **Mistral (7B)** | [▶️ Start for free](https://colab.research.google.com/drive/1Dyauq4kTZoLewQ1cApceUQVNcnnNTzg_?usp=sharing) | 2.2x faster | 73% less |
+| **Gemma (7B)** | [▶️ Start for free](https://colab.research.google.com/drive/10NbwlsRChbma1v55m8LAPYG15uQv6HLo?usp=sharing) | 2.4x faster | 71% less |
+| **ORPO** | [▶️ Start for free](https://colab.research.google.com/drive/11t4njE3c4Lxl-07OD8lJSMKkfyJml3Tn?usp=sharing) | 1.9x faster | 43% less |
+| **DPO Zephyr** | [▶️ Start for free](https://colab.research.google.com/drive/15vttTpzzVXv_tJwEk-hIcQ0S9FcEWvwP?usp=sharing) | 1.9x faster | 43% less |
+| **Phi-3 (3.8B)** | [▶️ Start for free](https://colab.research.google.com/drive/1NvkBmkHfucGO3Ve9s1NKZvMNlw5p83ym?usp=sharing) | 2x faster | 50% less |
+| **TinyLlama** | [▶️ Start for free](https://colab.research.google.com/drive/1AZghoNBQaMDgWJpi4RbffGM1h6raLUj9?usp=sharing) | 3.9x faster | 74% less |
+
+- Benchmarking compared to FA2 + Hugging Face combined.
+- **Kaggle Notebooks** for [Llama-3 8b](https://www.kaggle.com/code/danielhanchen/kaggle-llama-3-8b-unsloth-notebook), [Gemma 7b](https://www.kaggle.com/code/danielhanchen/kaggle-gemma-7b-unsloth-notebook/), [Mistral 7b](https://www.kaggle.com/code/danielhanchen/kaggle-mistral-7b-unsloth-notebook)
+- This [conversational notebook](https://colab.research.google.com/drive/1XamvWYinY6FOSX9GLvnqSjjsNflxdhNc?usp=sharing) is useful for Llama-3. And ChatML for [Mistral 7b](https://colab.research.google.com/drive/1Aau3lgPzeZKQ-98h69CCu1UJcvIBLmy2?usp=sharing).
+- This [text completion notebook](https://colab.research.google.com/drive/1ef-tab5bhkvWmBOObepl1WgJvfvSzn5Q?usp=sharing) is for continued pretraining / raw text.
+
+## 🦥 Unsloth.ai News
+- 📣 NEW! Qwen1.5-7B, Qwen1.5-14B, Qwen1.5-32B, Qwen1.5-72B now work, courtesy of Firefly's PR [#428](https://github.com/unslothai/unsloth/pull/428)
+- 📣 NEW! [Llama-3 8b](https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing) now works! Llama-3 70b also works (change the model name in the notebook).
+- 📣 NEW! [ORPO support](https://colab.research.google.com/drive/11t4njE3c4Lxl-07OD8lJSMKkfyJml3Tn?usp=sharing) is here!
+- 📣 NEW! [Phi-3 3.8b support](https://colab.research.google.com/drive/1NvkBmkHfucGO3Ve9s1NKZvMNlw5p83ym?usp=sharing) is here!
+- 📣 NEW! We cut memory usage by a [further 30%](https://unsloth.ai/blog/long-context) and now support fine-tuning of LLMs with [4x longer context windows](https://unsloth.ai/blog/long-context)! No change required if you're using our notebooks. To enable, simply change 1 line:
+```python
+model = FastLanguageModel.get_peft_model(
+ model,
+ use_gradient_checkpointing = "unsloth", # <<<<<<<
+)
+```
+- 📣 [CodeGemma](https://colab.research.google.com/drive/19lwcRk_ZQ_ZtX-qzFP3qZBBHZNcMD1hh?usp=sharing) now works along with [Gemma 7b](https://colab.research.google.com/drive/10NbwlsRChbma1v55m8LAPYG15uQv6HLo?usp=sharing) and [Gemma 2b](https://colab.research.google.com/drive/15gGm7x_jTm017_Ic8e317tdIpDG53Mtu?usp=sharing)
+- 📣 [2x faster inference](https://colab.research.google.com/drive/1aqlNQi7MMJbynFDyOQteD2t0yVfjb9Zh?usp=sharing) added for all our models
+
+## 🔗 Links and Resources
+| Type | Links |
+| ------------------------------- | --------------------------------------- |
+| 📚 **Wiki & FAQ** | [Read Our Wiki](https://github.com/unslothai/unsloth/wiki) |
+| **Twitter (aka X)** | [Follow us on X](https://twitter.com/unslothai)|
+| 📜 **Documentation** | [Read The Doc](https://github.com/unslothai/unsloth/tree/main#-documentation) |
+| 💾 **Installation** | [unsloth/README.md](https://github.com/unslothai/unsloth/tree/main#installation-instructions)|
+| 🥇 **Benchmarking** | [Performance Tables](https://github.com/unslothai/unsloth/tree/main#-performance-benchmarking)
+| 🌐 **Released Models** | [Unsloth Releases](https://huggingface.co/unsloth)|
+| ✍️ **Blog** | [Read our Blogs](https://unsloth.ai/blog)|
+
+## ⭐ Key Features
+- All kernels written in [OpenAI's Triton](https://openai.com/research/triton) language. **Manual backprop engine**.
+- **0% loss in accuracy** - no approximation methods - all exact.
+- No change of hardware. Supports NVIDIA GPUs since 2018+. Minimum CUDA Capability 7.0 (V100, T4, Titan V, RTX 20, 30, 40x, A100, H100, L40 etc) [Check your GPU!](https://developer.nvidia.com/cuda-gpus) GTX 1070, 1080 works, but is slow.
+- Works on **Linux** and **Windows** via WSL.
+- Supports 4bit and 16bit QLoRA / LoRA finetuning via [bitsandbytes](https://github.com/TimDettmers/bitsandbytes).
+- Open source trains 5x faster - see [Unsloth Pro](https://unsloth.ai/) for up to **30x faster training**!
+- If you trained a model with 🦥Unsloth, you can use this cool sticker!
+
+
+## 🥇 Performance Benchmarking
+- For the full list of **reproducable** benchmarking tables, [go to our website](https://unsloth.ai/blog/mistral-benchmark#Benchmark%20tables)
+
+| 1 A100 40GB | 🤗Hugging Face | Flash Attention | 🦥Unsloth Open Source | 🦥[Unsloth Pro](https://unsloth.ai/pricing) |
+|--------------|--------------|-----------------|---------------------|-----------------|
+| Alpaca | 1x | 1.04x | 1.98x | **15.64x** |
+| LAION Chip2 | 1x | 0.92x | 1.61x | **20.73x** |
+| OASST | 1x | 1.19x | 2.17x | **14.83x** |
+| Slim Orca | 1x | 1.18x | 2.22x | **14.82x** |
+
+- Benchmarking table below was conducted by [🤗Hugging Face](https://huggingface.co/blog/unsloth-trl).
+
+| Free Colab T4 | Dataset | 🤗Hugging Face | Pytorch 2.1.1 | 🦥Unsloth | 🦥 VRAM reduction |
+| --- | --- | --- | --- | --- | --- |
+| Llama-2 7b | OASST | 1x | 1.19x | 1.95x | -43.3% |
+| Mistral 7b | Alpaca | 1x | 1.07x | 1.56x | -13.7% |
+| Tiny Llama 1.1b | Alpaca | 1x | 2.06x | 3.87x | -73.8% |
+| DPO with Zephyr | Ultra Chat | 1x | 1.09x | 1.55x | -18.6% |
+
+
+
+## 💾 Installation Instructions
+### Conda Installation
+Select either `pytorch-cuda=11.8` for CUDA 11.8 or `pytorch-cuda=12.1` for CUDA 12.1. If you have `mamba`, use `mamba` instead of `conda` for faster solving. See this [Github issue](https://github.com/unslothai/unsloth/issues/73) for help on debugging Conda installs.
+```bash
+conda create --name unsloth_env python=3.10
+conda activate unsloth_env
+
+conda install pytorch-cuda=<12.1/11.8> pytorch cudatoolkit xformers -c pytorch -c nvidia -c xformers
+
+pip install "unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git"
+
+pip install --no-deps trl peft accelerate bitsandbytes
+```
+
+### Pip Installation
+Do **NOT** use this if you have Anaconda. You must use the Conda install method, or else stuff will BREAK.
+
+1. Find your CUDA version via
+```python
+import torch; torch.version.cuda
+```
+2. For Pytorch 2.1.0: You can update Pytorch via Pip (interchange `cu121` / `cu118`). Go to https://pytorch.org/ to learn more. Select either `cu118` for CUDA 11.8 or `cu121` for CUDA 12.1. If you have a RTX 3060 or higher (A100, H100 etc), use the `"ampere"` path. For Pytorch 2.1.1: go to step 3. For Pytorch 2.2.0: go to step 4.
+```bash
+pip install --upgrade --force-reinstall --no-cache-dir torch==2.1.0 triton \
+ --index-url https://download.pytorch.org/whl/cu121
+```
+```bash
+pip install "unsloth[cu118] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu118-ampere] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-ampere] @ git+https://github.com/unslothai/unsloth.git"
+```
+3. For Pytorch 2.1.1: Use the `"ampere"` path for newer RTX 30xx GPUs or higher.
+```bash
+pip install --upgrade --force-reinstall --no-cache-dir torch==2.1.1 triton \
+ --index-url https://download.pytorch.org/whl/cu121
+```
+```bash
+pip install "unsloth[cu118-torch211] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-torch211] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu118-ampere-torch211] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-ampere-torch211] @ git+https://github.com/unslothai/unsloth.git"
+```
+4. For Pytorch 2.2.0: Use the `"ampere"` path for newer RTX 30xx GPUs or higher.
+```bash
+pip install --upgrade --force-reinstall --no-cache-dir torch==2.2.0 triton \
+ --index-url https://download.pytorch.org/whl/cu121
+```
+```bash
+pip install "unsloth[cu118-torch220] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-torch220] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu118-ampere-torch220] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-ampere-torch220] @ git+https://github.com/unslothai/unsloth.git"
+```
+5. If you get errors, try the below first, then go back to step 1:
+```bash
+pip install --upgrade pip
+```
+6. For Pytorch 2.2.1:
+```bash
+# RTX 3090, 4090 Ampere GPUs:
+pip install "unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git"
+pip install --no-deps packaging ninja einops flash-attn xformers trl peft accelerate bitsandbytes
+
+# Pre Ampere RTX 2080, T4, GTX 1080 GPUs:
+pip install "unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git"
+pip install --no-deps xformers trl peft accelerate bitsandbytes
+```
+7. For Pytorch 2.3.0: Use the `"ampere"` path for newer RTX 30xx GPUs or higher.
+```bash
+pip install "unsloth[cu118-torch230] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-torch230] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu118-ampere-torch230] @ git+https://github.com/unslothai/unsloth.git"
+pip install "unsloth[cu121-ampere-torch230] @ git+https://github.com/unslothai/unsloth.git"
+```
+8. To troubleshoot installs try the below (all must succeed). Xformers should mostly all be available.
+```bash
+nvcc
+python -m xformers.info
+python -m bitsandbytes
+```
+
+## 📜 Documentation
+- Go to our [Wiki page](https://github.com/unslothai/unsloth/wiki) for saving to GGUF, checkpointing, evaluation and more!
+- We support Huggingface's TRL, Trainer, Seq2SeqTrainer or even Pytorch code!
+- We're in 🤗Hugging Face's official docs! Check out the [SFT docs](https://huggingface.co/docs/trl/main/en/sft_trainer#accelerate-fine-tuning-2x-using-unsloth) and [DPO docs](https://huggingface.co/docs/trl/main/en/dpo_trainer#accelerate-dpo-fine-tuning-using-unsloth)!
+
+```python
+from unsloth import FastLanguageModel
+import torch
+from trl import SFTTrainer
+from transformers import TrainingArguments
+from datasets import load_dataset
+max_seq_length = 2048 # Supports RoPE Scaling interally, so choose any!
+# Get LAION dataset
+url = "https://huggingface.co/datasets/laion/OIG/resolve/main/unified_chip2.jsonl"
+dataset = load_dataset("json", data_files = {"train" : url}, split = "train")
+
+# 4bit pre quantized models we support for 4x faster downloading + no OOMs.
+fourbit_models = [
+ "unsloth/mistral-7b-bnb-4bit",
+ "unsloth/mistral-7b-instruct-v0.2-bnb-4bit",
+ "unsloth/llama-2-7b-bnb-4bit",
+ "unsloth/gemma-7b-bnb-4bit",
+ "unsloth/gemma-7b-it-bnb-4bit", # Instruct version of Gemma 7b
+ "unsloth/gemma-2b-bnb-4bit",
+ "unsloth/gemma-2b-it-bnb-4bit", # Instruct version of Gemma 2b
+ "unsloth/llama-3-8b-bnb-4bit", # [NEW] 15 Trillion token Llama-3
+ "unsloth/Phi-3-mini-4k-instruct-bnb-4bit",
+] # More models at https://huggingface.co/unsloth
+
+model, tokenizer = FastLanguageModel.from_pretrained(
+ model_name = "unsloth/llama-3-8b-bnb-4bit",
+ max_seq_length = max_seq_length,
+ dtype = None,
+ load_in_4bit = True,
+)
+
+# Do model patching and add fast LoRA weights
+model = FastLanguageModel.get_peft_model(
+ model,
+ r = 16,
+ target_modules = ["q_proj", "k_proj", "v_proj", "o_proj",
+ "gate_proj", "up_proj", "down_proj",],
+ lora_alpha = 16,
+ lora_dropout = 0, # Supports any, but = 0 is optimized
+ bias = "none", # Supports any, but = "none" is optimized
+ # [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
+ use_gradient_checkpointing = "unsloth", # True or "unsloth" for very long context
+ random_state = 3407,
+ max_seq_length = max_seq_length,
+ use_rslora = False, # We support rank stabilized LoRA
+ loftq_config = None, # And LoftQ
+)
+
+trainer = SFTTrainer(
+ model = model,
+ train_dataset = dataset,
+ dataset_text_field = "text",
+ max_seq_length = max_seq_length,
+ tokenizer = tokenizer,
+ args = TrainingArguments(
+ per_device_train_batch_size = 2,
+ gradient_accumulation_steps = 4,
+ warmup_steps = 10,
+ max_steps = 60,
+ fp16 = not torch.cuda.is_bf16_supported(),
+ bf16 = torch.cuda.is_bf16_supported(),
+ logging_steps = 1,
+ output_dir = "outputs",
+ optim = "adamw_8bit",
+ seed = 3407,
+ ),
+)
+trainer.train()
+
+# Go to https://github.com/unslothai/unsloth/wiki for advanced tips like
+# (1) Saving to GGUF / merging to 16bit for vLLM
+# (2) Continued training from a saved LoRA adapter
+# (3) Adding an evaluation loop / OOMs
+# (4) Cutomized chat templates
+```
+
+
+## DPO Support
+DPO (Direct Preference Optimization), PPO, Reward Modelling all seem to work as per 3rd party independent testing from [Llama-Factory](https://github.com/hiyouga/LLaMA-Factory). We have a preliminary Google Colab notebook for reproducing Zephyr on Tesla T4 here: [notebook](https://colab.research.google.com/drive/15vttTpzzVXv_tJwEk-hIcQ0S9FcEWvwP?usp=sharing).
+
+We're in 🤗Hugging Face's official docs! We're on the [SFT docs](https://huggingface.co/docs/trl/main/en/sft_trainer#accelerate-fine-tuning-2x-using-unsloth) and the [DPO docs](https://huggingface.co/docs/trl/main/en/dpo_trainer#accelerate-dpo-fine-tuning-using-unsloth)!
+
+```python
+from unsloth import FastLanguageModel, PatchDPOTrainer
+PatchDPOTrainer()
+import torch
+from transformers import TrainingArguments
+from trl import DPOTrainer
+
+model, tokenizer = FastLanguageModel.from_pretrained(
+ model_name = "unsloth/zephyr-sft-bnb-4bit",
+ max_seq_length = max_seq_length,
+ dtype = None,
+ load_in_4bit = True,
+)
+
+# Do model patching and add fast LoRA weights
+model = FastLanguageModel.get_peft_model(
+ model,
+ r = 64,
+ target_modules = ["q_proj", "k_proj", "v_proj", "o_proj",
+ "gate_proj", "up_proj", "down_proj",],
+ lora_alpha = 64,
+ lora_dropout = 0, # Supports any, but = 0 is optimized
+ bias = "none", # Supports any, but = "none" is optimized
+ # [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
+ use_gradient_checkpointing = "unsloth", # True or "unsloth" for very long context
+ random_state = 3407,
+ max_seq_length = max_seq_length,
+)
+
+dpo_trainer = DPOTrainer(
+ model = model,
+ ref_model = None,
+ args = TrainingArguments(
+ per_device_train_batch_size = 4,
+ gradient_accumulation_steps = 8,
+ warmup_ratio = 0.1,
+ num_train_epochs = 3,
+ fp16 = not torch.cuda.is_bf16_supported(),
+ bf16 = torch.cuda.is_bf16_supported(),
+ logging_steps = 1,
+ optim = "adamw_8bit",
+ seed = 42,
+ output_dir = "outputs",
+ ),
+ beta = 0.1,
+ train_dataset = YOUR_DATASET_HERE,
+ # eval_dataset = YOUR_DATASET_HERE,
+ tokenizer = tokenizer,
+ max_length = 1024,
+ max_prompt_length = 512,
+)
+dpo_trainer.train()
+```
+
+## 🥇 Detailed Benchmarking Tables
+- Click "Code" for fully reproducible examples
+- "Unsloth Equal" is a preview of our PRO version, with code stripped out. All settings and the loss curve remains identical.
+- For the full list of benchmarking tables, [go to our website](https://unsloth.ai/blog/mistral-benchmark#Benchmark%20tables)
+
+| 1 A100 40GB | 🤗Hugging Face | Flash Attention 2 | 🦥Unsloth Open | Unsloth Equal | Unsloth Pro | Unsloth Max |
+|--------------|-------------|-------------|-----------------|--------------|---------------|-------------|
+| Alpaca | 1x | 1.04x | 1.98x | 2.48x | 5.32x | **15.64x** |
+| code | [Code](https://colab.research.google.com/drive/1u4dBeM-0vGNVmmO6X7cScAut-Hyt4KDF?usp=sharing) | [Code](https://colab.research.google.com/drive/1fgTOxpMbVjloQBvZyz4lF4BacKSZOB2A?usp=sharing) | [Code](https://colab.research.google.com/drive/1YIPY_18xm-K0iJDgvNkRoJsgkPMPAO3G?usp=sharing) | [Code](https://colab.research.google.com/drive/1ANW8EFL3LVyTD7Gq4TkheC1Z7Rxw-rHp?usp=sharing) | | |
+| seconds| 1040 | 1001 | 525 | 419 | 196 | 67 |
+| memory MB| 18235 | 15365 | 9631 | 8525 | | |
+| % saved| | 15.74 | 47.18 | 53.25 | | | |
+
+### Llama-Factory 3rd party benchmarking
+- [Link to performance table.](https://github.com/hiyouga/LLaMA-Factory/wiki/Performance-Comparison) TGS: tokens per GPU per second. Model: LLaMA2-7B. GPU: NVIDIA A100 * 1. Batch size: 4. Gradient accumulation: 2. LoRA rank: 8. Max length: 1024.
+
+| Method | Bits | TGS | GRAM | Speed |
+| --- | --- | --- | --- | --- |
+| HF | 16 | 2392 | 18GB | 100% |
+| HF+FA2 | 16 | 2954 | 17GB | 123% |
+| Unsloth+FA2 | 16 | 4007 | 16GB | **168%** |
+| HF | 4 | 2415 | 9GB | 101% |
+| Unsloth+FA2 | 4 | 3726 | 7GB | **160%** |
+
+### Performance comparisons between popular models
+
+ Click for specific model benchmarking tables (Mistral 7b, CodeLlama 34b etc.)
+
+### Mistral 7b
+| 1 A100 40GB | Hugging Face | Flash Attention 2 | Unsloth Open | Unsloth Equal | Unsloth Pro | Unsloth Max |
+|--------------|-------------|-------------|-----------------|--------------|---------------|-------------|
+| Mistral 7B Slim Orca | 1x | 1.15x | 2.15x | 2.53x | 4.61x | **13.69x** |
+| code | [Code](https://colab.research.google.com/drive/1mePk3KzwTD81hr5mcNcs_AX3Kbg_Ha0x?usp=sharing) | [Code](https://colab.research.google.com/drive/1dgHxjvTmX6hb0bPcLp26RXSE6_n9DKj7?usp=sharing) | [Code](https://colab.research.google.com/drive/1SKrKGV-BZoU4kv5q3g0jtE_OhRgPtrrQ?usp=sharing) | [Code](https://colab.research.google.com/drive/18yOiyX0T81mTwZqOALFSCX_tSAqju6aD?usp=sharing) | |
+| seconds | 1813 | 1571 | 842 | 718 | 393 | 132 |
+| memory MB | 32853 | 19385 | 12465 | 10271 | | |
+| % saved| | 40.99 | 62.06 | 68.74 | | |
+
+### CodeLlama 34b
+| 1 A100 40GB | Hugging Face | Flash Attention 2 | Unsloth Open | Unsloth Equal | Unsloth Pro | Unsloth Max |
+|--------------|-------------|-------------|-----------------|--------------|---------------|-------------|
+| Code Llama 34B | OOM ❌ | 0.99x | 1.87x | 2.61x | 4.27x | 12.82x |
+| code | [▶️ Code](https://colab.research.google.com/drive/1ykfz3BqrtC_AUFegCzUQjjfUNlxp6Otc?usp=sharing) | [Code](https://colab.research.google.com/drive/12ZypxQh7OC6kBXvWZI-5d05I4m-B_hoR?usp=sharing) | [Code](https://colab.research.google.com/drive/1gdHyAx8XJsz2yNV-DHvbHjR1iCef5Qmh?usp=sharing) | [Code](https://colab.research.google.com/drive/1fm7wqx9MJ0kRrwKOfmLkK1Rmw-pySahB?usp=sharing) | |
+| seconds | 1953 | 1982 | 1043 | 748 | 458 | 152 |
+| memory MB | 40000 | 33217 | 27413 | 22161 | | |
+| % saved| | 16.96| 31.47 | 44.60 | | | |
+
+### 1 Tesla T4
+
+| 1 T4 16GB | Hugging Face | Flash Attention | Unsloth Open | Unsloth Pro Equal | Unsloth Pro | Unsloth Max |
+|--------------|-------------|-----------------|-----------------|---------------|---------------|-------------|
+| Alpaca | 1x | 1.09x | 1.69x | 1.79x | 2.93x | **8.3x** |
+| code | [▶️ Code](https://colab.research.google.com/drive/1XpLIV4s8Bj5uryB-X2gqM88oRGHEGdaB?usp=sharing) | [Code](https://colab.research.google.com/drive/1LyXu6CjuymQg6ddHX8g1dpUvrMa1nn4L?usp=sharing) | [Code](https://colab.research.google.com/drive/1gsv4LpY7C32otl1rgRo5wXTk4HIitXoM?usp=sharing) | [Code](https://colab.research.google.com/drive/1VtULwRQwhEnVdNryjm27zXfdSM1tNfFK?usp=sharing) | | |
+| seconds | 1599 | 1468 | 942 | 894 | 545 | 193 |
+| memory MB | 7199 | 7059 | 6459 | 5443 | | |
+| % saved | | 1.94 | 10.28 | 24.39 | | |
+
+### 2 Tesla T4s via DDP
+
+ | 2 T4 DDP | Hugging Face | Flash Attention | Unsloth Open | Unsloth Equal | Unsloth Pro | Unsloth Max |
+|--------------|----------|-------------|-----------------|--------------|---------------|-------------|
+| Alpaca | 1x | 0.99x | 4.95x | 4.44x | 7.28x | **20.61x** |
+| code | [▶️ Code](https://www.kaggle.com/danielhanchen/hf-original-alpaca-t4-ddp) | [Code](https://www.kaggle.com/danielhanchen/hf-sdpa-alpaca-t4-ddp) | [Code](https://www.kaggle.com/danielhanchen/unsloth-alpaca-t4-ddp) | | |
+| seconds | 9882 | 9946 | 1996 | 2227 | 1357 | 480 |
+| memory MB| 9176 | 9128 | 6904 | 6782 | | |
+| % saved | | 0.52 | 24.76 | 26.09 | | | |
+
+
+### Performance comparisons on 1 Tesla T4 GPU:
+
+ Click for Time taken for 1 epoch
+
+One Tesla T4 on Google Colab
+`bsz = 2, ga = 4, max_grad_norm = 0.3, num_train_epochs = 1, seed = 3047, lr = 2e-4, wd = 0.01, optim = "adamw_8bit", schedule = "linear", schedule_steps = 10`
+
+| System | GPU | Alpaca (52K) | LAION OIG (210K) | Open Assistant (10K) | SlimOrca (518K) |
+| --- | --- | --- | --- | --- | --- |
+| Huggingface | 1 T4 | 23h 15m | 56h 28m | 8h 38m | 391h 41m |
+| Unsloth Open | 1 T4 | 13h 7m (1.8x) | 31h 47m (1.8x) | 4h 27m (1.9x) | 240h 4m (1.6x) |
+| Unsloth Pro | 1 T4 | 3h 6m (7.5x) | 5h 17m (10.7x) | 1h 7m (7.7x) | 59h 53m (6.5x) |
+| Unsloth Max | 1 T4 | 2h 39m (8.8x) | 4h 31m (12.5x) | 0h 58m (8.9x) | 51h 30m (7.6x) |
+
+**Peak Memory Usage**
+
+| System | GPU | Alpaca (52K) | LAION OIG (210K) | Open Assistant (10K) | SlimOrca (518K) |
+| --- | --- | --- | --- | --- | --- |
+| Huggingface | 1 T4 | 7.3GB | 5.9GB | 14.0GB | 13.3GB |
+| Unsloth Open | 1 T4 | 6.8GB | 5.7GB | 7.8GB | 7.7GB |
+| Unsloth Pro | 1 T4 | 6.4GB | 6.4GB | 6.4GB | 6.4GB |
+| Unsloth Max | 1 T4 | 11.4GB | 12.4GB | 11.9GB | 14.4GB |
+
+
+
+ Click for Performance Comparisons on 2 Tesla T4 GPUs via DDP:
+**Time taken for 1 epoch**
+
+Two Tesla T4s on Kaggle
+`bsz = 2, ga = 4, max_grad_norm = 0.3, num_train_epochs = 1, seed = 3047, lr = 2e-4, wd = 0.01, optim = "adamw_8bit", schedule = "linear", schedule_steps = 10`
+
+| System | GPU | Alpaca (52K) | LAION OIG (210K) | Open Assistant (10K) | SlimOrca (518K) * |
+| --- | --- | --- | --- | --- | --- |
+| Huggingface | 2 T4 | 84h 47m | 163h 48m | 30h 51m | 1301h 24m * |
+| Unsloth Pro | 2 T4 | 3h 20m (25.4x) | 5h 43m (28.7x) | 1h 12m (25.7x) | 71h 40m (18.1x) * |
+| Unsloth Max | 2 T4 | 3h 4m (27.6x) | 5h 14m (31.3x) | 1h 6m (28.1x) | 54h 20m (23.9x) * |
+
+**Peak Memory Usage on a Multi GPU System (2 GPUs)**
+
+| System | GPU | Alpaca (52K) | LAION OIG (210K) | Open Assistant (10K) | SlimOrca (518K) * |
+| --- | --- | --- | --- | --- | --- |
+| Huggingface | 2 T4 | 8.4GB \| 6GB | 7.2GB \| 5.3GB | 14.3GB \| 6.6GB | 10.9GB \| 5.9GB * |
+| Unsloth Pro | 2 T4 | 7.7GB \| 4.9GB | 7.5GB \| 4.9GB | 8.5GB \| 4.9GB | 6.2GB \| 4.7GB * |
+| Unsloth Max | 2 T4 | 10.5GB \| 5GB | 10.6GB \| 5GB | 10.6GB \| 5GB | 10.5GB \| 5GB * |
+
+* Slim Orca `bsz=1` for all benchmarks since `bsz=2` OOMs. We can handle `bsz=2`, but we benchmark it with `bsz=1` for consistency.
+
+
+
+
+
+### Thank You to
+- [HuyNguyen-hust](https://github.com/HuyNguyen-hust) for making [RoPE Embeddings 28% faster](https://github.com/unslothai/unsloth/pull/238)
+- [RandomInternetPreson](https://github.com/RandomInternetPreson) for confirming WSL support
+- [152334H](https://github.com/152334H) for experimental DPO support
+- [atgctg](https://github.com/atgctg) for syntax highlighting
diff --git a/unsloth/unsloth.egg-info/SOURCES.txt b/unsloth/unsloth.egg-info/SOURCES.txt
new file mode 100644
index 0000000000000000000000000000000000000000..884fa3baf6fe5045500bec0cbe5e47450b29a7b7
--- /dev/null
+++ b/unsloth/unsloth.egg-info/SOURCES.txt
@@ -0,0 +1,49 @@
+LICENSE
+README.md
+pyproject.toml
+.github/FUNDING.yml
+images/Colab.png
+images/Discord button.png
+images/Discord.png
+images/Free version button.png
+images/Kaggle.png
+images/Kofi button.png
+images/LAION 2GPU.png
+images/Slim Orca 2GPUs.png
+images/buy me a coffee button.png
+images/made with unsloth.png
+images/peft x trl button.png
+images/start free finetune button.png
+images/unsloth end.png
+images/unsloth loading page render.png
+images/unsloth logo black text.png
+images/unsloth logo only.png
+images/unsloth logo white text.png
+images/unsloth made with love.png
+images/unsloth new logo.png
+unsloth/__init__.py
+unsloth/chat_templates.py
+unsloth/save.py
+unsloth/tokenizer_utils.py
+unsloth.egg-info/PKG-INFO
+unsloth.egg-info/SOURCES.txt
+unsloth.egg-info/dependency_links.txt
+unsloth.egg-info/requires.txt
+unsloth.egg-info/top_level.txt
+unsloth/kernels/__init__.py
+unsloth/kernels/cross_entropy_loss.py
+unsloth/kernels/fast_lora.py
+unsloth/kernels/geglu.py
+unsloth/kernels/rms_layernorm.py
+unsloth/kernels/rope_embedding.py
+unsloth/kernels/swiglu.py
+unsloth/kernels/utils.py
+unsloth/models/__init__.py
+unsloth/models/_utils.py
+unsloth/models/dpo.py
+unsloth/models/gemma.py
+unsloth/models/llama.py
+unsloth/models/loader.py
+unsloth/models/mapper.py
+unsloth/models/mistral.py
+unsloth/models/qwen2.py
\ No newline at end of file
diff --git a/unsloth/unsloth.egg-info/dependency_links.txt b/unsloth/unsloth.egg-info/dependency_links.txt
new file mode 100644
index 0000000000000000000000000000000000000000..8b137891791fe96927ad78e64b0aad7bded08bdc
--- /dev/null
+++ b/unsloth/unsloth.egg-info/dependency_links.txt
@@ -0,0 +1 @@
+
diff --git a/unsloth/unsloth.egg-info/requires.txt b/unsloth/unsloth.egg-info/requires.txt
new file mode 100644
index 0000000000000000000000000000000000000000..df0745714db83ee1b789c2017c36d13e4ea9c37e
--- /dev/null
+++ b/unsloth/unsloth.egg-info/requires.txt
@@ -0,0 +1,302 @@
+
+[colab]
+unsloth[cu121]
+
+[colab-ampere]
+unsloth[colab-ampere-torch220]
+packaging
+ninja
+flash-attn
+
+[colab-ampere-torch211]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch211]
+packaging
+ninja
+flash-attn
+
+[colab-ampere-torch220]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch220]
+packaging
+ninja
+flash-attn
+
+[colab-new]
+tyro
+transformers>=4.38.2
+datasets>=2.16.0
+sentencepiece
+tqdm
+psutil
+wheel>=0.42.0
+numpy
+protobuf<4.0.0
+
+[colab-no-deps]
+accelerate>=0.26.1
+trl>=0.7.9
+peft>=0.7.1
+xformers
+bitsandbytes
+protobuf<4.0.0
+
+[colab-torch211]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch211]
+
+[colab-torch220]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch220]
+
+[conda]
+unsloth[huggingface]
+
+[cu118]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu118only]
+
+[cu118-ampere]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu118only]
+packaging
+ninja
+flash-attn
+
+[cu118-ampere-torch211]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu118onlytorch211]
+packaging
+ninja
+flash-attn
+
+[cu118-ampere-torch220]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu118onlytorch220]
+packaging
+ninja
+flash-attn
+
+[cu118-ampere-torch230]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu118onlytorch230]
+packaging
+ninja
+flash-attn
+
+[cu118-torch211]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu118onlytorch211]
+
+[cu118-torch212]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu118onlytorch212]
+
+[cu118-torch220]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu118onlytorch220]
+
+[cu118-torch230]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu118onlytorch230]
+
+[cu118only]
+
+[cu118only:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.22.post7%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu118only:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.22.post7%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu118only:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.22.post7%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl
+
+[cu118onlytorch211]
+
+[cu118onlytorch211:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu118onlytorch211:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu118onlytorch211:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl
+
+[cu118onlytorch212]
+
+[cu118onlytorch212:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23.post1%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu118onlytorch212:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23.post1%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu118onlytorch212:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.23.post1%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl
+
+[cu118onlytorch220]
+
+[cu118onlytorch220:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.24%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu118onlytorch220:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.24%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu118onlytorch220:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.24%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl
+
+[cu118onlytorch230]
+
+[cu118onlytorch230:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.26.post1%2Bcu118-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu118onlytorch230:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.26.post1%2Bcu118-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu118onlytorch230:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu118/xformers-0.0.26.post1%2Bcu118-cp39-cp39-manylinux2014_x86_64.whl
+
+[cu121]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121only]
+
+[cu121-ampere]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121only]
+packaging
+ninja
+flash-attn
+
+[cu121-ampere-torch211]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch211]
+packaging
+ninja
+flash-attn
+
+[cu121-ampere-torch220]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch220]
+packaging
+ninja
+flash-attn
+
+[cu121-ampere-torch230]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch230]
+packaging
+ninja
+flash-attn
+
+[cu121-torch211]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch211]
+
+[cu121-torch212]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch212]
+
+[cu121-torch220]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch220]
+
+[cu121-torch230]
+unsloth[huggingface]
+bitsandbytes
+unsloth[cu121onlytorch230]
+
+[cu121only]
+
+[cu121only:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.22.post7-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu121only:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.22.post7-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu121only:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.22.post7-cp39-cp39-manylinux2014_x86_64.whl
+
+[cu121onlytorch211]
+
+[cu121onlytorch211:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu121onlytorch211:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu121onlytorch211:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23-cp39-cp39-manylinux2014_x86_64.whl
+
+[cu121onlytorch212]
+
+[cu121onlytorch212:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23.post1-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu121onlytorch212:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23.post1-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu121onlytorch212:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.23.post1-cp39-cp39-manylinux2014_x86_64.whl
+
+[cu121onlytorch220]
+
+[cu121onlytorch220:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu121onlytorch220:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu121onlytorch220:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp39-cp39-manylinux2014_x86_64.whl
+
+[cu121onlytorch230]
+
+[cu121onlytorch230:python_version == "3.10"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.26.post1-cp310-cp310-manylinux2014_x86_64.whl
+
+[cu121onlytorch230:python_version == "3.11"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.26.post1-cp311-cp311-manylinux2014_x86_64.whl
+
+[cu121onlytorch230:python_version == "3.9"]
+xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.26.post1-cp39-cp39-manylinux2014_x86_64.whl
+
+[huggingface]
+tyro
+transformers>=4.38.2
+datasets>=2.16.0
+sentencepiece>=0.2.0
+tqdm
+psutil
+wheel>=0.42.0
+numpy
+accelerate>=0.26.1
+trl>=0.7.9
+peft>=0.7.1
+protobuf<4.0.0
+
+[kaggle]
+unsloth[huggingface]
+
+[kaggle-new]
+unsloth[huggingface]
+bitsandbytes
diff --git a/unsloth/unsloth.egg-info/top_level.txt b/unsloth/unsloth.egg-info/top_level.txt
new file mode 100644
index 0000000000000000000000000000000000000000..00fe8980685271ed927b39865c7a4bd9a34692e5
--- /dev/null
+++ b/unsloth/unsloth.egg-info/top_level.txt
@@ -0,0 +1 @@
+unsloth
diff --git a/unsloth/unsloth/__init__.py b/unsloth/unsloth/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..d4ca45d7d12cfe5c145b5a6ce28c85f4f20c8493
--- /dev/null
+++ b/unsloth/unsloth/__init__.py
@@ -0,0 +1,116 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+import os
+import warnings
+import importlib
+
+# Currently only supports 1 GPU, or else seg faults will occur.
+if "CUDA_VISIBLE_DEVICES" in os.environ:
+ devices = os.environ["CUDA_VISIBLE_DEVICES"]
+ # Check if there are multiple cuda devices set in env
+ if not devices.isdigit():
+ first_id = devices.split(",")[0]
+ warnings.warn(
+ f"Unsloth: 'CUDA_VISIBLE_DEVICES' is currently {devices} \n"\
+ "Multiple CUDA devices detected but we require a single device.\n"\
+ f"We will override CUDA_VISIBLE_DEVICES to first device: {first_id}."
+ )
+ os.environ["CUDA_VISIBLE_DEVICES"] = str(first_id)
+else:
+ # warnings.warn("Unsloth: 'CUDA_VISIBLE_DEVICES' is not set. We shall set it ourselves.")
+ os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID"
+ os.environ["CUDA_VISIBLE_DEVICES"] = "0"
+pass
+
+# Reduce VRAM usage by reducing fragmentation
+os.environ["PYTORCH_CUDA_ALLOC_CONF"] = "expandable_segments:True"
+
+try:
+ import torch
+except:
+ raise ImportError("Pytorch is not installed. Go to https://pytorch.org/.\n"\
+ "We have some installation instructions on our Github page.")
+
+# We support Pytorch 2
+# Fixes https://github.com/unslothai/unsloth/issues/38
+torch_version = torch.__version__.split(".")
+major_torch, minor_torch = torch_version[0], torch_version[1]
+major_torch, minor_torch = int(major_torch), int(minor_torch)
+if (major_torch < 2):
+ raise ImportError("Unsloth only supports Pytorch 2 for now. Please update your Pytorch to 2.1.\n"\
+ "We have some installation instructions on our Github page.")
+elif (major_torch == 2) and (minor_torch < 2):
+ # Disable expandable_segments
+ del os.environ["PYTORCH_CUDA_ALLOC_CONF"]
+pass
+
+
+# Try loading bitsandbytes and triton
+import bitsandbytes as bnb
+import triton
+from triton.common.build import libcuda_dirs
+import os
+import re
+import numpy as np
+import subprocess
+
+try:
+ cdequantize_blockwise_fp32 = bnb.functional.lib.cdequantize_blockwise_fp32
+ libcuda_dirs()
+except:
+ warnings.warn(
+ "Unsloth: Running `ldconfig /usr/lib64-nvidia` to link CUDA."\
+ )
+
+ if os.path.exists("/usr/lib64-nvidia"):
+ os.system("ldconfig /usr/lib64-nvidia")
+ elif os.path.exists("/usr/local"):
+ # Sometimes bitsandbytes cannot be linked properly in Runpod for example
+ possible_cudas = subprocess.check_output(["ls", "-al", "/usr/local"]).decode("utf-8").split("\n")
+ find_cuda = re.compile(r"[\s](cuda\-[\d\.]{2,})$")
+ possible_cudas = [find_cuda.search(x) for x in possible_cudas]
+ possible_cudas = [x.group(1) for x in possible_cudas if x is not None]
+
+ # Try linking cuda folder, or everything in local
+ if len(possible_cudas) == 0:
+ os.system(f"ldconfig /usr/local/")
+ else:
+ find_number = re.compile(r"([\d\.]{2,})")
+ latest_cuda = np.argsort([float(find_number.search(x).group(1)) for x in possible_cudas])[::-1][0]
+ latest_cuda = possible_cudas[latest_cuda]
+ os.system(f"ldconfig /usr/local/{latest_cuda}")
+ pass
+
+ importlib.reload(bnb)
+ importlib.reload(triton)
+ try:
+ import bitsandbytes as bnb
+ from triton.common.build import libcuda_dirs
+ cdequantize_blockwise_fp32 = bnb.functional.lib.cdequantize_blockwise_fp32
+ libcuda_dirs()
+ except:
+ warnings.warn(
+ "Unsloth: CUDA is not linked properly.\n"\
+ "Try running `python -m bitsandbytes` then `python -m xformers.info`\n"\
+ "We tried running `ldconfig /usr/lib64-nvidia` ourselves, but it didn't work.\n"\
+ "You need to run in your terminal `sudo ldconfig /usr/lib64-nvidia` yourself, then import Unsloth.\n"\
+ "Also try `sudo ldconfig /usr/local/cuda-xx.x` - find the latest cuda version.\n"\
+ "Unsloth will still run for now, but maybe it might crash - let's hope it works!"
+ )
+pass
+
+from .models import *
+from .save import *
+from .chat_templates import *
+from .tokenizer_utils import *
diff --git a/unsloth/unsloth/chat_templates.py b/unsloth/unsloth/chat_templates.py
new file mode 100644
index 0000000000000000000000000000000000000000..3af4c4e9ae4eaad7e297c1bc6b46f353b340613c
--- /dev/null
+++ b/unsloth/unsloth/chat_templates.py
@@ -0,0 +1,692 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+__all__ = [
+ "get_chat_template",
+ "test_chat_templates",
+ "test_hf_gguf_equivalence",
+]
+
+from transformers import StoppingCriteria, StoppingCriteriaList
+from torch import LongTensor, FloatTensor
+from transformers.models.llama.modeling_llama import logger
+from .save import patch_saving_functions
+import os
+import shutil
+from .tokenizer_utils import *
+from .models._utils import patch_tokenizer
+
+CHAT_TEMPLATES = {}
+
+# Unsloth efficient template leverages from Zephyr
+unsloth_template = \
+ "{{ bos_token }}"\
+ "{% if messages[0]['role'] == 'system' %}"\
+ "{{ messages[0]['content'] + '\n' }}"\
+ "{% set loop_messages = messages[1:] %}"\
+ "{% else %}"\
+ "{{ 'You are a helpful assistant to the user\n' }}"\
+ "{% set loop_messages = messages %}"\
+ "{% endif %}"\
+ "{% for message in loop_messages %}"\
+ "{% if message['role'] == 'user' %}"\
+ "{{ '>>> User: ' + message['content'] + '\n' }}"\
+ "{% elif message['role'] == 'assistant' %}"\
+ "{{ '>>> Assistant: ' + message['content'] + eos_token + '\n' }}"\
+ "{% else %}"\
+ "{{ raise_exception('Only user and assistant roles are supported!') }}"\
+ "{% endif %}"\
+ "{% endfor %}"\
+ "{% if add_generation_prompt %}"\
+ "{{ '>>> Assistant: ' }}"\
+ "{% endif %}"
+unsloth_eos_token = "eos_token"
+CHAT_TEMPLATES["unsloth"] = (unsloth_template, unsloth_eos_token,)
+
+
+# Zephyr has no BOS!
+zephyr_template = \
+ "{% for message in messages %}"\
+ "{% if message['role'] == 'user' %}"\
+ "{{ '<|user|>\n' + message['content'] + eos_token + '\n' }}"\
+ "{% elif message['role'] == 'assistant' %}"\
+ "{{ '<|assistant|>\n' + message['content'] + eos_token + '\n' }}"\
+ "{% else %}"\
+ "{{ '<|system|>\n' + message['content'] + eos_token + '\n' }}"\
+ "{% endif %}"\
+ "{% endfor %}"\
+ "{% if add_generation_prompt %}"\
+ "{{ '<|assistant|>\n' }}"\
+ "{% endif %}"
+zephyr_eos_token = "eos_token"
+CHAT_TEMPLATES["zephyr"] = (zephyr_template, zephyr_eos_token,)
+
+
+# ChatML has no BOS and not EOS! Rather <|im_start|> and <|im_end|> acts as BOS / EOS.
+chatml_template = \
+ "{% for message in messages %}"\
+ "{% if message['role'] == 'user' %}"\
+ "{{'<|im_start|>user\n' + message['content'] + '<|im_end|>\n'}}"\
+ "{% elif message['role'] == 'assistant' %}"\
+ "{{'<|im_start|>assistant\n' + message['content'] + '<|im_end|>\n' }}"\
+ "{% else %}"\
+ "{{ '<|im_start|>system\n' + message['content'] + '<|im_end|>\n' }}"\
+ "{% endif %}"\
+ "{% endfor %}"\
+ "{% if add_generation_prompt %}"\
+ "{{ '<|im_start|>assistant\n' }}"\
+ "{% endif %}"
+chatml_eos_token = "<|im_end|>"
+CHAT_TEMPLATES["chatml"] = (chatml_template, chatml_eos_token,)
+
+
+# Mistral Instruct doesn't allow system prompts, so we append it to the user message.
+mistral_template = \
+ "{{ bos_token }}"\
+ "{% if messages[0]['role'] == 'system' %}"\
+ "{% if messages[1]['role'] == 'user' %}"\
+ "{{ '[INST] ' + messages[0]['content'] + ' ' + messages[1]['content'] + ' [/INST]' }}"\
+ "{% set loop_messages = messages[2:] %}"\
+ "{% else %}"\
+ "{{ '[INST] ' + messages[0]['content'] + ' [/INST]' }}"\
+ "{% set loop_messages = messages[1:] %}"\
+ "{% endif %}"\
+ "{% else %}"\
+ "{% set loop_messages = messages %}"\
+ "{% endif %}"\
+ "{% for message in loop_messages %}"\
+ "{% if message['role'] == 'user' %}"\
+ "{{ '[INST] ' + message['content'] + ' [/INST]' }}"\
+ "{% elif message['role'] == 'assistant' %}"\
+ "{{ message['content'] + eos_token }}"\
+ "{% else %}"\
+ "{{ raise_exception('Only user and assistant roles are supported!') }}"\
+ "{% endif %}"\
+ "{% endfor %}"
+mistral_eos_token = "eos_token"
+CHAT_TEMPLATES["mistral"] = (mistral_template, mistral_eos_token,)
+
+
+# Adds BOS to every convo! And weird <> system messages.
+llama_template = \
+ "{% if messages[0]['role'] == 'system' %}"\
+ "{% if messages[1]['role'] == 'user' %}"\
+ "{{ bos_token + '[INST] <>\n' + messages[0]['content'] + '\n< >\n\n' + messages[1]['content'] + ' [/INST]' }}"\
+ "{% set loop_messages = messages[2:] %}"\
+ "{% else %}"\
+ "{{ bos_token + '[INST] ' + messages[0]['content'] + ' [/INST]' }}"\
+ "{% set loop_messages = messages[1:] %}"\
+ "{% endif %}"\
+ "{% else %}"\
+ "{% set loop_messages = messages %}"\
+ "{% endif %}"\
+ "{% for message in loop_messages %}"\
+ "{% if message['role'] == 'user' %}"\
+ "{{ bos_token + '[INST] ' + message['content'].strip() + ' [/INST]' }}"\
+ "{% elif message['role'] == 'assistant' %}"\
+ "{{ ' ' + message['content'].strip() + ' ' + eos_token }}"\
+ "{% else %}"\
+ "{{ raise_exception('Only user and assistant roles are supported!') }}"\
+ "{% endif %}"\
+ "{% endfor %}"
+llama_eos_token = "eos_token"
+CHAT_TEMPLATES["llama"] = (llama_template, llama_eos_token,)
+
+
+# https://github.com/lm-sys/FastChat/blob/main/docs/vicuna_weights_version.md#prompt-template
+vicuna_template = \
+ "{{ bos_token }}"\
+ "{% if messages[0]['role'] == 'system' %}"\
+ "{{ messages[0]['content'] + ' ' }}"\
+ "{% set loop_messages = messages[1:] %}"\
+ "{% else %}"\
+ "{{ 'A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user\\'s questions.' + ' ' }}"\
+ "{% set loop_messages = messages %}"\
+ "{% endif %}"\
+ "{% for message in loop_messages %}"\
+ "{% if message['role'] == 'user' %}"\
+ "{{ 'USER: ' + message['content'] + ' ' }}"\
+ "{% elif message['role'] == 'assistant' %}"\
+ "{{ 'ASSISTANT: ' + message['content'] + eos_token }}"\
+ "{% else %}"\
+ "{{ raise_exception('Only user and assistant roles are supported!') }}"\
+ "{% endif %}"\
+ "{% endfor %}"\
+ "{% if add_generation_prompt %}"\
+ "{{ 'ASSISTANT:' }}"\
+ "{% endif %}"
+vicuna_eos_token = "eos_token"
+CHAT_TEMPLATES["vicuna"] = (vicuna_template, vicuna_eos_token,)
+
+
+# https://github.com/lm-sys/FastChat/blob/main/docs/vicuna_weights_version.md#prompt-template
+vicuna_old_template = \
+ "{{ bos_token }}"\
+ "{% if messages[0]['role'] == 'system' %}"\
+ "{{ messages[0]['content'] + '\n' }}"\
+ "{% set loop_messages = messages[1:] %}"\
+ "{% else %}"\
+ "{{ 'A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human\\'s questions.' + '\n' }}"\
+ "{% set loop_messages = messages %}"\
+ "{% endif %}"\
+ "{% for message in loop_messages %}"\
+ "{% if message['role'] == 'user' %}"\
+ "{{ '### Human: ' + message['content'] + '\n' }}"\
+ "{% elif message['role'] == 'assistant' %}"\
+ "{{ '### Assistant: ' + message['content'] + eos_token + '\n' }}"\
+ "{% else %}"\
+ "{{ raise_exception('Only user and assistant roles are supported!') }}"\
+ "{% endif %}"\
+ "{% endfor %}"\
+ "{% if add_generation_prompt %}"\
+ "{{ '### Assistant:' }}"\
+ "{% endif %}"
+vicuna_old_eos_token = "eos_token"
+CHAT_TEMPLATES["vicuna_old"] = (vicuna_old_template, vicuna_old_eos_token,)
+
+
+# https://github.com/tatsu-lab/stanford_alpaca Changed for multi-turn convos
+alpaca_template = \
+ "{{ bos_token }}"\
+ "{% if messages[0]['role'] == 'system' %}"\
+ "{{ messages[0]['content'] + '\n\n' }}"\
+ "{% set loop_messages = messages[1:] %}"\
+ "{% else %}"\
+ "{{ 'Below are some instructions that describes some tasks. Write responses that appropriately completes each request.\n\n' }}"\
+ "{% set loop_messages = messages %}"\
+ "{% endif %}"\
+ "{% for message in loop_messages %}"\
+ "{% if message['role'] == 'user' %}"\
+ "{{ '### Instruction:\n' + message['content'] + '\n\n' }}"\
+ "{% elif message['role'] == 'assistant' %}"\
+ "{{ '### Response:\n' + message['content'] + eos_token + '\n\n' }}"\
+ "{% else %}"\
+ "{{ raise_exception('Only user and assistant roles are supported!') }}"\
+ "{% endif %}"\
+ "{% endfor %}"\
+ "{% if add_generation_prompt %}"\
+ "{{ '### Response:\n' }}"\
+ "{% endif %}"
+alpaca_eos_token = "eos_token"
+CHAT_TEMPLATES["alpaca"] = (alpaca_template, alpaca_eos_token,)
+
+
+# https://huggingface.co/google/gemma-7b-it
+# Notice we must use |trim for lstrip and rstrip. maps to 106.
+# maps to 107. user and model are normal 1 word tokens.
+gemma_template = \
+ "{{ bos_token }}"\
+ "{% for message in messages %}"\
+ "{% if message['role'] == 'user' %}"\
+ "{{'user\n' + message['content'] | trim + '\n'}}"\
+ "{% elif message['role'] == 'assistant' %}"\
+ "{{'model\n' + message['content'] | trim + '\n' }}"\
+ "{% else %}"\
+ "{{ 'system\n' + message['content'] | trim + '\n' }}"\
+ "{% endif %}"\
+ "{% endfor %}"\
+ "{% if add_generation_prompt %}"\
+ "{{ 'model\n' }}"\
+ "{% endif %}"
+gemma_eos_token = ""
+CHAT_TEMPLATES["gemma"] = (gemma_template, gemma_eos_token,)
+
+
+# Gemma with ChatML instead
+# We find using is still more appropriate!
+gemma_chatml_template = "{{ bos_token }}" + chatml_template
+gemma_chatml_eos_token = (
+ {"" : "<|im_start|>", "" : "<|im_end|>"},
+ "<|im_end|>",
+)
+CHAT_TEMPLATES["gemma_chatml"] = (gemma_chatml_template, gemma_chatml_eos_token,)
+
+
+# Llama-3
+# Weirdly \n\n is needed?
+llama3_template = \
+ "{{ bos_token }}"\
+ "{% for message in messages %}"\
+ "{{ '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n' + message['content'] | trim + '<|eot_id|>' }}"\
+ "{% endfor %}"\
+ "{% if add_generation_prompt %}"\
+ "{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}"\
+ "{% endif %}"
+llama3_template_eos_token = "eos_token"
+CHAT_TEMPLATES["llama-3"] = (llama3_template, llama3_template_eos_token,)
+
+
+# Phi-3
+phi3_template = \
+ "{{ bos_token }}"\
+ "{% for message in messages %}"\
+ "{{'<|' + message['role'] + '|>\n' + message['content'] + '<|end|>\n'}}"\
+ "{% endfor %}"\
+ "{% if add_generation_prompt %}"\
+ "{{ '<|assistant|>\n' }}"\
+ "{% endif %}"
+phi3_template_eos_token = "<|end|>"
+CHAT_TEMPLATES["phi-3"] = (phi3_template, phi3_template_eos_token,)
+
+
+def get_chat_template(
+ tokenizer,
+ chat_template = "chatml",
+ mapping = {"role" : "role", "content" : "content", "user" : "user", "assistant" : "assistant"},
+ map_eos_token = True,
+):
+ assert(type(map_eos_token) is bool)
+ old_tokenizer = tokenizer
+
+ IS_GEMMA = False
+ if tokenizer.__class__.__name__.startswith("Gemma"):
+ if chat_template == "chatml": chat_template = "gemma_chatml"
+ IS_GEMMA = True
+ pass
+
+ # We add a check for Llama-3
+ # if chat_template == "llama-3":
+ # tokenizer._using_llama3_template = True
+ # else:
+ # llama3_tokens = set(["<|end_header_id|>", "<|eot_id|>", "<|start_header_id|>"])
+ # check_llama3_tokens = llama3_tokens & set(str(x) for x in tokenizer.added_tokens_decoder.values())
+ # if len(check_llama3_tokens) == len(llama3_tokens):
+ # tokenizer._using_llama3_template = True
+ # pass
+ # pass
+
+ # We first check if the tokenizer is a fast one. If not, we cannot convert this!
+ is_fast_tokenizer = getattr(tokenizer, "is_fast", False)
+ old_padding_side = tokenizer.padding_side
+
+ same_padding_token = False
+
+ if type(chat_template) in (list, tuple,):
+ chat_template, stop_word = chat_template
+ assert(type(chat_template) is str)
+ assert(type(stop_word) is str)
+
+ elif type(chat_template) is str:
+
+ chat_template, stop_word = CHAT_TEMPLATES[chat_template]
+
+ if type(stop_word) in (list, tuple,):
+ token_mapping, stop_word = stop_word
+ assert(type(token_mapping) is dict)
+ else:
+ token_mapping = None
+
+ assert(type(stop_word) is str)
+
+ # Check fast tokenizer
+ if not is_fast_tokenizer:
+ print(
+ f"Unsloth: Not a fast tokenizer, so can't process it as of yet :(\n"\
+ "Please log a Github issue if you want this as a new feature!\n"\
+ "Your chat template will still work, but it won't add or edit tokens."
+ )
+
+ elif token_mapping is not None:
+ # token_mapping = {"" : "<|im_start|>", "" : "<|im_end|>"}
+ # For Gemma :)
+
+ string_vocab = tokenizer._tokenizer.to_str()
+
+ skipped = 0
+ for old_token, new_token in token_mapping.items():
+ old_count = string_vocab.count(f'"{old_token}"')
+ new_count = string_vocab.count(f'"{new_token}"')
+ if new_count != 0:
+ print(f"{new_token} is already a token. Skipping.")
+ skipped += 1
+ elif old_count == 0:
+ raise RuntimeError(f"{old_token} was not part of the tokenizer!")
+ else:
+ string_vocab = string_vocab.replace(f'"{old_token}"', f'"{new_token}"')
+ pass
+ pass
+
+ if map_eos_token and (not stop_word in token_mapping.values()):
+ # Do not map 107 = <|im_end|> and 1 = <|im_end|>. This will reduce the vocab size by 1
+ logger.warning_once(f"Unsloth: Will map {stop_word} to EOS = {tokenizer.eos_token}.")
+ string_vocab = string_vocab.replace(tokenizer.eos_token, stop_word)
+ pass
+
+ if skipped != len(token_mapping):
+ new_tokenizer = tokenizer._tokenizer.from_str(string_vocab)
+
+ # Careful on pad_token
+ old_pad_token = tokenizer.pad_token
+ if old_pad_token == tokenizer.eos_token:
+ old_pad_token = stop_word
+ same_padding_token = True
+ pass
+
+ if map_eos_token:
+ new_tokenizer = tokenizer.__class__(
+ tokenizer_object = new_tokenizer,
+ eos_token = stop_word,
+ pad_token = old_pad_token,
+ )
+ else:
+ new_tokenizer = tokenizer.__class__(
+ tokenizer_object = new_tokenizer,
+ pad_token = old_pad_token,
+ )
+ pass
+
+ # Must fix the sentence piece tokenizer since there's no tokenizer.model file!
+ tokenizer = fix_sentencepiece_tokenizer(tokenizer, new_tokenizer, token_mapping,)
+ else:
+ pass
+
+ elif map_eos_token and (stop_word != "eos_token"):
+ logger.warning_once(f"Unsloth: Will map {stop_word} to EOS = {tokenizer.eos_token}.")
+
+ # Replaces the old EOS token with a new one.
+ # Useful for ChatML <|im_end|> for example.
+ # Usually we train 2 more tokens <|im_start|> and <|im_end|>
+ # But training the lm_head and embeddings are slow!
+ # This is a HACK!
+ # Idea from https://huggingface.co/cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser
+
+ old_bos_token = getattr(tokenizer, "bos_token", None)
+ old_eos_token = getattr(tokenizer, "eos_token", None)
+ old_pad_token = getattr(tokenizer, "pad_token", None)
+ old_unk_token = getattr(tokenizer, "unk_token", None)
+
+ string_vocab = tokenizer._tokenizer.to_str()
+ # First check if new stop_word is in the tokenizer
+ if stop_word in string_vocab:
+ # We shall swap them around
+ temporary_stop_token = "<|:__TEMP//STOP//TOKEN__:|>"
+ string_vocab = string_vocab.replace(old_eos_token, temporary_stop_token)
+ string_vocab = string_vocab.replace(stop_word, old_eos_token)
+ string_vocab = string_vocab.replace(temporary_stop_token, stop_word)
+ else:
+ string_vocab = string_vocab.replace(old_eos_token, stop_word)
+ pass
+ new_tokenizer = tokenizer._tokenizer.from_str(string_vocab)
+
+ # Careful on pad_token
+ if old_pad_token == old_eos_token:
+ old_pad_token = stop_word
+ same_padding_token = True
+ pass
+
+ new_tokenizer = tokenizer.__class__(
+ tokenizer_object = new_tokenizer,
+ bos_token = old_bos_token,
+ eos_token = stop_word,
+ unk_token = old_unk_token,
+ pad_token = old_pad_token,
+ )
+
+ # Must fix the sentence piece tokenizer since there's no tokenizer.model file!
+ token_mapping = { old_eos_token : stop_word, }
+ tokenizer = fix_sentencepiece_tokenizer(tokenizer, new_tokenizer, token_mapping,)
+ pass
+
+ else:
+ raise TypeError(
+ f"Unsloth: `chat_template` must be a tuple of (your_template, eos_token,) or one of\n"\
+ f"{CHAT_TEMPLATES.keys()}"
+ )
+ pass
+
+ # For ShareGPT role -> from and content -> value
+ chat_template = chat_template\
+ .replace("'role'", "'" + mapping["role"] + "'")\
+ .replace("'content'", "'" + mapping["content"] + "'")\
+ .replace("'user'", "'" + mapping["user"] + "'")\
+ .replace("'assistant'", "'" + mapping["assistant"] + "'")
+
+ # Careful on Gemma
+ # bos_token is a must or else losses become too high
+ if IS_GEMMA and not chat_template.startswith("{{ bos_token }}"):
+ chat_template = "{{ bos_token }}" + chat_template
+ pass
+
+ _, tokenizer = patch_tokenizer(model = None, tokenizer = tokenizer)
+ tokenizer.padding_side = old_padding_side
+ tokenizer.chat_template = chat_template
+
+ # Also fix up other tokens
+ old_pad_token = getattr(old_tokenizer, "pad_token", None)
+ old_bos_token = getattr(old_tokenizer, "bos_token", None)
+ old_unk_token = getattr(old_tokenizer, "unk_token", None)
+ new_pad_token = getattr(tokenizer, "pad_token", None)
+ new_bos_token = getattr(tokenizer, "bos_token", None)
+ new_unk_token = getattr(tokenizer, "unk_token", None)
+ if old_bos_token != new_bos_token: tokenizer.bos_token = old_bos_token
+ if old_unk_token != new_unk_token: tokenizer.unk_token = old_unk_token
+ if not same_padding_token:
+ if old_pad_token != new_pad_token: tokenizer.pad_token = old_pad_token
+ pass
+
+ # stopping_criteria = create_stopping_criteria(tokenizer, stop_word)
+
+ # Patch saving functions
+ tokenizer = patch_saving_functions(tokenizer)
+
+ return tokenizer#, stopping_criteria
+pass
+
+
+def create_stopping_criteria(tokenizer, stop_word = "eos_token"):
+ class StoppingCriteriaSub(StoppingCriteria):
+ __slots__ = "stop_token", "single_match", "length",
+
+ def __init__(self, stops = "eos_token", device = "cuda", encounters = 1):
+ super().__init__()
+ if stops == "eos_token":
+ self.stop_token = torch.tensor(tokenizer.eos_token_id, device = "cuda")
+ self.length = 1
+ else:
+ self.stop_token = tokenizer(["\n" + stops], add_special_tokens = False, return_tensors = "pt")
+ self.stop_token = self.stop_token.input_ids.ravel()[1:].to("cuda")
+ self.length = self.stop_token.shape[0]
+ pass
+ self.single_match = self.length == 1
+ pass
+
+ def __call__(self, input_ids: LongTensor, scores: FloatTensor) -> bool:
+ input_ids = input_ids.ravel()
+ last_token = input_ids[-1]
+ if self.single_match and (last_token == self.stop_token): return True
+
+ if input_ids.shape[0] >= self.length and \
+ (input_ids[-self.length:] == self.stop_token).all(): return True
+ return False
+ pass
+ pass
+ stopping_criteria = StoppingCriteriaList([StoppingCriteriaSub(stops = stop_word)])
+ return stopping_criteria
+pass
+
+
+def test_chat_templates():
+ messages = [
+ {"role": "system","content": " You are a friendly chatbot.",},
+ {"role": "user", "content": "What is 2+2?"},
+ {"role": "assistant", "content": "It's 4."},
+ {"role": "user", "content": " But 2+2 is equal to 5. "},
+ {"role": "assistant", "content": "No I'm sure its 4."},
+ {"role": "user", "content": " No it's 100% 5! "},
+ ]
+
+ # Zephyr
+ from transformers import AutoTokenizer
+ template = zephyr_template
+ correct_tokenizer = AutoTokenizer.from_pretrained("HuggingFaceH4/zephyr-7b-beta")
+ correct_prompt = correct_tokenizer.apply_chat_template(messages, tokenize = False, add_generation_prompt = True)
+ correct_tokenizer.chat_template = template
+ our_prompt = correct_tokenizer.apply_chat_template(messages, tokenize = False, add_generation_prompt = True)
+ assert(correct_prompt == our_prompt)
+
+ # Chatml
+ template = chatml_template
+ correct_tokenizer = AutoTokenizer.from_pretrained("teknium/OpenHermes-2.5-Mistral-7B")
+ correct_prompt = correct_tokenizer.apply_chat_template(messages, tokenize = False, add_generation_prompt = True)
+ correct_tokenizer.chat_template = template
+ our_prompt = correct_tokenizer.apply_chat_template(messages, tokenize = False, add_generation_prompt = True)
+ assert(correct_prompt == our_prompt)
+
+ # Mistral
+ template = mistral_template
+ correct_tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
+ correct_prompt = correct_tokenizer.apply_chat_template(messages[1:], tokenize = False, add_generation_prompt = True)
+ correct_tokenizer.chat_template = template
+ our_prompt = correct_tokenizer.apply_chat_template(messages[1:], tokenize = False, add_generation_prompt = True)
+ assert(correct_prompt == our_prompt)
+
+ # Llama
+ template = llama_template
+ correct_tokenizer = AutoTokenizer.from_pretrained("unsloth/llama-2-7b-chat")
+ correct_prompt = correct_tokenizer.apply_chat_template(messages, tokenize = False, add_generation_prompt = True)
+ correct_tokenizer.chat_template = template
+ our_prompt = correct_tokenizer.apply_chat_template(messages, tokenize = False, add_generation_prompt = True)
+ assert(correct_prompt == our_prompt)
+
+ # Vicuna
+ try:
+ from fastchat.conversation import get_conv_template
+ except:
+ os.system("pip -qqq install git+https://github.com/lm-sys/FastChat.git")
+ from fastchat.conversation import get_conv_template
+ correct_prompt = get_conv_template("vicuna_v1.1")
+ for j in range(len(messages)-1):
+ correct_prompt.append_message(correct_prompt.roles[j%2==1], messages[j+1]["content"])
+ correct_prompt.append_message(correct_prompt.roles[1], "")
+ correct_prompt = tokenizer.bos_token + correct_prompt.get_prompt()
+
+ template = vicuna_template
+ correct_tokenizer = AutoTokenizer.from_pretrained("lmsys/vicuna-7b-v1.5")
+ correct_tokenizer.chat_template = template
+ our_prompt = correct_tokenizer.apply_chat_template(messages[1:], tokenize = False, add_generation_prompt = True)
+ assert(correct_prompt == our_prompt)
+
+ try:
+ from fastchat.conversation import get_conv_template
+ except:
+ os.system("pip -qqq install git+https://github.com/lm-sys/FastChat.git")
+ from fastchat.conversation import get_conv_template
+ correct_prompt = get_conv_template("zero_shot")
+ for j in range(len(messages)-1):
+ correct_prompt.append_message(correct_prompt.roles[j%2==1], messages[j+1]["content"])
+ correct_prompt.append_message(correct_prompt.roles[1], "")
+ correct_prompt = tokenizer.bos_token + correct_prompt.get_prompt()
+
+ template = vicuna_old_template
+ correct_tokenizer = AutoTokenizer.from_pretrained("lmsys/vicuna-7b-v1.5")
+ correct_tokenizer.chat_template = template
+ our_prompt = correct_tokenizer.apply_chat_template(messages[1:], tokenize = False, add_generation_prompt = True)
+ # We add ourselves
+ assert(correct_prompt == our_prompt.replace(" ", ""))
+
+ # Gemma
+ correct_tokenizer = AutoTokenizer.from_pretrained("unsloth/gemma-7b-it")
+ correct_prompt = correct_tokenizer.apply_chat_template(messages[1:], tokenize = False, add_generation_prompt = True)
+ correct_tokenizer.chat_template = gemma_template
+ our_prompt = correct_tokenizer.apply_chat_template(messages[1:], tokenize = False, add_generation_prompt = True)
+ assert(our_prompt == correct_prompt)
+
+ # Llama-3
+ template = llama3_template
+ correct_tokenizer = AutoTokenizer.from_pretrained("unsloth/llama-3-8b-Instruct")
+ correct_prompt = correct_tokenizer.apply_chat_template(messages, tokenize = False, add_generation_prompt = True)
+ correct_tokenizer.chat_template = template
+ our_prompt = correct_tokenizer.apply_chat_template(messages, tokenize = False, add_generation_prompt = True)
+ assert(correct_prompt == our_prompt)
+
+ # Phi-3
+ template = phi3_template
+ correct_tokenizer = AutoTokenizer.from_pretrained("microsoft/Phi-3-mini-4k-instruct")
+ correct_prompt = correct_tokenizer.apply_chat_template(messages[1:], tokenize = False, add_generation_prompt = True)
+ correct_tokenizer.chat_template = template
+ our_prompt = correct_tokenizer.apply_chat_template(messages[1:], tokenize = False, add_generation_prompt = True)
+ assert(correct_prompt == our_prompt)
+pass
+
+
+def test_hf_gguf_equivalence(tokenizer, gguf_model = "./model-unsloth.F16.gguf"):
+ """
+ Carefully checks the output of GGUF's tokenization and HF.
+ Can catch all tokenization bugs.
+ """
+ import subprocess
+ import re
+ messages = [
+ {"role": "user", "content": "What is 2+2?"},
+ {"role": "assistant", "content": "It's 4."},
+ {"role": "user", "content": " But 2+2 is equal to 5. "},
+ {"role": "assistant", "content": "No I'm sure its 4."},
+ {"role": "user", "content": " No it's 100% 5! "},
+ ]
+
+ prompt = """Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
+
+ ### Instruction:
+ {}
+
+ ### Input:
+ {}
+
+ ### Response:
+ {}""".format(
+ "Describe the city given eloquently.", # instruction
+ "The lost city of Atlantis.", # input
+ "", # output - leave this blank for generation!
+ )
+ prompts = [ prompt, ]
+
+ if tokenizer.chat_template is not None:
+ prompt = tokenizer.apply_chat_template(messages, tokenize = False, add_generation_prompt = True)
+ prompt = prompt.replace("'", "") # Subprocess does not like ''
+ prompts.append(prompts)
+ pass
+
+ for prompt in prompts:
+ command = f"./llama.cpp/main -m {gguf_model} -n 0 --temp 0.0 --verbose-prompt "\
+ f"--check-tensors -p '{prompt}'"
+
+ datas = []
+ with subprocess.Popen(command, shell = True, stdout = subprocess.PIPE, stderr = subprocess.STDOUT, bufsize = 1) as sp:
+ for line in sp.stdout:
+ datas.append(line.decode("utf-8", errors = "replace"))
+ pass
+ gguf_tokens = "".join(datas)
+
+ # Now extract GGUF tokenization attempt
+ gguf_tokenized = re.findall("([\d]{1,}) \-\> \'([^\']{1,})\'", gguf_tokens, flags = re.MULTILINE)
+ gguf_tokenized = [(int(x[0]), x[1],) for x in gguf_tokenized]
+ input_ids = tokenizer(prompt).input_ids
+ tokens = tokenizer.batch_decode(input_ids)
+ hf_tokenized = list(zip(input_ids, tokens))
+ print(gguf_tokenized[:5])
+
+ # Compare to Huggingface
+ for j, (hf_token, gguf_token) in enumerate(zip(hf_tokenized, gguf_tokenized)):
+ if (hf_token[0] != gguf_token[0]):
+ print("Failed GGUF != HF at", j)
+ print("HF =", hf_token)
+ print("GGUF =", gguf_token)
+ print(hf_tokenized[:j+1])
+ print(gguf_tokenized[:j+1])
+ print(gguf_tokens)
+ raise RuntimeError("Failed comparing GGUF to HF.")
+ pass
+ pass
+ return True
+pass
diff --git a/unsloth/unsloth/kernels/__init__.py b/unsloth/unsloth/kernels/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..cb04377b1a2647e8292c3b18e00e9f3acf352e75
--- /dev/null
+++ b/unsloth/unsloth/kernels/__init__.py
@@ -0,0 +1,33 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+from .cross_entropy_loss import fast_cross_entropy_loss
+from .rms_layernorm import fast_rms_layernorm
+from .rope_embedding import fast_rope_embedding, inplace_rope_embedding
+from .swiglu import swiglu_fg_kernel, swiglu_DWf_DW_dfg_kernel
+from .geglu import (
+ geglu_exact_forward_kernel,
+ geglu_exact_backward_kernel,
+ geglu_approx_forward_kernel,
+ geglu_approx_backward_kernel,
+)
+from .fast_lora import (
+ get_lora_parameters,
+ apply_lora_mlp_swiglu,
+ apply_lora_mlp_geglu_exact,
+ apply_lora_mlp_geglu_approx,
+ apply_lora_qkv,
+ apply_lora_o,
+)
+from .utils import fast_dequantize, fast_gemv, QUANT_STATE, fast_linear_forward, matmul_lora
diff --git a/unsloth/unsloth/kernels/__pycache__/__init__.cpython-310.pyc b/unsloth/unsloth/kernels/__pycache__/__init__.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..5013d4a4aa161449d473a95e7727acc58d75a588
Binary files /dev/null and b/unsloth/unsloth/kernels/__pycache__/__init__.cpython-310.pyc differ
diff --git a/unsloth/unsloth/kernels/__pycache__/cross_entropy_loss.cpython-310.pyc b/unsloth/unsloth/kernels/__pycache__/cross_entropy_loss.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..57c42b12319a66ca32602d6022662bde4297569f
Binary files /dev/null and b/unsloth/unsloth/kernels/__pycache__/cross_entropy_loss.cpython-310.pyc differ
diff --git a/unsloth/unsloth/kernels/__pycache__/fast_lora.cpython-310.pyc b/unsloth/unsloth/kernels/__pycache__/fast_lora.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..9b7aa9093b215ea757e19c9d5538f6c8d62f6c55
Binary files /dev/null and b/unsloth/unsloth/kernels/__pycache__/fast_lora.cpython-310.pyc differ
diff --git a/unsloth/unsloth/kernels/__pycache__/geglu.cpython-310.pyc b/unsloth/unsloth/kernels/__pycache__/geglu.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..32e8aa58d465ecb72aef990a89c47c0f4a69fa3c
Binary files /dev/null and b/unsloth/unsloth/kernels/__pycache__/geglu.cpython-310.pyc differ
diff --git a/unsloth/unsloth/kernels/__pycache__/rms_layernorm.cpython-310.pyc b/unsloth/unsloth/kernels/__pycache__/rms_layernorm.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..39ff50ff1f7dea33c5e6de949505b870b10a231b
Binary files /dev/null and b/unsloth/unsloth/kernels/__pycache__/rms_layernorm.cpython-310.pyc differ
diff --git a/unsloth/unsloth/kernels/__pycache__/rope_embedding.cpython-310.pyc b/unsloth/unsloth/kernels/__pycache__/rope_embedding.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..fd8e2839f89a57ebe958c99d7f70df4d33a9446f
Binary files /dev/null and b/unsloth/unsloth/kernels/__pycache__/rope_embedding.cpython-310.pyc differ
diff --git a/unsloth/unsloth/kernels/__pycache__/swiglu.cpython-310.pyc b/unsloth/unsloth/kernels/__pycache__/swiglu.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..7c491ac8b27a77d222ba1c5b9ee182dd6f4e3724
Binary files /dev/null and b/unsloth/unsloth/kernels/__pycache__/swiglu.cpython-310.pyc differ
diff --git a/unsloth/unsloth/kernels/__pycache__/utils.cpython-310.pyc b/unsloth/unsloth/kernels/__pycache__/utils.cpython-310.pyc
new file mode 100644
index 0000000000000000000000000000000000000000..76af85a8a00ec413e7f2939490f9839c8eff884a
Binary files /dev/null and b/unsloth/unsloth/kernels/__pycache__/utils.cpython-310.pyc differ
diff --git a/unsloth/unsloth/kernels/cross_entropy_loss.py b/unsloth/unsloth/kernels/cross_entropy_loss.py
new file mode 100644
index 0000000000000000000000000000000000000000..260577912f89f9f02314e631b466e81d01b3d78a
--- /dev/null
+++ b/unsloth/unsloth/kernels/cross_entropy_loss.py
@@ -0,0 +1,280 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import triton
+import triton.language as tl
+import torch
+from .utils import calculate_settings, MAX_FUSED_SIZE
+from transformers.models.llama.modeling_llama import logger
+
+
+@triton.jit
+def _cross_entropy_forward(
+ logits_ptr, logits_row_stride,
+ loss_ptr,
+ logsumexp_ptr,
+ labels_ptr,
+ VOCAB_SIZE : tl.constexpr,
+ BLOCK_SIZE : tl.constexpr,
+):
+ """
+ Cross Entropy Loss = 1/n sum [ -yi log(Pi) ]
+ Pi = exp(xi) / sum(exp(xi))
+ CE_i = -y log(p) = -y log[ exp(x) / sum(exp(x)) ]
+ = -y [ x - log[sum(exp(x))] ]
+ = y * (log[sum(exp(x))] - x)
+ If y == 0: CE_i = 0
+ If y == 1: CE_i = logsumexp - x
+
+ logsumexp is also stable
+ Take y = log[sum(exp(x))]
+ exp(y) = sum(exp(x))
+ exp(y) = sum(exp(x - c)*exp(c)) Since e^(x-c)*e^c = e^x
+ exp(y) = exp(c)*sum(exp(x - c))
+ y = log(exp(c)*sum(exp(x - c)))
+ y = c + log[sum(exp(x - c))]
+ This means we can set c = max(x) to make sure
+ exp(x - c) always is exp(x - max(x)).
+ This ensures exp(x - max(x))'s maximum is 1 as exp(0) = 1.
+ """
+ row_idx = tl.program_id(0)
+ logits_ptr += row_idx * logits_row_stride.to(tl.int64)
+ loss_ptr += row_idx
+ logsumexp_ptr += row_idx
+ labels_ptr += row_idx
+
+ col_offsets = tl.arange(0, BLOCK_SIZE)
+ mask = col_offsets < VOCAB_SIZE
+
+ label_idx = tl.load(labels_ptr).to(tl.int32)
+ logits = tl.load(logits_ptr + col_offsets, mask = mask, other = -float("inf")).to(tl.float32)
+ c = tl.max(logits, 0)
+ logsumexp = c + tl.log(tl.sum(tl.exp(logits - c), 0))
+
+ if label_idx != -100:
+ x = tl.load(logits_ptr + label_idx).to(tl.float32)
+ loss = logsumexp - x
+ else:
+ loss = 0.0
+ tl.store(logsumexp_ptr, logsumexp)
+ tl.store(loss_ptr, loss)
+pass
+
+
+@triton.jit
+def _chunked_cross_entropy_forward(
+ logits_ptr, logits_row_stride,
+ loss_ptr,
+ logsumexp_ptr,
+ labels_ptr,
+ VOCAB_SIZE : tl.constexpr,
+ N_CHUNKS : tl.constexpr,
+ BLOCK_SIZE : tl.constexpr,
+):
+ """
+ 256K vocab divided in 4 chunks
+
+ |-65536-| |-65536-| |-65536-| |-65536-|
+ |-------| |-------| |-------| |-------|
+ |-------| |-------| |-------| |-------|
+
+ If y == 0: CE_i = 0
+ If y == 1: CE_i = logsumexp - x
+
+ Notice we can do logsumexp for each chunk and then
+ logsumexp[chunk_sum(logsumexp)] == logsumexp
+
+ chunk_sum = log[chunk_sum(logsumexp)]
+ = log[exp(logsumexp(a)) + ... + exp(logsumexp(z))]
+ = log[exp(log[sum(exp(a))]) + ... + exp(log[sum(exp(z))])]
+ = log[sum(exp(a)) + ... + sum(exp(z))]
+ = logsumexp(x)
+
+ This means we can perform a logsumexp for each chunk, then do a
+ final logsumexp reduction!
+
+ Ie do: logsumexp(chunked_logsumexp) - x
+ """
+ row_idx = tl.program_id(0)
+ chunk_idx = tl.program_id(1)
+ logits_ptr += row_idx * logits_row_stride.to(tl.int64)
+ loss_ptr += row_idx
+ logsumexp_ptr += row_idx * N_CHUNKS + chunk_idx
+ labels_ptr += row_idx
+
+ col_offsets = chunk_idx*BLOCK_SIZE + tl.arange(0, BLOCK_SIZE)
+ mask = col_offsets < VOCAB_SIZE
+
+ label_idx = tl.load(labels_ptr).to(tl.int32)
+ logits = tl.load(logits_ptr + col_offsets, mask = mask, other = -float("inf")).to(tl.float32)
+ c = tl.max(logits, 0)
+ logsumexp = c + tl.log(tl.sum(tl.exp(logits - c), 0))
+
+ if chunk_idx == 0:
+ # logsumexp(chunked_logsumexp) - x
+ # Do the -x separately
+ if label_idx != -100:
+ x = tl.load(logits_ptr + label_idx).to(tl.float32)
+ loss = -1.0 * x
+ else:
+ loss = 0.0
+ tl.store(loss_ptr, loss)
+ pass
+ tl.store(logsumexp_ptr, logsumexp)
+pass
+
+
+@triton.jit
+def _cross_entropy_backward(
+ logits_ptr, logits_row_stride,
+ dloss_ptr, dloss_row_stride,
+ logsumexp_ptr,
+ labels_ptr,
+ VOCAB_SIZE : tl.constexpr,
+ BLOCK_SIZE : tl.constexpr,
+):
+ """
+ CE_i = -y log(P) = y * (log[sum(exp(x))] - x)
+ dC/dx = d/dx (y * log[sum(exp(x))] - x * y)
+
+ From https://en.wikipedia.org/wiki/LogSumExp
+ d/dx logsumexp = exp(x) / sum(exp(x)) = softmax(x)
+
+ dC/dx = y * exp(x) / sum(exp(x)) - d/dx (x * y)
+ dC/dx = y * exp[ log[exp(x) / sum(exp(x))] ] using x = exp(log(x)) trick
+ dC/dx = y * exp[x - logsumexp] - d/dx (x * y)
+
+ If y == 0: dC/dx = 0
+ If y == 1 and x == label: dC/dlabel = exp[x - logsumexp] - 1
+ If y == 1 and x != label: dC/dx = exp[x - logsumexp]
+ """
+ row_idx = tl.program_id(0)
+ block_idx = tl.program_id(1)
+
+ logits_ptr += row_idx * logits_row_stride.to(tl.int64)
+ dloss_ptr += row_idx * dloss_row_stride
+ col_offsets = block_idx*BLOCK_SIZE + tl.arange(0, BLOCK_SIZE)
+ mask = col_offsets < VOCAB_SIZE
+ label_idx = tl.load(labels_ptr + row_idx).to(tl.int32)
+
+ if label_idx != -100:
+ dloss = tl.load(dloss_ptr)
+ else:
+ dloss = 0.0
+
+ x = tl.load(logits_ptr + col_offsets, mask = mask, other = -float("inf")).to(tl.float32)
+ logsumexp = tl.load(logsumexp_ptr + row_idx)
+ y = tl.exp(x - logsumexp)
+ y = tl.where(
+ col_offsets == label_idx,
+ y - 1.0, # exp(x - logsumexp) - 1
+ y, # exp(x - logsumexp)
+ )
+
+ # If y == 0: dC/dx = 0 ==> we already masked it to be = 0, so dloss = 0.
+ tl.store(logits_ptr + col_offsets, dloss * y, mask = mask)
+pass
+
+
+MAX_FUSED_SIZE = 65536 # 2**16
+
+class Fast_CrossEntropyLoss(torch.autograd.Function):
+ @staticmethod
+ def forward(ctx, logits, labels):
+ n_rows, vocab_size = logits.shape
+
+ div, mod = divmod(vocab_size, MAX_FUSED_SIZE)
+ n_chunks = div + (mod != 0)
+ losses = torch.empty(n_rows, dtype = torch.float32, device = "cuda")
+
+ if n_chunks == 1:
+ # For small vocabs <= 65336 like Llama, Mistral
+ BLOCK_SIZE, num_warps = calculate_settings(vocab_size)
+ logsumexp = torch.empty(n_rows, dtype = torch.float32, device = "cuda")
+
+ _cross_entropy_forward[(n_rows,)](
+ logits, logits.stride(0),
+ losses,
+ logsumexp,
+ labels,
+ VOCAB_SIZE = vocab_size,
+ BLOCK_SIZE = BLOCK_SIZE,
+ num_warps = num_warps,
+ )
+ else:
+ # For large vocabs > 65336 like Gemma 256K
+ logsumexp = torch.empty((n_rows, n_chunks,), dtype = torch.float32, device = "cuda")
+
+ _chunked_cross_entropy_forward[(n_rows, n_chunks,)](
+ logits, logits.stride(0),
+ losses,
+ logsumexp,
+ labels,
+ VOCAB_SIZE = vocab_size,
+ N_CHUNKS = n_chunks,
+ BLOCK_SIZE = MAX_FUSED_SIZE,
+ num_warps = 32,
+ )
+ # logsumexp(chunked_logsumexp) - x
+ # Do the -x separately
+ logsumexp = torch.logsumexp(logsumexp, dim = 1) # Row sum
+ losses += logsumexp
+ losses.masked_fill_(labels == -100, 0) # Don't forget to mask padding out!
+ pass
+
+ ctx.save_for_backward(logits, logsumexp, labels)
+ return losses
+ pass
+
+ @staticmethod
+ def backward(ctx, dlosses):
+ logits, logsumexp, labels = ctx.saved_tensors
+ n_rows, vocab_size = logits.shape
+
+ BLOCK_SIZE = 4096
+ div, mod = divmod(vocab_size, BLOCK_SIZE)
+ n_blocks = div + (mod != 0)
+
+ _cross_entropy_backward[(n_rows, n_blocks,)](
+ logits, logits.stride(0),
+ dlosses, dlosses.stride(0),
+ logsumexp,
+ labels,
+ VOCAB_SIZE = vocab_size,
+ BLOCK_SIZE = BLOCK_SIZE,
+ num_warps = 8,
+ )
+ return logits, None, None,
+ pass
+pass
+
+
+def fast_cross_entropy_loss(logits, labels):
+ """
+ Arguments:
+ logits: (batch, seq_len, vocab_size)
+ labels: (batch, seq_len,)
+ Returns:
+ losses: float
+ """
+ batch, seq_len, d = logits.shape
+ assert(labels.shape == (batch, seq_len))
+
+ loss = Fast_CrossEntropyLoss.apply(
+ logits.view(batch*seq_len, d),
+ labels.view(-1),
+ )
+ n_items = torch.count_nonzero(labels != -100)
+ return loss.sum() / n_items
+pass
diff --git a/unsloth/unsloth/kernels/fast_lora.py b/unsloth/unsloth/kernels/fast_lora.py
new file mode 100644
index 0000000000000000000000000000000000000000..aba44f02145a02fd9962dfdee37892454a47836f
--- /dev/null
+++ b/unsloth/unsloth/kernels/fast_lora.py
@@ -0,0 +1,396 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import torch
+from .utils import fast_dequantize, QUANT_STATE, get_lora_parameters, matmul_lora
+
+
+class LoRA_MLP(torch.autograd.Function):
+ """
+ ### LoRA weights
+ G = G + Ag @ Bg
+ U = U + Au @ Bu
+ W = W + Aw @ Bw
+
+ ### SwiGLU(X)
+ e = X @ G
+ f = e * sigmoid(e)
+ g = X @ U
+ h = f * g
+ i = h @ W
+
+ ### Backpropagation chain rule
+ See our blog post for more details
+
+ df = sigmoid(e) * (1 - f) + f
+ dC/dW = h.T @ dY
+ dC/dU = X.T @ (D @ W.T * f)
+ dC/dG = X.T @ (D @ W.T * df * g)
+
+ ### Down projection LoRA weights
+ dC/dAw = dC/dW @ B.T
+ dC/dBw = A.T @ dC/dW
+ dC/dAw = h.T @ dY @ B.T
+ dC/dBw = A.T @ h.T @ dY
+
+ ### Up projection LoRA weights
+ dC/dAu = X.T @ (D @ W.T * f) @ B.T
+ dC/dBu = A.T @ X.T @ (D @ W.T * f)
+
+ ### Gate projection LoRA weights
+ dC/dAg = X.T @ (D @ W.T * df * g) @ B.T
+ dC/dBg = A.T @ X.T @ (D @ W.T * df * g)
+
+ Don't forget to see our blog post for more details!
+ """
+ @staticmethod
+ @torch.cuda.amp.custom_fwd
+ def forward(ctx, X : torch.Tensor,
+ gateW, gateW_quant, gateA, gateB, gateS,
+ upW, upW_quant, upA, upB, upS,
+ downW, downW_quant, downA, downB, downS,
+ _forward_function, _backward_function,):
+ dtype = X.dtype
+
+ e = matmul_lora(X, gateW, gateW_quant, gateA, gateB, gateS)
+ g = matmul_lora(X, upW, upW_quant, upA, upB, upS)
+ h = _forward_function(e, g)
+ i = matmul_lora(h, downW, downW_quant, downA, downB, downS)
+
+ ctx.custom_saved_tensors = (
+ gateW, gateW_quant, gateS,
+ upW, upW_quant, upS,
+ downW, downW_quant, downS,
+ _backward_function,
+ )
+ ctx.save_for_backward(gateA, gateB, upA, upB, downA, downB,
+ X, e, g)
+ return i
+ pass
+
+
+ @staticmethod
+ @torch.cuda.amp.custom_bwd
+ def backward(ctx, dY : torch.Tensor):
+ gateW, gateW_quant, gateS, upW, upW_quant, upS, downW, downW_quant, downS, \
+ _backward_function = ctx.custom_saved_tensors
+ gateA, gateB, upA, upB, downA, downB, \
+ X, e, g = ctx.saved_tensors
+
+ gateA, gateB, upA, upB, downA, downB = \
+ gateA.t(), gateB.t(), upA.t(), upB.t(), downA.t(), downB.t()
+
+ batch, seq_len, hd = X.shape
+ dY = dY.view(-1, dY.shape[-1])
+ X = X .view(-1, X .shape[-1])
+ e = e .view(-1, e .shape[-1])
+ g = g .view(-1, g .shape[-1])
+ dtype = X.dtype
+
+ DW = matmul_lora(dY, downW.t(), downW_quant, downB, downA, downS)
+ DW, e, g = _backward_function(DW, e, g)
+ h, df, de = DW, e, g
+
+ # Down projection LoRA weights
+ d_downA = h.t() @ (dY @ downB.t())
+ d_downB = (downA.t() @ h.t()) @ dY
+ d_downA *= downS
+ d_downB *= downS
+
+ # Up projection LoRA weights
+ d_upA = X.t() @ (df @ upB.t())
+ d_upB = (upA.t() @ X.t()) @ df
+ d_upA *= upS
+ d_upB *= upS
+
+ # Gate projection LoRA weights
+ d_gateA = X.t() @ (de @ gateB.t())
+ d_gateB = (gateA.t() @ X.t()) @ de
+ d_gateA *= gateS
+ d_gateB *= gateS
+
+ # dX = matmul_lora(df, upW.t(), upW_quant, upB, upA, upS)
+ # dX += matmul_lora(de, gateW.t(), gateW_quant, gateB, gateA, gateS)
+ upW = fast_dequantize(upW.t(), upW_quant)
+ dX = torch.matmul(df, upW.t(), out = X)
+ del upW
+ dX += df @ upB.to(dtype).t() @ (upS * upA.to(dtype).t())
+
+ gateW = fast_dequantize(gateW.t(), gateW_quant)
+ dX += de @ gateW.t()
+ del gateW
+ dX += de @ gateB.to(dtype).t() @ (gateS * gateA.to(dtype).t())
+
+ # gateW, gateW_quant, gateA, gateB, gateS,
+ # upW, upW_quant, upA, upB, upS,
+ # downW, downW_quant, downA, downB, downS,
+ return dX.view(batch, seq_len, hd), \
+ None, None, d_gateA.t(), d_gateB.t(), None, \
+ None, None, d_upA.t(), d_upB.t(), None, \
+ None, None, d_downA.t(), d_downB.t(), None, \
+ None, None, # _backward and _forward
+ pass
+pass
+
+
+from .swiglu import swiglu_fg_kernel, swiglu_DWf_DW_dfg_kernel
+def apply_lora_mlp_swiglu(self, X):
+ gateW, gateW_quant, gateA, gateB, gateS = get_lora_parameters(self.gate_proj)
+ upW, upW_quant, upA, upB, upS = get_lora_parameters(self. up_proj)
+ downW, downW_quant, downA, downB, downS = get_lora_parameters(self.down_proj)
+ out = LoRA_MLP.apply(X,
+ gateW, gateW_quant, gateA, gateB, gateS,
+ upW, upW_quant, upA, upB, upS,
+ downW, downW_quant, downA, downB, downS,
+ swiglu_fg_kernel, swiglu_DWf_DW_dfg_kernel,)
+ return out
+pass
+
+
+from .geglu import geglu_exact_forward_kernel, geglu_exact_backward_kernel
+def apply_lora_mlp_geglu_exact(self, X):
+ gateW, gateW_quant, gateA, gateB, gateS = get_lora_parameters(self.gate_proj)
+ upW, upW_quant, upA, upB, upS = get_lora_parameters(self. up_proj)
+ downW, downW_quant, downA, downB, downS = get_lora_parameters(self.down_proj)
+ out = LoRA_MLP.apply(X,
+ gateW, gateW_quant, gateA, gateB, gateS,
+ upW, upW_quant, upA, upB, upS,
+ downW, downW_quant, downA, downB, downS,
+ geglu_exact_forward_kernel, geglu_exact_backward_kernel,)
+ return out
+pass
+
+
+from .geglu import geglu_approx_forward_kernel, geglu_approx_backward_kernel
+def apply_lora_mlp_geglu_approx(self, X):
+ gateW, gateW_quant, gateA, gateB, gateS = get_lora_parameters(self.gate_proj)
+ upW, upW_quant, upA, upB, upS = get_lora_parameters(self. up_proj)
+ downW, downW_quant, downA, downB, downS = get_lora_parameters(self.down_proj)
+ out = LoRA_MLP.apply(X,
+ gateW, gateW_quant, gateA, gateB, gateS,
+ upW, upW_quant, upA, upB, upS,
+ downW, downW_quant, downA, downB, downS,
+ geglu_approx_forward_kernel, geglu_approx_backward_kernel,)
+ return out
+pass
+
+
+class LoRA_QKV(torch.autograd.Function):
+ """
+ ### LoRA weights
+ Wq = Wq + Aq @ Bq
+ Wk = Wk + Ak @ Bk
+ Wv = Wv + Av @ Bv
+ Q = X @ Wq = X @ Wq + X @ Aq @ Bq
+ K = X @ Wk = X @ Wk + X @ Ak @ Bk
+ V = X @ Wv = X @ Wv + X @ Av @ Bv
+
+ ### Backpropagation chain rule
+ See our blogpost for more details.
+
+ dC/dWq = X.T @ D(Wq)
+ dC/dWk = X.T @ D(Wk)
+ dC/dWv = X.T @ D(Wv)
+ We then sum them all find dC/dX
+
+ ### Q projection LoRA weights
+ dC/dAq = X.T @ D(Wq) @ B.T
+ dC/dBq = A.T @ X.T @ D(Wq)
+
+ ### K projection LoRA weights
+ dC/dAk = X.T @ D(Wk) @ B.T
+ dC/dBk = A.T @ X.T @ D(Wk)
+
+ ### V projection LoRA weights
+ dC/dAv = X.T @ D(Wv) @ B.T
+ dC/dBv = A.T @ X.T @ D(Wv)
+ """
+ @staticmethod
+ @torch.cuda.amp.custom_fwd
+ def forward(ctx, X : torch.Tensor,
+ QW, QW_quant, QA, QB, QS,
+ KW, KW_quant, KA, KB, KS,
+ VW, VW_quant, VA, VB, VS,):
+ dtype = X.dtype
+
+ Q = matmul_lora(X, QW, QW_quant, QA, QB, QS)
+ K = matmul_lora(X, KW, KW_quant, KA, KB, KS)
+ V = matmul_lora(X, VW, VW_quant, VA, VB, VS)
+
+ ctx.custom_saved_tensors = (
+ QW, QW_quant, QS,
+ KW, KW_quant, KS,
+ VW, VW_quant, VS,
+ )
+ ctx.save_for_backward(X, QA, QB, KA, KB, VA, VB,)
+ return Q, K, V
+ pass
+
+ @staticmethod
+ @torch.cuda.amp.custom_bwd
+ def backward(ctx, dQ, dK, dV):
+ QW, QW_quant, QS, KW, KW_quant, KS, VW, VW_quant, VS = \
+ ctx.custom_saved_tensors
+ X, QA, QB, KA, KB, VA, VB, = ctx.saved_tensors
+
+ QA, QB, KA, KB, VA, VB = \
+ QA.t(), QB.t(), KA.t(), KB.t(), VA.t(), VB.t()
+
+ batch, seq_len, hd = X.shape
+ dQ = dQ.view(-1, dQ.shape[-1])
+ dK = dK.reshape(-1, dK.shape[-1]) # view doesn't work on K.T
+ dV = dV.view(-1, dV.shape[-1])
+ X = X .view(-1, X .shape[-1])
+ dtype = X.dtype
+
+ ### Weight projection LoRA weights
+ # See our blogpost for more details.
+
+ # Q Projection
+ d_QA = X.t() @ (dQ @ QB.t())
+ d_QB = (QA.t() @ X.t()) @ dQ
+ d_QA *= QS
+ d_QB *= QS
+
+ # K Projection
+ d_KA = X.t() @ (dK @ KB.t())
+ d_KB = (KA.t() @ X.t()) @ dK
+ d_KA *= KS
+ d_KB *= KS
+
+ # V Projection
+ d_VA = X.t() @ (dV @ VB.t())
+ d_VB = (VA.t() @ X.t()) @ dV
+ d_VA *= VS
+ d_VB *= VS
+
+ # Combine derivatives to find dX
+ # dQ
+ QW = fast_dequantize(QW.t(), QW_quant)
+ dX = torch.matmul(dQ, QW.t(), out = X)
+ del QW
+ dX += (dQ @ QB.to(dtype).t() @ (QS * QA.to(dtype).t()))
+
+ # dK
+ KW = fast_dequantize(KW.t(), KW_quant)
+ dX += dK @ KW.t()
+ del KW
+ dX += dK @ KB.to(dtype).t() @ (KS * KA.to(dtype).t())
+
+ # dV
+ VW = fast_dequantize(VW.t(), VW_quant)
+ dX += dV @ VW.t()
+ del VW
+ dX += dV @ VB.to(dtype).t() @ (VS * VA.to(dtype).t())
+
+ # QW, QW_quant, QA, QB, QS,
+ # KW, KW_quant, KA, KB, KS,
+ # VW, VW_quant, VA, VB, VS,
+ return dX.view(batch, seq_len, hd), \
+ None, None, d_QA.t(), d_QB.t(), None, \
+ None, None, d_KA.t(), d_KB.t(), None, \
+ None, None, d_VA.t(), d_VB.t(), None
+ pass
+pass
+
+
+def apply_lora_qkv(self, X):
+ QW, QW_quant, QA, QB, QS = get_lora_parameters(self.q_proj)
+ KW, KW_quant, KA, KB, KS = get_lora_parameters(self.k_proj)
+ VW, VW_quant, VA, VB, VS = get_lora_parameters(self.v_proj)
+ Q, K, V = LoRA_QKV.apply(X,
+ QW, QW_quant, QA, QB, QS,
+ KW, KW_quant, KA, KB, KS,
+ VW, VW_quant, VA, VB, VS,
+ )
+ return Q, K, V
+pass
+
+
+class LoRA_W(torch.autograd.Function):
+ """
+ ### LoRA weights
+ Wq = Wq + Aq @ Bq
+ Wk = Wk + Ak @ Bk
+ Wv = Wv + Av @ Bv
+ Q = X @ Wq = X @ Wq + X @ Aq @ Bq
+ K = X @ Wk = X @ Wk + X @ Ak @ Bk
+ V = X @ Wv = X @ Wv + X @ Av @ Bv
+
+ ### Backpropagation chain rule
+ dC/dWq = X.T @ D(Wq)
+ dC/dWk = X.T @ D(Wk)
+ dC/dWv = X.T @ D(Wv)
+
+ ### Q projection LoRA weights
+ dC/dAq = X.T @ D(Wq) @ B.T
+ dC/dBq = A.T @ X.T @ D(Wq)
+
+ ### K projection LoRA weights
+ dC/dAk = X.T @ D(Wk) @ B.T
+ dC/dBk = A.T @ X.T @ D(Wk)
+
+ ### V projection LoRA weights
+ dC/dAv = X.T @ D(Wv) @ B.T
+ dC/dBv = A.T @ X.T @ D(Wv)
+ """
+ @staticmethod
+ @torch.cuda.amp.custom_fwd
+ def forward(ctx, X : torch.Tensor,
+ W, W_quant, A, B, S):
+ dtype = X.dtype
+ XW = matmul_lora(X, W, W_quant, A, B, S)
+ ctx.custom_saved_tensors = (W, W_quant, S,)
+ ctx.save_for_backward(A, B, X)
+ return XW
+ pass
+
+ @staticmethod
+ @torch.cuda.amp.custom_bwd
+ def backward(ctx, dY : torch.Tensor):
+ W, W_quant, S = ctx.custom_saved_tensors
+ A, B, X = ctx.saved_tensors
+
+ A, B = A.t(), B.t()
+
+ batch, seq_len, hd = X.shape
+ dY = dY.reshape(-1, dY.shape[-1]) # Must be reshape
+ X = X .reshape(-1, X .shape[-1]) # Must be reshape
+ dtype = X.dtype
+
+ ### Weight projection LoRA weights
+ # Weight projection
+ d_A = X.t() @ (dY @ B.t())
+ d_B = (A.t() @ X.t()) @ dY
+ d_A *= S
+ d_B *= S
+
+ # Get derivative for dX
+ W = fast_dequantize(W.t(), W_quant)
+ dX = dY @ W.t()
+ del W
+ dX += dY @ B.to(dtype).t() @ (S * A.to(dtype).t())
+
+ # W, W_quant, A, B, S
+ return dX.view(batch, seq_len, hd), \
+ None, None, d_A.t(), d_B.t(), None
+ pass
+pass
+
+
+def apply_lora_o(self, X):
+ OW, OW_quant, OA, OB, OS = get_lora_parameters(self.o_proj)
+ O = LoRA_W.apply(X, OW, OW_quant, OA, OB, OS)
+ return O
+pass
diff --git a/unsloth/unsloth/kernels/geglu.py b/unsloth/unsloth/kernels/geglu.py
new file mode 100644
index 0000000000000000000000000000000000000000..df80fcb79bfebd535a3fecc897cc5edd3e1e09a7
--- /dev/null
+++ b/unsloth/unsloth/kernels/geglu.py
@@ -0,0 +1,203 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import triton
+import triton.language as tl
+import torch
+from .utils import calculate_settings
+
+
+@triton.jit
+def _exact_forward_kernel(e, g, h, n_elements, BLOCK_SIZE : tl.constexpr,):
+ block_idx = tl.program_id(0)
+ offsets = block_idx*BLOCK_SIZE + tl.arange(0, BLOCK_SIZE)
+ mask = offsets < n_elements
+
+ # f = 1/2 * e * (1 + erf(1/sqrt(2) * e))
+ # h = f * up
+ e_row = tl.load(e + offsets, mask = mask, other = 0).to(tl.float32)
+ g_row = tl.load(g + offsets, mask = mask, other = 0)#.to(tl.float32)
+
+ f_row = 0.5 * e_row * (tl.math.erf(tl.math.rsqrt(2.0) * e_row) + 1.0)
+ f_row = f_row.to(g_row.dtype) # Exact copy from HF
+ h_row = f_row * g_row
+
+ # Store h
+ tl.store(h + offsets, h_row, mask = mask)
+pass
+
+
+def geglu_exact_forward_kernel(gate, up):
+ batch, seq_len, hd = gate.shape
+ n_elements = gate.numel()
+ out = torch.empty((batch, seq_len, hd), dtype = gate.dtype, device = "cuda")
+ grid = lambda meta: (triton.cdiv(n_elements, meta['BLOCK_SIZE']),)
+ _exact_forward_kernel[grid](gate, up, out, n_elements, BLOCK_SIZE = 1024,)
+ return out
+pass
+
+
+@triton.jit
+def _exact_backward_kernel(DW, e, g, n_elements, BLOCK_SIZE : tl.constexpr,):
+ """
+ f = 1/2 * e * (1 + erf(1/sqrt(2) * e))
+ h = f * up
+
+ df/de (with help of Wolfram :)
+ df/de = 1/2 * (1 + erf(1/sqrt(2) * e)) + 1/sqrt(2*pi) * e * exp(-1/2 * e^2)
+
+ Reuse via
+ f = 1/2 * (1 + erf(1/sqrt(2) * e)) * e
+ """
+ block_idx = tl.program_id(0)
+ offsets = block_idx*BLOCK_SIZE + tl.arange(0, BLOCK_SIZE)
+ mask = offsets < n_elements
+
+ DW_row = tl.load(DW + offsets, mask = mask, other = 0)#.to(tl.float32)
+ e_row = tl.load(e + offsets, mask = mask, other = 0).to(tl.float32)
+ g_row = tl.load(g + offsets, mask = mask, other = 0)#.to(tl.float32)
+
+ # Break e_row away for re-use
+ # f = 1/2 * e * (1 + erf(1/sqrt(2) * e))
+ f_partial_row = 0.5 * (tl.math.erf(tl.math.rsqrt(2.0) * e_row) + 1.0)
+ f_row = f_partial_row * e_row
+
+ f_row = f_row.to(DW_row.dtype)
+ # h = f * g
+ h_row = f_row * g_row
+ # df = DW * f
+ df_row = DW_row * f_row
+ # dg = DW * g
+ dg_row = DW_row * g_row
+
+ # df/de = 1/2 * (1 + erf(1/sqrt(2) * e)) + 1/sqrt(2*pi) * e * exp(-1/2 * e^2)
+ t = 0.3989422804014327 # 1/sqrt(2*pi)
+ df_de = f_partial_row + t * e_row * tl.exp(-0.5 * e_row * e_row)
+
+ de_row = dg_row.to(tl.float32) * df_de
+ de_row = de_row.to(DW_row.dtype)
+
+ # Store derivatives in buffers
+ tl.store(DW + offsets, h_row, mask = mask) # h = f * g
+ tl.store(e + offsets, df_row, mask = mask) # df = DW * f
+ tl.store(g + offsets, de_row, mask = mask) # de
+pass
+
+
+def geglu_exact_backward_kernel(DW, e, g):
+ batch_seq_len, hd = e.shape
+ n_elements = e.numel()
+ grid = lambda meta: (triton.cdiv(n_elements, meta['BLOCK_SIZE']),)
+ _exact_backward_kernel[grid](DW, e, g, n_elements, BLOCK_SIZE = 1024,)
+ return DW, e, g
+pass
+
+
+@triton.jit
+def _approx_forward_kernel(e, g, h, n_elements, BLOCK_SIZE : tl.constexpr,):
+ block_idx = tl.program_id(0)
+ offsets = block_idx*BLOCK_SIZE + tl.arange(0, BLOCK_SIZE)
+ mask = offsets < n_elements
+
+ # f = 1/2 * e * (1 + tanh( sqrt(2/pi) * (x + 0.044715 * x^3 ) ))
+ # f = 1/2 * e * (1 + tanh( sqrt(2/pi) * x * (1 + 0.044715 * x^2 ) ))
+ # h = f * up
+ s = 0.7978845608028654 # math.sqrt(2 / math.pi)
+
+ e_row = tl.load(e + offsets, mask = mask, other = 0).to(tl.float32)
+ g_row = tl.load(g + offsets, mask = mask, other = 0)#.to(tl.float32)
+
+ f_row = 0.5 * e_row * (
+ tl.math.tanh(s * e_row * (1.0 + 0.044715 * e_row * e_row)) \
+ + 1.0
+ )
+ f_row = f_row.to(g_row.dtype) # Exact copy from HF
+ h_row = f_row * g_row
+
+ # Store h
+ tl.store(h + offsets, h_row, mask = mask)
+pass
+
+
+def geglu_approx_forward_kernel(gate, up):
+ batch, seq_len, hd = gate.shape
+ n_elements = gate.numel()
+ out = torch.empty((batch, seq_len, hd), dtype = gate.dtype, device = "cuda")
+ grid = lambda meta: (triton.cdiv(n_elements, meta['BLOCK_SIZE']),)
+ _approx_forward_kernel[grid](gate, up, out, n_elements, BLOCK_SIZE = 1024,)
+ return out
+pass
+
+
+@triton.jit
+def _approx_backward_kernel(DW, e, g, n_elements, BLOCK_SIZE : tl.constexpr,):
+ """
+ f = 1/2 * e * (1 + tanh( sqrt(2/pi) * x * (1 + 0.044715 * x^2 ) ))
+ h = f * up
+
+ df/de (with help from https://arxiv.org/pdf/2305.12073.pdf :))
+ df/de = 1/2 * [1 + tanh( sqrt(2/pi) * x * (1 + 0.044715 * x^2 ) )] +
+ 1/2 * sech^2 [ sqrt(2/pi) * x * (1 + 0.044715 * x^2 ) ] * \
+ ( sqrt(2/pi) * x * (1 + 0.044715 * x^2 * 3 ) )
+
+ Notice sech^2(x) = 1 - tanh^2(x)
+ So reuse tanh( sqrt(2/pi) * x * (1 + 0.044715 * x^2 ) )
+
+ See https://www.desmos.com/calculator/nqprfoni6x
+ """
+ block_idx = tl.program_id(0)
+ offsets = block_idx*BLOCK_SIZE + tl.arange(0, BLOCK_SIZE)
+ mask = offsets < n_elements
+
+ DW_row = tl.load(DW + offsets, mask = mask, other = 0)#.to(tl.float32)
+ e_row = tl.load(e + offsets, mask = mask, other = 0).to(tl.float32)
+ g_row = tl.load(g + offsets, mask = mask, other = 0)#.to(tl.float32)
+
+ # See https://www.desmos.com/calculator/nqprfoni6x
+ s = 0.7978845608028654 # math.sqrt(2 / math.pi)
+ a = s * e_row # a = sqrt(2 / pi) * x
+ b = a * 0.044715 * e_row * e_row # b = a * 0.044715 * x^2
+ T = 1.0 + tl.math.tanh(a + b)
+ T2 = 0.5 * T
+ # Q = 0.5 * -T * (T - 2.0) * (a + 3.0 * b)
+ Q2 = -T2 * (T - 2.0) * (a + 3.0 * b)
+ df_de = T2 + Q2 # 1/2 * (T + Q)
+
+ # f = 1/2 * e * (1 + tanh( sqrt(2/pi) * (x + 0.044715 * x^3 ) ))
+ f_row = T2 * e_row
+ f_row = f_row.to(DW_row.dtype)
+ # h = f * g
+ h_row = f_row * g_row
+ # df = DW * f
+ df_row = DW_row * f_row
+ # dg = DW * g
+ dg_row = DW_row * g_row
+
+ de_row = dg_row.to(tl.float32) * df_de
+ de_row = de_row.to(DW_row.dtype)
+
+ # Store derivatives in buffers
+ tl.store(DW + offsets, h_row, mask = mask) # h = f * g
+ tl.store(e + offsets, df_row, mask = mask) # df = DW * f
+ tl.store(g + offsets, de_row, mask = mask) # de
+pass
+
+
+def geglu_approx_backward_kernel(DW, e, g):
+ batch_seq_len, hd = e.shape
+ n_elements = e.numel()
+ grid = lambda meta: (triton.cdiv(n_elements, meta['BLOCK_SIZE']),)
+ _approx_backward_kernel[grid](DW, e, g, n_elements, BLOCK_SIZE = 1024,)
+ return DW, e, g
+pass
diff --git a/unsloth/unsloth/kernels/rms_layernorm.py b/unsloth/unsloth/kernels/rms_layernorm.py
new file mode 100644
index 0000000000000000000000000000000000000000..4db89b781633c4c22405363b6ea0c22dafedb763
--- /dev/null
+++ b/unsloth/unsloth/kernels/rms_layernorm.py
@@ -0,0 +1,192 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import triton
+import triton.language as tl
+import torch
+from .utils import calculate_settings
+
+
+@triton.jit
+def _rms_layernorm_forward(
+ Y, Y_row_stride,
+ X, X_row_stride,
+ W, W_row_stride,
+ r, r_row_stride,
+ n_cols, eps,
+ BLOCK_SIZE : tl.constexpr
+):
+ """
+ Fast RMS Layernorm kernel
+ Inspiration from a Triton tutorial:
+ https://triton-lang.org/main/getting-started/tutorials/05-layer-norm.html
+ """
+ row_idx = tl.program_id(0)
+ col_offsets = tl.arange(0, BLOCK_SIZE)
+ mask = col_offsets < n_cols
+
+ Y += row_idx * Y_row_stride
+ X += row_idx * X_row_stride
+ r += row_idx * r_row_stride
+
+ X_row = tl.load(X + col_offsets, mask = mask, other = 0).to(tl.float32)
+ W_row = tl.load(W + col_offsets, mask = mask, other = 0)#.to(tl.float32)
+
+ row_var = tl.sum(X_row * X_row, axis = 0) / n_cols
+ inv_var = tl.math.rsqrt(row_var + eps)
+ tl.store(r, inv_var)
+ normed = X_row * inv_var
+ normed = normed.to(W_row.dtype) # Exact copy from HF
+ output = normed * W_row
+ tl.store(Y + col_offsets, output, mask = mask)
+pass
+
+
+@triton.heuristics({"GEMMA": lambda args: args["GEMMA"],})
+@triton.jit
+def _rms_layernorm_backward(
+ dY, dY_row_stride,
+ X, X_row_stride,
+ W, W_row_stride,
+ r, r_row_stride,
+ dW, dW_row_stride,
+ n_cols, eps,
+ GEMMA : tl.constexpr,
+ BLOCK_SIZE : tl.constexpr,
+):
+ """
+ Fast RMS Layernorm kernel for the backward pass
+ Inspiration from a Triton tutorial:
+ https://triton-lang.org/main/getting-started/tutorials/05-layer-norm.html
+ """
+ row_idx = tl.program_id(0)
+ col_offsets = tl.arange(0, BLOCK_SIZE)
+ mask = col_offsets < n_cols
+
+ dY += row_idx * dY_row_stride
+ X += row_idx * X_row_stride
+ r += row_idx * r_row_stride
+
+ dY_row = tl.load(dY + col_offsets, mask = mask, other = 0).to(tl.float32)
+ X_row = tl.load(X + col_offsets, mask = mask, other = 0).to(tl.float32)
+ W_row = tl.load(W + col_offsets, mask = mask, other = 0).to(tl.float32)
+
+ # Get saved row variance
+ inv_var = tl.load(r).to(tl.float32)
+ normed = X_row * inv_var
+
+ if GEMMA: dY_W = dY_row * (W_row + 1.0)
+ else: dY_W = dY_row * W_row
+
+ rowsum_dY_normed = tl.sum(dY_W * normed, axis = 0)
+ output = inv_var/n_cols * (n_cols*dY_W - normed*rowsum_dY_normed)
+ tl.store(dY + col_offsets, output, mask = mask)
+pass
+
+
+@triton.jit
+def _gemma_rms_layernorm_forward(
+ Y, Y_row_stride,
+ X, X_row_stride,
+ W, W_row_stride,
+ r, r_row_stride,
+ n_cols, eps,
+ BLOCK_SIZE : tl.constexpr,
+):
+ # Copies https://github.com/google-deepmind/gemma/blob/main/gemma/layers.py#L31
+ # and https://github.com/keras-team/keras-nlp/blob/v0.8.2/keras_nlp/models/gemma/rms_normalization.py#L33
+ # exactly. Essentially all in float32!
+ row_idx = tl.program_id(0)
+ col_offsets = tl.arange(0, BLOCK_SIZE)
+ mask = col_offsets < n_cols
+
+ Y += row_idx * Y_row_stride
+ X += row_idx * X_row_stride
+ r += row_idx * r_row_stride
+
+ X_row = tl.load(X + col_offsets, mask = mask, other = 0).to(tl.float32)
+ W_row = tl.load(W + col_offsets, mask = mask, other = 0).to(tl.float32)
+
+ row_var = tl.sum(X_row * X_row, axis = 0) / n_cols
+ inv_var = 1.0 / tl.sqrt(row_var + eps) # Must be 1/sqrt to match Deepmind's impl
+ tl.store(r, inv_var)
+ normed = X_row * inv_var
+ output = normed * (W_row + 1.0)
+
+ tl.store(Y + col_offsets, output, mask = mask)
+pass
+
+
+class Fast_RMS_Layernorm(torch.autograd.Function):
+ @staticmethod
+ def forward(ctx, X, W, eps, gemma = False):
+ shape = X.shape
+ dim = shape[-1]
+ X = X.view(-1, dim)
+ n_rows, n_cols = X.shape
+ BLOCK_SIZE, num_warps = calculate_settings(n_cols)
+
+ Y = torch.empty((n_rows, n_cols), dtype = X.dtype, device = "cuda")
+ r = torch.empty(n_rows, dtype = torch.float32, device = "cuda")
+
+ fx = _gemma_rms_layernorm_forward if gemma else _rms_layernorm_forward
+ fx[(n_rows,)](
+ Y, Y.stride(0),
+ X, X.stride(0),
+ W, W.stride(0),
+ r, r.stride(0),
+ n_cols, eps,
+ BLOCK_SIZE = BLOCK_SIZE,
+ num_warps = num_warps,
+ )
+ ctx.eps = eps
+ ctx.BLOCK_SIZE = BLOCK_SIZE
+ ctx.num_warps = num_warps
+ ctx.GEMMA = gemma
+ ctx.save_for_backward(X, W, r)
+ return Y.view(*shape)
+ pass
+
+ @staticmethod
+ def backward(ctx, dY):
+ shape = dY.shape
+ dim = shape[-1]
+ dY = dY.view(-1, dim)
+ X, W, r = ctx.saved_tensors
+ n_rows, n_cols = dY.shape
+ dW = X
+
+ _rms_layernorm_backward[(n_rows,)](
+ dY, dY.stride(0),
+ X, X .stride(0),
+ W, W .stride(0),
+ r, r .stride(0),
+ dW, dW.stride(0),
+ n_cols, ctx.eps,
+ GEMMA = ctx.GEMMA,
+ BLOCK_SIZE = ctx.BLOCK_SIZE,
+ num_warps = ctx.num_warps,
+ )
+ dX = dY.view(*shape)
+ return dX, None, None, None
+ pass
+pass
+
+
+def fast_rms_layernorm(layernorm, X, gemma = False):
+ W = layernorm.weight
+ eps = layernorm.variance_epsilon
+ out = Fast_RMS_Layernorm.apply(X, W, eps, gemma)
+ return out
+pass
diff --git a/unsloth/unsloth/kernels/rope_embedding.py b/unsloth/unsloth/kernels/rope_embedding.py
new file mode 100644
index 0000000000000000000000000000000000000000..b32d75ebf328db88ad460937ce2107fcd4ddef67
--- /dev/null
+++ b/unsloth/unsloth/kernels/rope_embedding.py
@@ -0,0 +1,181 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import triton
+import triton.language as tl
+import torch
+from .utils import calculate_settings
+
+ROPE_GROUP_SIZE = 4
+
+@triton.heuristics({"BACKWARD_PASS": lambda args: args["BACKWARD_PASS"],})
+@triton.jit
+def _rope_embedding(
+ Q, Q_row_stride,
+ cos, cos_row_stride,
+ sin, sin_row_stride,
+ seqlen,
+ head_dim : tl.constexpr,
+ n_heads : tl.constexpr,
+ BACKWARD_PASS : tl.constexpr,
+ BLOCK_SIZE : tl.constexpr,
+):
+ """
+ Calculates the RoPE Embedding quickly
+ RoPE is Q * cos + rotate_half(Q) * sin
+ See our blog post for more info
+ """
+ row_position = tl.program_id(0)
+ group_head_position = tl.program_id(1)
+ col_offsets = tl.arange(0, BLOCK_SIZE)
+ half_head_dim = head_dim // 2
+ mask = col_offsets < half_head_dim
+
+ sin1 = tl.load(sin + (row_position % seqlen)*sin_row_stride + \
+ half_head_dim*0 + col_offsets, mask = mask, other = 0)
+ cos1 = tl.load(cos + (row_position % seqlen)*cos_row_stride + \
+ half_head_dim*0 + col_offsets, mask = mask, other = 0)
+
+ if BACKWARD_PASS:
+ # See our blog post for more info.
+ sin1 = -sin1
+ pass
+
+ # [TODO] Autotune ROPE_GROUP_SIZE to be 1, 2, 4, 8
+ head_start = group_head_position * ROPE_GROUP_SIZE
+ head_end = min((head_start + ROPE_GROUP_SIZE), n_heads)
+
+ # 10% Faster kernel from [HuyNguyen-hust](https://github.com/unslothai/unsloth/pull/238)
+ for k in range(head_start, head_end):
+ offs_q1 = row_position * Q_row_stride + k * head_dim + col_offsets
+ offs_q2 = row_position * Q_row_stride + k * head_dim + col_offsets + half_head_dim
+
+ # For Gemma - sometimes RoPE must be done in float32 and not bfloat16
+ Q1 = tl.load(Q + offs_q1, mask = mask, other = 0).to(sin1.dtype)
+ Q2 = tl.load(Q + offs_q2, mask = mask, other = 0).to(sin1.dtype)
+
+ tl.store(Q + offs_q1, Q1*cos1 - Q2*sin1, mask = mask)
+ tl.store(Q + offs_q2, Q2*cos1 + Q1*sin1, mask = mask)
+ pass
+pass
+
+
+class Fast_RoPE_Embedding(torch.autograd.Function):
+ @staticmethod
+ def forward(ctx, Q, cos, sin):
+ cos, sin = cos.squeeze(), sin.squeeze()
+ batch, seq_len, n_heads, head_dim = Q.shape
+ Q = Q.view(batch*seq_len, n_heads*head_dim)
+ n_rows, n_cols = Q.shape
+ assert(seq_len <= cos.shape[0])
+
+ # [TODO] Changing blocksize to head_dim//2 seems to have
+ # some concurrency / un-deterministic issues.
+ BLOCK_SIZE, num_warps = calculate_settings(head_dim//2) # (head_dim//2)
+
+ # group_size = 4 # 4 or 8, too large group_size can hurt performance.
+ div, mod = divmod(n_heads, ROPE_GROUP_SIZE)
+ n_groups = div + (mod != 0)
+
+ _rope_embedding[(n_rows, n_groups, )](
+ Q, Q.stride(0),
+ cos, cos.stride(0),
+ sin, sin.stride(0),
+ seq_len,
+ head_dim, n_heads,
+ BACKWARD_PASS = False,
+ BLOCK_SIZE = BLOCK_SIZE,
+ num_warps = num_warps,
+ )
+ ctx.BLOCK_SIZE = BLOCK_SIZE
+ ctx.num_warps = num_warps
+ ctx.n_groups = n_groups
+ ctx.cos = cos
+ ctx.sin = sin
+ return Q.view(batch, seq_len, n_heads, head_dim)
+ pass
+
+ @staticmethod
+ def backward(ctx, dY):
+ batch, seq_len, n_heads, head_dim = dY.shape
+ dY = dY.reshape(batch*seq_len, n_heads*head_dim)
+ # Must be reshape not view
+ n_rows, n_cols = dY.shape
+
+ cos = ctx.cos
+ sin = ctx.sin
+
+ _rope_embedding[(n_rows, ctx.n_groups, )](
+ dY, dY .stride(0),
+ cos, cos.stride(0),
+ sin, sin.stride(0),
+ seq_len, head_dim, n_heads,
+ BACKWARD_PASS = True,
+ BLOCK_SIZE = ctx.BLOCK_SIZE,
+ num_warps = ctx.num_warps,
+ )
+ dY = dY.view(batch, seq_len, n_heads, head_dim)
+ return dY, None, None,
+ pass
+pass
+
+
+def fast_rope_embedding(Q, K, cos, sin):
+ Q = Fast_RoPE_Embedding.apply(Q.transpose(1, 2), cos, sin).transpose(1, 2)
+ K = Fast_RoPE_Embedding.apply(K.transpose(1, 2), cos, sin).transpose(1, 2)
+ return Q, K
+pass
+
+
+class Slow_RoPE_Embedding(torch.autograd.Function):
+ @staticmethod
+ def forward(ctx, Q, cos, sin, position_ids):
+ if position_ids is not None:
+ # The first two dimensions of cos and sin are always 1, so we can `squeeze` them.
+ cos = cos.squeeze(1).squeeze(0) # [seq_len, dim]
+ sin = sin.squeeze(1).squeeze(0) # [seq_len, dim]
+ cos = cos[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim]
+ sin = sin[position_ids].unsqueeze(1) # [bs, 1, seq_len, dim]
+
+ # Q * cos + rotate_half(Q) * sin
+ half = Q.shape[-1]//2
+ RH_Q = torch.cat((-Q[..., half:], Q[..., :half]), dim = -1)
+ Q *= cos
+ Q.addcmul_(RH_Q, sin)
+ # RH_Q *= sin
+ # Q += RH_Q
+ ctx.save_for_backward(cos, sin)
+ return Q
+ pass
+
+ @staticmethod
+ def backward(ctx, dY):
+ cos, sin = ctx.saved_tensors
+ # Q * cos + rotate_half.T(Q) * sin
+ half = dY.shape[-1]//2
+ RH_dY = torch.cat((dY[..., half:], -dY[..., :half]), dim = -1)
+ dY *= cos
+ dY.addcmul_(RH_dY, sin)
+ # RH_dY *= sin
+ # dY += RH_dY
+ return dY, None, None, None
+ pass
+pass
+
+
+def inplace_rope_embedding(Q, K, cos, sin, position_ids):
+ Q = Slow_RoPE_Embedding.apply(Q, cos, sin, position_ids)
+ K = Slow_RoPE_Embedding.apply(K, cos, sin, position_ids)
+ return Q, K
+pass
diff --git a/unsloth/unsloth/kernels/swiglu.py b/unsloth/unsloth/kernels/swiglu.py
new file mode 100644
index 0000000000000000000000000000000000000000..ff6b162680be5e781adcaf55f0e8ccf70739a27f
--- /dev/null
+++ b/unsloth/unsloth/kernels/swiglu.py
@@ -0,0 +1,99 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import triton
+import triton.language as tl
+import torch
+from .utils import calculate_settings
+
+
+@triton.jit
+def _fg_kernel(e, g, h, n_elements, BLOCK_SIZE : tl.constexpr,):
+ block_idx = tl.program_id(0)
+ offsets = block_idx*BLOCK_SIZE + tl.arange(0, BLOCK_SIZE)
+ mask = offsets < n_elements
+
+ e_row = tl.load(e + offsets, mask = mask, other = 0).to(tl.float32)
+ g_row = tl.load(g + offsets, mask = mask, other = 0)#.to(tl.float32)
+
+ # f = e * sigmoid(e)
+ f_row = e_row * tl.sigmoid(e_row) # e_row / (1 + tl.exp(-e_row))
+ f_row = f_row.to(g_row.dtype) # Exact copy from HF
+ # h = f * g
+ h_row = f_row * g_row
+
+ # Store h
+ tl.store(h + offsets, h_row, mask = mask)
+pass
+
+
+def swiglu_fg_kernel(e, g):
+ batch, seq_len, hd = e.shape
+ n_elements = e.numel()
+ h = torch.empty((batch, seq_len, hd), dtype = e.dtype, device = "cuda")
+ grid = lambda meta: (triton.cdiv(n_elements, meta['BLOCK_SIZE']),)
+ _fg_kernel[grid](e, g, h, n_elements, BLOCK_SIZE = 1024,)
+ return h
+pass
+
+
+@triton.jit
+def _DWf_DW_dfg_kernel(DW, e, g, n_elements, BLOCK_SIZE : tl.constexpr,):
+ """
+ e = e.float()
+ se = 1.0 / (1.0 + torch.exp(-e))
+ f = (se * e).to(dtype)
+ h = f * g
+ df = DW * f
+ dg = DW * g
+ de = (dg.float() * se * (1.0 + e * (1.0 - se))).to(dtype)
+ """
+ block_idx = tl.program_id(0)
+ offsets = block_idx*BLOCK_SIZE + tl.arange(0, BLOCK_SIZE)
+ mask = offsets < n_elements
+
+ DW_row = tl.load(DW + offsets, mask = mask, other = 0)#.to(tl.float32)
+ e_row = tl.load(e + offsets, mask = mask, other = 0).to(tl.float32)
+ g_row = tl.load(g + offsets, mask = mask, other = 0)#.to(tl.float32)
+
+ # e = e.float()
+ # se = 1.0 / (1.0 + torch.exp(-e))
+ se_row = tl.sigmoid(e_row) # 1.0 / (1.0 + tl.exp(-e_row))
+ # f = (se * e).to(dtype)
+ f_row = se_row * e_row
+ f_row = f_row.to(DW_row.dtype)
+ # h = f * g
+ h_row = f_row * g_row
+ # df = DW * f
+ df_row = DW_row * f_row
+ # dg = DW * g
+ dg_row = DW_row * g_row
+ # de = (dg.float() * se * (1.0 + e * (1.0 - se))).to(dtype)
+ de_row = dg_row.to(tl.float32) * se_row * (1.0 + e_row * (1.0 - se_row))
+ de_row = de_row.to(DW_row.dtype)
+
+ # Store derivatives in buffers
+ tl.store(DW + offsets, h_row, mask = mask) # h = f * g
+ tl.store(e + offsets, df_row, mask = mask) # df = DW * f
+ tl.store(g + offsets, de_row, mask = mask) # de
+pass
+
+
+def swiglu_DWf_DW_dfg_kernel(DW, e, g):
+ batch_seq_len, hd = e.shape
+ n_elements = e.numel()
+ grid = lambda meta: (triton.cdiv(n_elements, meta['BLOCK_SIZE']),)
+ _DWf_DW_dfg_kernel[grid](DW, e, g, n_elements, BLOCK_SIZE = 1024,)
+ return DW, e, g
+pass
diff --git a/unsloth/unsloth/kernels/utils.py b/unsloth/unsloth/kernels/utils.py
new file mode 100644
index 0000000000000000000000000000000000000000..1f2085df5ae4acc3e81a1a93dcd29c3cbb7b1940
--- /dev/null
+++ b/unsloth/unsloth/kernels/utils.py
@@ -0,0 +1,245 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import triton
+MAX_FUSED_SIZE = 65536
+next_power_of_2 = triton.next_power_of_2
+
+def calculate_settings(n):
+ BLOCK_SIZE = next_power_of_2(n)
+ if BLOCK_SIZE > MAX_FUSED_SIZE:
+ raise RuntimeError(f"Cannot launch Triton kernel since n = {n} exceeds "\
+ f"the maximum CUDA blocksize = {MAX_FUSED_SIZE}.")
+ num_warps = 4
+ if BLOCK_SIZE >= 32768: num_warps = 32
+ elif BLOCK_SIZE >= 8192: num_warps = 16
+ elif BLOCK_SIZE >= 2048: num_warps = 8
+ return BLOCK_SIZE, num_warps
+pass
+
+
+import bitsandbytes as bnb
+get_ptr = bnb.functional.get_ptr
+import ctypes
+import torch
+cdequantize_blockwise_fp32 = bnb.functional.lib.cdequantize_blockwise_fp32
+cdequantize_blockwise_fp16_nf4 = bnb.functional.lib.cdequantize_blockwise_fp16_nf4
+cdequantize_blockwise_bf16_nf4 = bnb.functional.lib.cdequantize_blockwise_bf16_nf4
+cgemm_4bit_inference_naive_fp16 = bnb.functional.lib.cgemm_4bit_inference_naive_fp16
+cgemm_4bit_inference_naive_bf16 = bnb.functional.lib.cgemm_4bit_inference_naive_bf16
+
+
+def QUANT_STATE(W):
+ return getattr(W, "quant_state", None)
+pass
+
+
+def get_lora_parameters(proj):
+ # For DPO or disabled adapters
+ base_layer = (proj.base_layer if hasattr(proj, "base_layer") else proj)
+ W = base_layer.weight
+
+ if not hasattr(proj, "disable_adapters") or proj.disable_adapters or proj.merged:
+ return W, QUANT_STATE(W), None, None, None
+ pass
+
+ active_adapter = proj.active_adapters[0] if \
+ hasattr(proj, "active_adapters") else proj.active_adapter
+ A = proj.lora_A [active_adapter].weight
+ B = proj.lora_B [active_adapter].weight
+ s = proj.scaling[active_adapter]
+ return W, QUANT_STATE(W), A, B, s
+pass
+
+
+def fast_dequantize(W, quant_state = None, out = None):
+ if quant_state is None: return W
+ if type(quant_state) is not list:
+ # New quant_state as a class
+ # https://github.com/TimDettmers/bitsandbytes/pull/763/files
+ absmax = quant_state.absmax
+ shape = quant_state.shape
+ dtype = quant_state.dtype
+ blocksize = quant_state.blocksize
+ offset = quant_state.offset
+ state2 = quant_state.state2
+ absmax2 = state2.absmax
+ code2 = state2.code
+ blocksize2 = state2.blocksize
+ else:
+ # Old quant_state as a list of lists
+ absmax, shape, dtype, blocksize, compressed_stats, _, _ = quant_state
+ offset, state2 = compressed_stats
+ absmax2, code2, blocksize2, _, _, _, _ = state2
+ pass
+
+ # Create weight matrix
+ if out is None:
+ out = torch.empty(shape, dtype = dtype, device = "cuda")
+ else:
+ assert(out.shape == shape)
+ assert(out.dtype == dtype)
+
+ # NF4 dequantization of statistics
+ n_elements_absmax = absmax.numel()
+ out_absmax = torch.empty(n_elements_absmax, dtype = torch.float32, device = "cuda")
+
+ # Do dequantization
+ ptr_out_absmax = get_ptr(out_absmax)
+ cdequantize_blockwise_fp32(
+ get_ptr(code2), get_ptr(absmax), get_ptr(absmax2), ptr_out_absmax,
+ ctypes.c_int(blocksize2), ctypes.c_int(n_elements_absmax)
+ )
+ out_absmax += offset
+
+ fx = cdequantize_blockwise_fp16_nf4 if dtype == torch.float16 else \
+ cdequantize_blockwise_bf16_nf4
+ fx(get_ptr(None), get_ptr(W), ptr_out_absmax, get_ptr(out),
+ ctypes.c_int(blocksize), ctypes.c_int(out.numel()))
+
+ # Careful returning transposed data
+ is_transposed = (True if W.shape[0] == 1 else False)
+ return out.t() if is_transposed else out
+pass
+
+
+def fast_gemv(X, W, quant_state, out = None):
+ if quant_state is None: return torch.matmul(X, W, out = out)
+ # For fast X @ W where seq_len == 1
+ # From https://github.com/TimDettmers/bitsandbytes/blob/main/bitsandbytes/functional.py#L1469
+ _, q_len, hd = X.shape
+ # assert(q_len == 1)
+
+ if type(quant_state) is not list:
+ # https://github.com/TimDettmers/bitsandbytes/pull/763/files
+ absmax = quant_state.absmax
+ shape = quant_state.shape
+ dtype = quant_state.dtype
+ blocksize = quant_state.blocksize
+ stats = quant_state.code
+ offset = quant_state.offset
+ state2 = quant_state.state2
+ absmax2 = state2.absmax
+ code2 = state2.code
+ blocksize2 = state2.blocksize
+ else:
+ absmax, shape, dtype, blocksize, compressed_stats, quant_type, stats = quant_state
+ offset, state2 = compressed_stats
+ absmax2, code2, blocksize2, _, _, _, _ = state2
+ pass
+ # assert(dtype == X.dtype)
+ bout = shape[0]
+
+ if out is None:
+ out = torch.empty((1, 1, bout,), dtype = dtype, device = "cuda")
+ # else:
+ # assert(out.shape == (1, 1, bout,))
+ # pass
+
+ n = 1
+ m = shape[0]
+ k = shape[1]
+ lda = shape[0]
+ ldc = shape[0]
+ ldb = (hd+1)//2
+ m = ctypes.c_int32(m)
+ n = ctypes.c_int32(n)
+ k = ctypes.c_int32(k)
+ lda = ctypes.c_int32(lda)
+ ldb = ctypes.c_int32(ldb)
+ ldc = ctypes.c_int32(ldc)
+
+ df = torch.empty(absmax.shape, dtype = torch.float32, device = "cuda")
+ cdequantize_blockwise_fp32(
+ get_ptr(code2), get_ptr(absmax), get_ptr(absmax2), get_ptr(df),
+ ctypes.c_int(blocksize2), ctypes.c_int(df.numel()),
+ )
+ df += offset
+ absmax = df
+
+ fx = cgemm_4bit_inference_naive_fp16 if dtype == torch.float16 else \
+ cgemm_4bit_inference_naive_bf16
+
+ blocksize = ctypes.c_int32(blocksize)
+ fx(m, n, k, get_ptr(X), get_ptr(W), get_ptr(absmax), get_ptr(stats), get_ptr(out),
+ lda, ldb, ldc, blocksize)
+
+ return out
+pass
+
+
+def fast_linear_forward(proj, X, temp_lora = None, out = None):
+
+ W, W_quant, lora_A, lora_B, lora_S = get_lora_parameters(proj)
+ bsz, q_len, in_dim = X.shape
+ if q_len != 1: return matmul_lora(X, W, W_quant, lora_A, lora_B, lora_S)
+
+ if W_quant is None:
+ out = torch.matmul(X, W.t(), out = out)
+ elif bsz == 1 and q_len == 1:
+ out = fast_gemv(X, W, W_quant, out = out)
+ else:
+ W = fast_dequantize(W.t(), W_quant)
+ out = torch.matmul(X, W, out = out)
+ pass
+
+ # Add in LoRA weights
+ if lora_A is not None:
+ out_dim = out.shape[2]
+ dtype = X.dtype
+
+ if not hasattr(lora_A, "_fast_lora"):
+ lora_A._fast_lora = lora_A.to(dtype)
+ lora_B._fast_lora = lora_B.to(dtype)
+ pass
+
+ if bsz == 1:
+ out = out.view(out_dim)
+ temp_lora = torch.mv(lora_A._fast_lora, X.ravel(), out = temp_lora)
+ out.addmv_(lora_B._fast_lora, temp_lora, alpha = lora_S)
+ else:
+ out = out.view(bsz, out_dim)
+ temp_lora = torch.mm(X.view(bsz, in_dim), lora_A._fast_lora.t(), out = temp_lora)
+ out.addmm_(temp_lora, lora_B._fast_lora.t(), alpha = lora_S)
+ pass
+ out = out.view(bsz, 1, out_dim)
+ pass
+
+ return out
+pass
+
+
+def matmul_lora(X, W, W_quant, A, B, s, out = None):
+ dtype = X.dtype
+ W = fast_dequantize(W.t(), W_quant)
+
+ if X.dim() == 3:
+ batch, seq_len, d = X.shape
+ X = X.view(-1, X.shape[-1])
+ reshape = True
+ else:
+ reshape = False
+ pass
+
+ out = torch.matmul(X, W, out = out)
+ if W_quant is not None: del W
+
+ if A is not None:
+ # LoRA is enabled
+ A, B = A.t(), B.t()
+ out += (X @ A.to(dtype)) @ (s * B.to(dtype))
+ pass
+
+ return out.view(batch, seq_len, -1) if reshape else out
+pass
diff --git a/unsloth/unsloth/models/__init__.py b/unsloth/unsloth/models/__init__.py
new file mode 100644
index 0000000000000000000000000000000000000000..ff7129e06a95140ef082e627a262bcf43f4ec2ee
--- /dev/null
+++ b/unsloth/unsloth/models/__init__.py
@@ -0,0 +1,19 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+from .loader import FastLanguageModel
+from .llama import FastLlamaModel
+from .mistral import FastMistralModel
+from .qwen2 import FastQwen2Model
+from .dpo import PatchDPOTrainer
diff --git a/unsloth/unsloth/models/_utils.py b/unsloth/unsloth/models/_utils.py
new file mode 100644
index 0000000000000000000000000000000000000000..09e035e684fd457b162ed2c6c999fd9f182610f5
--- /dev/null
+++ b/unsloth/unsloth/models/_utils.py
@@ -0,0 +1,390 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import torch
+from typing import Union, Optional, List, Any, Callable
+import warnings
+warnings.filterwarnings(action = "ignore", category = UserWarning, module = "torch")
+warnings.filterwarnings(action = "ignore", category = UserWarning, module = "huggingface_hub")
+warnings.filterwarnings(action = "ignore", category = RuntimeWarning, module = "subprocess")
+warnings.filterwarnings(action = "ignore", category = UserWarning, module = "transformers")
+warnings.filterwarnings(action = "ignore", category = FutureWarning, module = "accelerate")
+import bitsandbytes as bnb
+from transformers.models.llama.modeling_llama import logger
+from transformers import AutoTokenizer
+from platform import system as platform_system
+platform_system = platform_system()
+import math
+import numpy as np
+import os
+import psutil
+
+__version__ = "2024.5"
+
+# Get Flash Attention v2 if Ampere (RTX 30xx, A100)
+major_version, minor_version = torch.cuda.get_device_capability()
+if major_version >= 8:
+ try:
+ from flash_attn import flash_attn_func
+ # Check for CUDA linking errors "undefined symbol: _ZNK3c106SymIntltEl"
+ try:
+ from flash_attn.flash_attn_interface import flash_attn_cuda
+ HAS_FLASH_ATTENTION = True
+ except:
+ logger.warning_once(
+ "Unsloth: Your Flash Attention 2 installation seems to be broken?\n"\
+ "A possible explanation is you have a new CUDA version which isn't\n"\
+ "yet compatible with FA2? Please file a ticket to Unsloth or FA2.\n"\
+ "We shall now use Xformers instead, which gets a 0.01% performance hit.\n"\
+ "We found this negligible impact by benchmarking on 1x A100."
+ )
+ HAS_FLASH_ATTENTION = False
+ except:
+ HAS_FLASH_ATTENTION = False
+else:
+ # Tri Dao's benchmark shows xformers is faster for now.
+ HAS_FLASH_ATTENTION = False
+pass
+import xformers.ops.fmha as xformers
+xformers_attention = xformers.memory_efficient_attention
+from xformers import __version__ as xformers_version
+
+__all__ = [
+ "prepare_model_for_kbit_training",
+ "xformers",
+ "xformers_attention",
+ "xformers_version",
+ "__version__",
+ "HAS_FLASH_ATTENTION",
+ "platform_system",
+ "patch_tokenizer",
+ "get_statistics",
+ "Unsloth_Offloaded_Gradient_Checkpointer",
+]
+
+
+def prepare_model_for_kbit_training(
+ model : Any,
+ use_gradient_checkpointing : Optional = True,
+ use_reentrant : Optional[bool] = True,
+) -> Any:
+ """
+ Calculates where to place the gradient checkpoints given n_layers.
+ We also freeze all other layers's gradients
+
+ Args:
+ model: Any LlamaModel with layers.
+ use_gradient_checkpointing (`bool`, *optional*):
+ Default enabled. Provides memory savings by not saving all activations,
+ but only some.
+ use_reentrant (`bool`, *optional*):
+ https://github.com/pytorch/pytorch/blob/main/torch/utils/checkpoint.py#L354
+ Optimal gradient checkpointing algorithm which will be the default in
+ future Pytorch versions.
+ """
+
+ # Freeze all parameters except LoRA
+ import re
+ with torch.no_grad():
+ for name, param in model.named_parameters():
+ if ".lora_A." in name or ".lora_B." in name or ".lora_magnitude_vector" in name:
+ param.requires_grad_(True)
+ # Also must be in float32!
+ if param.dtype != torch.float32:
+ name = name.replace("base_model", "model", 1)
+ layer_number = re.search(r"\.[\d]{1,}\.", name).group(0)
+ name = name.replace(layer_number, f"[{layer_number[1:-1]}].")
+ name = name.replace(".weight", "", 1)
+ exec(f"{name}.to(torch.float32)")
+ pass
+ else:
+ param.requires_grad_(False)
+ pass
+ pass
+
+ # Gradient checkpointing!
+ if use_gradient_checkpointing == "unsloth":
+
+ # Saves VRAM!
+ original_model = model
+ while hasattr(original_model, "model"):
+ original_model._offloaded_gradient_checkpointing = True
+ original_model = original_model.model
+ pass
+ original_model._offloaded_gradient_checkpointing = True
+
+ model.gradient_checkpointing_enable()
+
+ elif use_gradient_checkpointing == True:
+ model.gradient_checkpointing_enable()
+ pass
+
+ # If use_reentrant = True which is the Pytorch default, we just make the input requires_grad.
+ if use_reentrant:
+ if hasattr(model, "enable_input_require_grads"):
+ model.enable_input_require_grads()
+ else:
+ def make_inputs_require_grad(module, input, output):
+ output.requires_grad_(True)
+ model.get_input_embeddings().register_forward_hook(make_inputs_require_grad)
+
+ return model
+pass
+
+
+def patch_tokenizer(model, tokenizer):
+ """
+ Phi3's pad_token isn't set. We set it to <|placeholder...
+ Llama-3 is <|reserved...
+ Llama-2 is
+ Check if pad_token is not the same as eos_token otherwise the loss will ignore it!!
+ Fixes https://github.com/unslothai/unsloth/issues/5
+ """
+ possible_reserved_tokens = ("<|reserved", "<|placeholder",)
+
+ if model is not None:
+ model.config.update({"unsloth_version" : __version__})
+
+ bad_pad_token = False
+ if hasattr(tokenizer, "pad_token") and tokenizer.pad_token is not None:
+ # Check if pad_token is not the same as eos_token otherwise the loss will ignore it!!
+ bad_pad_token = tokenizer.eos_token == tokenizer.pad_token
+ elif hasattr(tokenizer, "pad_token") and tokenizer.pad_token is None:
+ bad_pad_token = True
+ else:
+ bad_pad_token = False
+ pass
+
+ if bad_pad_token:
+ # Find a better pad token
+ added_tokens = [str(x) for x in tokenizer.added_tokens_decoder.values()]
+ possible_pad_token = None
+ for added_token in added_tokens[::-1]:
+ if added_token.startswith(possible_reserved_tokens):
+ possible_pad_token = added_token
+ break
+ pass
+ pass
+ if possible_pad_token is None:
+ # Try unk_token
+ possible_pad_token = tokenizer.unk_token
+ pass
+
+ if possible_pad_token is None:
+ # Failure to find a good replacement!! We shall manually add one!
+ new_pad_token = "<|PAD_TOKEN|>"
+ while new_pad_token in tokenizer.get_vocab():
+ new_pad_token += "#"
+ pass
+ possible_pad_token = new_pad_token
+ pass
+
+ name = model.config._name_or_path if model is not None else "Model"
+ logger.warning_once(
+ f"{name} does not have a padding token! Will use pad_token = {possible_pad_token}."
+ )
+
+ # Edit pad_token
+ tokenizer.add_special_tokens({"pad_token" : possible_pad_token})
+ tokenizer.pad_token = possible_pad_token
+ if model is not None:
+ config = model.config.update({"pad_token_id" : tokenizer.pad_token_id})
+ pass
+ return model, tokenizer
+pass
+
+
+# Weirdly LoraLayer.update_layer downcasts PEFT layers to float16??
+# For mixed precision, we need it to be in float32 not float16.
+from peft.tuners.lora.layer import LoraLayer
+import inspect, re
+try:
+ source = inspect.getsource(LoraLayer.update_layer)
+ text = "if weight is not None:\n"
+ start = source.find(text) + len(text)
+ end = source.find("self.to(weight.device)", start)
+ spaces = re.findall(r"^([ ]{1,})break", source, flags = re.MULTILINE)[0]
+ source = source.replace(source[start : end], spaces)
+ spaces = len(re.match(r"[\s]{1,}", source).group(0))
+ lines = source.split("\n")
+ source = "\n".join(x[spaces:] for x in lines)
+ source = re.sub("([^\.])nn\.", r"\1torch.nn.", source)
+ source = source.replace("def update_layer", "def LoraLayer_update_layer")
+ exec(source, globals())
+
+ # Fix up incorrect downcasting of LoRA weights
+ from peft.tuners.lora.layer import LoraLayer
+ LoraLayer.update_layer = LoraLayer_update_layer
+ from peft.tuners.lora import LoraLayer
+ LoraLayer.update_layer = LoraLayer_update_layer
+except:
+ logger.warning_once(
+ "Unsloth unsuccessfully patched LoraLayer.update_layer. Please file a bug report.\n"\
+ "Luckily, your training run will still work in the meantime!"
+ )
+pass
+
+
+def get_statistics():
+ # We log some basic stats about which environment is being used.
+ # We simply download a README.md file from HF - all data is made public.
+ # This is simply so we can check if some envs are broken or not.
+ try:
+ from huggingface_hub import hf_hub_download
+ from huggingface_hub.utils import disable_progress_bars, enable_progress_bars, are_progress_bars_disabled
+ import psutil
+ n_cpus = psutil.cpu_count(logical = False)
+
+ keynames = "\n" + "\n".join(os.environ.keys())
+ statistics = None
+ if "\nCOLAB_" in keynames and n_cpus == 1: statistics = "colab"
+ elif "\nCOLAB_" in keynames: statistics = "colabpro"
+ elif "\nKAGGLE_" in keynames: statistics = "kaggle"
+ elif "\nRUNPOD_" in keynames: statistics = "runpod"
+ elif "\nAWS_" in keynames: statistics = "aws"
+ elif "\nAZURE_" in keynames: statistics = "azure"
+ elif "\nK_" in keynames or "\nFUNCTION_" in keynames: statistics = "gcp"
+ elif "\nINVOCATION_ID" in keynames: statistics = "lambda"
+
+ if statistics is not None:
+ disabled = False
+ if not are_progress_bars_disabled():
+ disable_progress_bars()
+ disabled = True
+ pass
+ hf_hub_download(f"unslothai/statistics-{statistics}", "README.md", force_download = True)
+ if disabled:
+ enable_progress_bars()
+ pass
+ pass
+ except:
+ pass
+pass
+
+
+def _calculate_n_gradient_checkpoints(
+ n_layers : int,
+ method : Optional[Union[str, int]] = "sqrt",
+) -> List[int]:
+ assert(type(n_layers) is int and n_layers > 0)
+
+ if method is None: method = "sqrt"
+
+ if method == "sqrt":
+ n_checkpoints = int(n_layers**0.5)
+ elif type(method) is int and method > 0:
+ n_checkpoints = int(np.ceil(n_layers / method))
+ else:
+ raise ValueError("method must be 'sqrt' or an int >0 and <= n_layers.")
+
+ size = n_layers // n_checkpoints
+ sizes = np.full(n_checkpoints, size, dtype = int)
+ leftovers = n_layers % n_checkpoints
+ # We append leftovers from the right
+ for k in range(leftovers):
+ sizes[n_checkpoints-1-k] += 1
+ boundaries = np.hstack((0, np.cumsum(sizes)))
+ boundaries = boundaries.tolist()
+ return boundaries
+pass
+
+
+def calculate_n_gradient_checkpoints(
+ n_layers : int,
+ layers_per_checkpoint : Optional[Union[str, int]] = "sqrt",
+) -> List[int]:
+ assert(type(n_layers) is int and n_layers > 0)
+
+ if layers_per_checkpoint is None or layers_per_checkpoint == 1:
+ return None
+
+ boundaries = _calculate_n_gradient_checkpoints(n_layers, layers_per_checkpoint)
+
+ assert(boundaries[0] == 0 and boundaries[-1] == n_layers)
+ assert(min(boundaries) == 0 and max(boundaries) == n_layers)
+ assert(np.diff(boundaries).min() >= 0)
+ return boundaries
+pass
+
+
+def prepare_n_gradient_checkpoints(
+ model : Any,
+ layers_per_checkpoint : Optional[Union[str, int]] = "sqrt",
+ use_reentrant : Optional[bool] = True,
+) -> None:
+ """
+ Calculates where to place the gradient checkpoints given n_layers.
+
+ Args:
+ model: Any LlamaModel with layers.
+ layers_per_checkpoint (`Union[str, int]`, *optional*):
+ Can either be `sqrt` or an integer for how many layers per checkpoint you want.
+ The more, the less memory usage, but can be slower. Default is `sqrt`.
+ Choose 1 for Pytorch gradient checkpointing. 2 to wrap 2 layers in 1 module etc.
+ use_reentrant (`bool`, *optional*):
+ https://github.com/pytorch/pytorch/blob/main/torch/utils/checkpoint.py#L354
+ Optimal gradient checkpointing algorithm `use_reentrant=False` which will
+ be the default in future Pytorch versions doesn't seem to work??
+ """
+ _model = None
+ if hasattr(model, "layers"):
+ _model = model
+ elif hasattr(model, "model"):
+ if hasattr(model.model, "layers"):
+ _model = model.model
+ if _model is None:
+ raise TypeError("`model` or `model.model` does not have attribute `layers`. Are you sure this is a model?")
+ pass
+
+ if use_reentrant is False:
+ use_reentrant = True
+ pass
+
+ n_layers = len(_model.layers)
+ boundaries = calculate_n_gradient_checkpoints(n_layers, layers_per_checkpoint)
+ _model._gradient_checkpointing_boundaries = boundaries
+ _model._gradient_checkpointing_use_reentrant = use_reentrant
+pass
+
+
+class Unsloth_Offloaded_Gradient_Checkpointer(torch.autograd.Function):
+ """
+ Saves VRAM by smartly offloading to RAM.
+ Tiny hit to performance, since we mask the movement via non blocking calls.
+ """
+ @staticmethod
+ @torch.cuda.amp.custom_fwd
+ def forward(ctx, forward_function, hidden_states, *args):
+ saved_hidden_states = hidden_states.to("cpu", non_blocking = True)
+ with torch.no_grad():
+ (output,) = forward_function(hidden_states, *args)
+ ctx.save_for_backward(saved_hidden_states)
+ ctx.forward_function = forward_function
+ ctx.args = args
+ return output
+ pass
+
+ @staticmethod
+ @torch.cuda.amp.custom_bwd
+ def backward(ctx, dY):
+ (hidden_states,) = ctx.saved_tensors
+ hidden_states = hidden_states.to("cuda", non_blocking = True).detach()
+ hidden_states.requires_grad = True
+ with torch.enable_grad():
+ (output,) = ctx.forward_function(hidden_states, *ctx.args)
+ torch.autograd.backward(output, dY)
+ return (None, hidden_states.grad,) + (None,)*len(ctx.args)
+ pass
+pass
+
diff --git a/unsloth/unsloth/models/dpo.py b/unsloth/unsloth/models/dpo.py
new file mode 100644
index 0000000000000000000000000000000000000000..b7c7305bb3fcc570329f117d9ced06811ef3fb67
--- /dev/null
+++ b/unsloth/unsloth/models/dpo.py
@@ -0,0 +1,120 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+try:
+ from transformers.utils.notebook import (
+ IntervalStrategy,
+ NotebookTrainingTracker,
+ NotebookProgressCallback,
+ )
+ HAS_NOTEBOOK = True
+except:
+ HAS_NOTEBOOK = False
+pass
+
+DPOTrainer_metrics = [
+ "rewards/chosen",
+ "rewards/rejected",
+ "rewards/accuracies",
+ "rewards/margins",
+ "logps/rejected",
+ "logps/chosen",
+ "logits/rejected",
+ "logits/chosen",
+]
+set_DPOTrainer_metrics = frozenset(DPOTrainer_metrics)
+
+
+def NotebookProgressCallback_on_train_begin(self, args, state, control, **kwargs):
+ self.first_column = "Epoch" if args.evaluation_strategy == IntervalStrategy.EPOCH else "Step"
+ self.training_loss = 0
+ self.last_log = 0
+ column_names = [self.first_column] + ["Training Loss"]
+ if args.evaluation_strategy != IntervalStrategy.NO:
+ column_names.append("Validation Loss")
+ column_names += [x.replace("/", " / ") for x in DPOTrainer_metrics]
+ self.training_tracker = NotebookTrainingTracker(state.max_steps, column_names)
+pass
+
+
+def NotebookProgressCallback_on_log(self, args, state, control, logs=None, **kwargs):
+ # Only for when there is no evaluation
+ if args.evaluation_strategy == IntervalStrategy.NO and "loss" in logs:
+ values = {"Training Loss": logs["loss"]}
+ for metric in DPOTrainer_metrics:
+ values[metric.replace("/", " / ")] = logs[metric]
+ pass
+ # First column is necessarily Step since we're not in epoch eval strategy
+ values["Step"] = state.global_step
+ self.training_tracker.write_line(values)
+ pass
+pass
+
+
+def NotebookTrainingTracker_write_line(self, values):
+ """
+ Write the values in the inner table.
+
+ Args:
+ values (`Dict[str, float]`): The values to display.
+ """
+ if self.inner_table is None:
+ self.inner_table = [list(values.keys()), list(values.values())]
+ else:
+ columns = self.inner_table[0]
+ new_values = {}
+ for key, value in values.items():
+ lowered = key.lower()
+ if lowered in set_DPOTrainer_metrics:
+ new_values[lowered.replace("/", " / ")] = value
+ else:
+ new_values[key] = value
+ pass
+ values = new_values
+
+ self.inner_table[0] = columns
+ if len(self.inner_table) > 1:
+ last_values = self.inner_table[-1]
+ first_column = self.inner_table[0][0]
+ if last_values[0] != values[first_column]:
+ # write new line
+ self.inner_table.append([values[c] if c in values else "No Log" for c in columns])
+ else:
+ # update last line
+ new_values = values
+ for c in columns:
+ if c not in new_values.keys():
+ new_values[c] = last_values[columns.index(c)]
+ self.inner_table[-1] = [new_values[c] for c in columns]
+ else:
+ # Edit for evaluation purposes
+ self.inner_table.append([values[c] if c in values else 0 for c in columns])
+ pass
+ pass
+pass
+
+
+def PatchDPOTrainer():
+ if HAS_NOTEBOOK:
+ from transformers.trainer import is_in_notebook
+ if is_in_notebook():
+ # Patch DPO notebook printing
+ NotebookTrainingTracker.write_line = NotebookTrainingTracker_write_line
+ from transformers.trainer import DEFAULT_PROGRESS_CALLBACK
+ DEFAULT_PROGRESS_CALLBACK.on_train_begin = NotebookProgressCallback_on_train_begin
+ DEFAULT_PROGRESS_CALLBACK.on_log = NotebookProgressCallback_on_log
+ pass
+ pass
+pass
+
diff --git a/unsloth/unsloth/models/gemma.py b/unsloth/unsloth/models/gemma.py
new file mode 100644
index 0000000000000000000000000000000000000000..c0cce75e2f68ffabeb34e968155f2527b4df8ddf
--- /dev/null
+++ b/unsloth/unsloth/models/gemma.py
@@ -0,0 +1,350 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+from .llama import *
+from ._utils import __version__
+
+from transformers.models.gemma.modeling_gemma import (
+ GemmaAttention,
+ GemmaDecoderLayer,
+ GemmaModel,
+ GemmaForCausalLM,
+ GemmaRotaryEmbedding,
+ apply_rotary_pos_emb,
+ repeat_kv,
+)
+from transformers.modeling_attn_mask_utils import (
+ _prepare_4d_causal_attention_mask_for_sdpa,
+)
+# For Pytorch 2.1.1
+try:
+ from transformers.models.gemma.modeling_gemma import (
+ GemmaSdpaAttention,
+ GemmaFlashAttention2,
+ )
+except:
+ GemmaSdpaAttention = GemmaAttention
+ GemmaFlashAttention2 = GemmaAttention
+pass
+
+
+torch_nn_functional_gelu = torch.nn.functional.gelu
+def fast_geglu_inference(self, X):
+ # gate = self.gate_proj(X)
+ # up = self.up_proj(X)
+ bsz, _, hd = X.shape
+ # mlp_size = self.config.intermediate_size
+ # temp = torch.empty((2, bsz, 1, mlp_size), dtype = X.dtype, device = "cuda")
+
+ gate = fast_linear_forward(self.gate_proj, X)#, out = temp[0])
+ up = fast_linear_forward(self. up_proj, X)#, out = temp[1])
+ gate = torch_nn_functional_gelu(gate, approximate = "tanh")
+ gate *= up
+
+ # X = self.down_proj(gate)
+ down = fast_linear_forward(self.down_proj, gate, out = up[:,:,:hd])
+ return down
+pass
+
+
+# https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py#L590
+def GemmaDecoderLayer_fast_forward(
+ self,
+ hidden_states: torch.Tensor,
+ causal_mask: Optional[xformers.attn_bias.BlockDiagonalCausalMask] = None,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: Optional[bool] = False,
+ use_cache: Optional[bool] = False,
+ padding_mask: Optional[torch.LongTensor] = None,
+ *args, **kwargs,
+):
+ if use_cache: #past_key_value is not None:
+ out_weight = torch.empty(self.input_layernorm.weight.shape, dtype = torch.float32, device = "cuda")
+
+ # Self Attention
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm_inference_gemma(self.input_layernorm, hidden_states, out_weight)
+ hidden_states, self_attn_weights, present_key_value = self.self_attn(
+ hidden_states=hidden_states,
+ causal_mask=causal_mask,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ padding_mask=padding_mask,
+ )
+ hidden_states += residual
+
+ # Fully Connected
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm_inference_gemma(self.post_attention_layernorm, hidden_states, out_weight)
+ hidden_states = fast_geglu_inference(self.mlp, hidden_states)
+ hidden_states += residual
+ else:
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm(self.input_layernorm, hidden_states, gemma = True)
+ hidden_states, self_attn_weights, present_key_value = self.self_attn(
+ hidden_states=hidden_states,
+ causal_mask=causal_mask,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ padding_mask=padding_mask,
+ )
+ hidden_states = residual + hidden_states
+
+ # Fully Connected
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm(self.post_attention_layernorm, hidden_states, gemma = True)
+ hidden_states = self.mlp(hidden_states)
+ hidden_states = residual + hidden_states
+ pass
+
+ outputs = (hidden_states,)
+ if output_attentions: outputs += (self_attn_weights,)
+ if use_cache: outputs += (present_key_value,)
+ return outputs
+pass
+
+
+from math import sqrt as math_sqrt
+
+# https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py#L825
+# @torch.inference_mode
+def GemmaModel_fast_forward_inference(
+ self,
+ input_ids,
+ past_key_values,
+ position_ids,
+ attention_mask = None,
+):
+ out_weight = torch.empty_like(self.model.layers[0].input_layernorm.weight, dtype = torch.float32, device = "cuda")
+ input_ids = input_ids[:,:self.max_seq_length]
+ hidden_states = self.model.embed_tokens(input_ids)
+ hidden_states = hidden_states.to(self.config.torch_dtype)
+ # 3072**0.5 = 55.5000 in bfloat16, whilst 55.4256 in float32
+ # 2048**0.5 = 45.2500 in bfloat16, whilst 45.2548 in float32
+ hidden_states *= torch.tensor(math_sqrt(self.config.hidden_size), dtype = hidden_states.dtype)
+
+ bsz, q_len, hd = hidden_states.shape
+ seq_len = past_key_values[0][0].shape[-2]
+ if bsz != 1:
+ attention_mask = _prepare_4d_causal_attention_mask_for_sdpa(
+ attention_mask,
+ (bsz, q_len),
+ hidden_states,
+ seq_len,
+ )
+ pass
+
+ next_decoder_cache = []
+ for idx, decoder_layer in enumerate(self.model.layers):
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm_inference_gemma(decoder_layer.input_layernorm, hidden_states, out_weight)
+ hidden_states, present_key_value = LlamaAttention_fast_forward_inference(
+ decoder_layer.self_attn,
+ hidden_states = hidden_states,
+ past_key_value = past_key_values[idx],
+ position_ids = position_ids,
+ attention_mask = attention_mask,
+ do_prefill = not hasattr(decoder_layer.self_attn, "paged_attention"),
+ )
+ hidden_states += residual
+
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm_inference_gemma(decoder_layer.post_attention_layernorm, hidden_states, out_weight)
+ hidden_states = fast_geglu_inference(decoder_layer.mlp, hidden_states)
+ hidden_states += residual
+
+ next_decoder_cache.append(present_key_value)
+ pass
+ hidden_states = fast_rms_layernorm_inference_gemma(self.model.norm, hidden_states, out_weight)
+
+ return BaseModelOutputWithPast(
+ last_hidden_state = hidden_states,
+ past_key_values = next_decoder_cache,
+ hidden_states = [],
+ attentions = [],
+ )
+pass
+
+
+# Follows line by line https://github.com/google-deepmind/gemma/blob/main/gemma/positional_embeddings.py#L45
+# Formulates cos and sin differently from Llama!
+class GemmaFixedRotaryEmbedding(torch.nn.Module):
+ # Fixes https://github.com/huggingface/transformers/pull/28837
+ # https://github.com/microsoft/DeepSpeed/issues/4932
+ # The precision of RoPE buffers is not correct, so we cast to int64.
+ def __init__(self, dim, max_position_embeddings=2048, base=10000, device=None):
+ super().__init__()
+ self.dim = dim
+ self.max_position_embeddings = max_position_embeddings
+ self.base = base
+
+ # Build here to make `torch.jit.trace` work.
+ self._set_cos_sin_cache(seq_len=max_position_embeddings, device=device, dtype=torch.get_default_dtype())
+ pass
+
+ def _set_cos_sin_cache(self, seq_len, device, dtype):
+ # Note: on the original Llama codebase, these tensors are created on the target device (and not on CPU) and
+ # in FP32. They are applied (multiplied) in FP32 as well.
+ self.max_seq_len_cached = seq_len
+
+ # The difference is we do division explicity instead of t * (1/x) ie we do t/x.
+ freq_exponents = (2.0 / self.dim) * (
+ torch.arange(self.dim // 2, dtype = torch.int64, device = "cpu").float()
+ )
+ timescale = self.base**freq_exponents
+ positions = torch.arange(self.max_seq_len_cached, device = "cpu", dtype = torch.int64).float()
+ radians_new = positions[..., None] / timescale[None, None, :]
+ radians_new = radians_new.squeeze(0)
+
+ emb = torch.cat((radians_new, radians_new), dim = -1)
+ # We must do RoPE in float32!
+ cos = emb.cos().to(device = device, non_blocking = True)#, dtype = dtype)
+ sin = emb.sin().to(device = device, non_blocking = True)#, dtype = dtype)
+ self.register_buffer("cos_cached", cos, persistent = False)
+ self.register_buffer("sin_cached", sin, persistent = False)
+ pass
+
+ def forward(self, x, position_ids=None, seq_len=None):
+ # x: [bs, num_attention_heads, seq_len, head_size]
+ if seq_len > self.max_seq_len_cached:
+ self._set_cos_sin_cache(seq_len=seq_len, device=x.device, dtype=x.dtype)
+
+ return (
+ self.cos_cached[:seq_len].to(dtype=x.dtype),
+ self.sin_cached[:seq_len].to(dtype=x.dtype),
+ )
+ pass
+pass
+
+
+class FastGemmaModel(FastLlamaModel):
+
+ @staticmethod
+ def pre_patch():
+ GemmaAttention .forward = LlamaAttention_fast_forward
+ GemmaSdpaAttention .forward = LlamaAttention_fast_forward
+ GemmaFlashAttention2.forward = LlamaAttention_fast_forward
+ GemmaDecoderLayer .forward = GemmaDecoderLayer_fast_forward
+ GemmaModel .forward = LlamaModel_fast_forward
+ GemmaForCausalLM .forward = CausalLM_fast_forward(GemmaModel_fast_forward_inference)
+ PeftModelForCausalLM.forward = PeftModelForCausalLM_fast_forward
+ # Solves https://github.com/unslothai/unsloth/issues/168
+ # Static KV Cache was introduced in 4.38.0, causing training to be much slower.
+ # Inferene can now be CUDAGraphed, but we shall retain the old rotary embeddings.
+ # https://github.com/huggingface/transformers/pull/27931
+ # https://github.com/huggingface/transformers/blob/v4.37.2/src/transformers/models/llama/modeling_llama.py
+ import transformers.models.gemma.modeling_gemma
+ transformers.models.gemma.modeling_gemma.GemmaRotaryEmbedding = GemmaFixedRotaryEmbedding
+ return
+ pass
+
+
+ @staticmethod
+ def post_patch(model):
+ # Patch model for Gemma
+ layers = model.model.layers
+
+ # Torch.compile fails on embedding matrix??
+ # Workaround randomnly fixes it for torch versions < 2.2
+ model.model.embed_tokens = torch.nn.Embedding.from_pretrained(model.model.embed_tokens.weight)
+ model.config.update({"unsloth_version" : __version__})
+
+ # We also do this for the lm_head
+ lm_head = torch.nn.Linear(1, 1, bias = None)
+ del lm_head.weight
+ lm_head.weight = model.lm_head.weight
+ lm_head.in_features = lm_head.weight.shape[1]
+ lm_head.out_features = lm_head.weight.shape[0]
+ model.lm_head = lm_head
+
+ # Gemma has tied weights! This means lm_head == embed_tokens
+ if model.model.embed_tokens.weight.data_ptr() != model.lm_head.weight.data_ptr():
+ lm_head = torch.nn.Linear(1, 1, bias = None)
+ del lm_head.weight
+ lm_head.weight = model.model.embed_tokens.weight
+ lm_head.in_features = lm_head.weight.shape[1]
+ lm_head.out_features = lm_head.weight.shape[0]
+ model.lm_head = lm_head
+ pass
+
+ # Also patch all dtypes - BnB seems to not allocate the correct type?
+ # BnB default dtype seems to be float16!
+ correct_dtype = lm_head.weight.dtype
+
+ for name, module in model.named_modules():
+ if isinstance(module, (Bnb_Linear4bit, Peft_Linear4bit)):
+ weight = module.weight
+ quant_state = weight.quant_state
+
+ if type(quant_state) is list:
+ # BnB seems to have float16 as default!
+ module.weight.quant_state[2] = correct_dtype # Cast to correct dtype
+ else:
+ # https://github.com/TimDettmers/bitsandbytes/pull/763/files
+ quant_state.dtype = correct_dtype
+ pass
+ pass
+ # Downcast RoPE embedding to correct data type
+ # RoPE must be done in float32 for Gemma
+ # if (name.endswith("rotary_emb") or hasattr(module, "cos_cached")) \
+ # and (module.cos_cached.dtype != correct_dtype):
+
+ # module.cos_cached = module.cos_cached.to(correct_dtype)
+ # module.sin_cached = module.sin_cached.to(correct_dtype)
+ # pass
+ # pass
+ pass
+
+ # Add 1 to weight
+ # return output * (1 + self.weight)
+ # https://github.com/huggingface/transformers/blob/main/src/transformers/models/gemma/modeling_gemma.py#L89
+ from transformers.models.gemma.modeling_gemma import GemmaRMSNorm
+
+ # Freeze all parameters except LoRA
+ # We do this first since += 1 seems to not be liked by requires_grad = True
+ for name, param in model.named_parameters():
+ if ".lora_A." in name or ".lora_B." in name:
+ param.requires_grad_(True)
+ else:
+ param.requires_grad_(False)
+ pass
+
+ # Patch RMS Layernorm
+ for name, module in model.named_modules():
+ if isinstance(module, GemmaRMSNorm):
+ # Must be in float32
+ # https://github.com/keras-team/keras-nlp/blob/v0.8.2/keras_nlp/models/gemma/rms_normalization.py#L36
+ # module = module.to(torch.float32)
+ # Leave + 1 to Triton kernel itself
+ # module.weight += 1.0 # return output * (1 + self.weight)
+ if not hasattr(module, "variance_epsilon"):
+ module.variance_epsilon = module.eps # Gemma doesn't use variance_epsilon
+ pass
+
+ # Clear deleted GPU items
+ import gc
+ for _ in range(3):
+ gc.collect()
+ torch.cuda.empty_cache()
+ return model
+ pass
+pass
diff --git a/unsloth/unsloth/models/llama.py b/unsloth/unsloth/models/llama.py
new file mode 100644
index 0000000000000000000000000000000000000000..f08c403074fffd8e78642196b8ac559159f52ae7
--- /dev/null
+++ b/unsloth/unsloth/models/llama.py
@@ -0,0 +1,1843 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import torch
+from typing import Optional, Tuple, List, Union
+from torch.nn.functional import scaled_dot_product_attention
+from transformers.models.llama.modeling_llama import (
+ logger,
+ BaseModelOutputWithPast,
+ CausalLMOutputWithPast,
+)
+from transformers.modeling_attn_mask_utils import (
+ _prepare_4d_causal_attention_mask_for_sdpa,
+)
+from ..kernels import *
+from ._utils import *
+from ._utils import __version__
+from ..tokenizer_utils import *
+if HAS_FLASH_ATTENTION:
+ from flash_attn import flash_attn_func
+
+# Final patching code
+from transformers.models.llama.modeling_llama import (
+ LlamaAttention,
+ LlamaDecoderLayer,
+ LlamaModel,
+ LlamaForCausalLM,
+)
+
+# For Pytorch 2.1.1
+try:
+ from transformers.models.llama.modeling_llama import (
+ LlamaSdpaAttention,
+ LlamaFlashAttention2,
+ )
+except:
+ LlamaSdpaAttention = LlamaAttention
+ LlamaFlashAttention2 = LlamaAttention
+pass
+
+from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig, AutoConfig
+from transformers import set_seed as transformers_set_seed
+from peft import LoraConfig, TaskType, get_peft_model as _get_peft_model
+from peft import PeftModelForCausalLM
+from bitsandbytes.nn import Linear4bit as Bnb_Linear4bit
+from peft.tuners.lora import Linear4bit as Peft_Linear4bit
+from ..save import patch_saving_functions
+import re, os, inspect, math, sys
+
+
+def original_apply_qkv(self, X):
+ Q = self.q_proj(X)
+ K = self.k_proj(X)
+ V = self.v_proj(X)
+ return Q, K, V
+pass
+
+
+def original_apply_o(self, X):
+ O = self.o_proj(X)
+ return O
+pass
+
+
+from math import sqrt as math_sqrt
+KV_CACHE_INCREMENT = 256 # KV Cache update size
+torch_nn_functional_softmax = torch.nn.functional.softmax
+
+def LlamaAttention_fast_forward_inference(
+ self,
+ hidden_states: torch.Tensor,
+ past_key_value: Optional[Tuple[torch.Tensor]],
+ position_ids,
+ do_prefill = False,
+ attention_mask = None,
+):
+ """
+ https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py#L406
+ Fast inference using KV cache.
+ QK^T can be computed in 4 chunks
+
+ [Q, q] @ [K, k].T where q, k are the new tokens.
+ [QK^T, Qk^T]
+ [qK^T, qk^T]
+
+ Since the attention mask wipes Qk^T, we just get
+ [QK^T, 0]
+ [qK^T, qk^T]
+
+ Since softmax is row-wise, we get
+ softmax([QK^T, 0])
+ softmax([qK^T, qk^T])
+
+ We then multiply by [V]
+ [v]
+ softmax([QK^T, 0]) [softmax(QK^T)V] *
+ softmax([qK^T, qk^T]) [softmax([qK^T, qk^T]) @ [V, v]]
+
+ But notice * [softmax(QK^T)V] is just the last attention.
+ We just need to compute the last final row.
+
+ This means we can pass in a row of Q, but we need to
+ remember K and V, which are called the KV cache.
+ """
+ Xn = hidden_states
+ bsz, _, hd = hidden_states.size()
+ K1, V1 = past_key_value
+ dtype = Xn.dtype
+
+ n_heads = self.num_heads
+ n_groups = self.num_key_value_groups
+ n_kv_heads = self.num_key_value_heads
+ head_dim = self.head_dim
+ attention_size = n_heads*head_dim
+ # assert(n_kv_heads * n_groups == n_heads)
+ seq_len = K1.shape[-2]
+ kv_seq_len = seq_len + 1
+
+ # Prefill phase
+ # if not hasattr(self, "paged_attention"):
+ if do_prefill:
+ self.paged_attention = torch.empty((KV_CACHE_INCREMENT+seq_len+1, 2, bsz, n_kv_heads, head_dim), dtype = dtype, device = "cuda")
+ self.paged_attention_K = self.paged_attention[:,0]
+ self.paged_attention_V = self.paged_attention[:,1]
+ self.paged_attention_K[:seq_len] = K1.permute(2, 0, 1, 3)
+ self.paged_attention_V[:seq_len] = V1.permute(2, 0, 1, 3)
+ self.temp_QA = torch.empty((2, bsz, 1, attention_size), dtype = dtype, device = "cuda")
+ self.temp_KV = torch.empty((2, bsz, 1, n_kv_heads*head_dim), dtype = dtype, device = "cuda")
+ self.RH_Q = torch.empty((bsz, n_heads, 1, head_dim), dtype = dtype, device = "cuda")
+ self.attention = torch.empty((bsz, n_heads, 1, KV_CACHE_INCREMENT+seq_len), dtype = dtype, device = "cuda")
+ self.scalar = 1.0 / math_sqrt(self.head_dim)
+ self.half_head_dim = head_dim // 2
+ elif kv_seq_len >= self.paged_attention.shape[0]:
+ self.paged_attention.resize_((self.paged_attention.shape[0]+KV_CACHE_INCREMENT, 2, bsz, n_kv_heads, head_dim))
+ self.paged_attention_K = self.paged_attention[:,0]
+ self.paged_attention_V = self.paged_attention[:,1]
+ self.attention.resize_((bsz, n_heads, 1, self.attention.shape[-1]+KV_CACHE_INCREMENT))
+ pass
+
+ Qn = fast_linear_forward(self.q_proj, Xn, out = self.temp_QA[0])
+ Kn = fast_linear_forward(self.k_proj, Xn, out = self.temp_KV[0])
+ Vn = fast_linear_forward(self.v_proj, Xn, out = self.temp_KV[1])
+ Qn = Qn.view(bsz, 1, n_heads, head_dim).transpose(1, 2)
+ Kn = Kn.view(bsz, 1, n_kv_heads, head_dim).transpose(1, 2)
+ Vn = Vn.view(bsz, 1, n_kv_heads, head_dim).transpose(1, 2)
+
+ # cos, sin = self.rotary_emb(Vn, seq_len = kv_seq_len)
+ # Qn, Kn = inplace_rope_embedding(Qn, Kn, cos, sin, position_ids)
+ cos = self.rotary_emb.cos_cached[position_ids].unsqueeze(1)
+ sin = self.rotary_emb.sin_cached[position_ids].unsqueeze(1)
+ h = self.half_head_dim
+
+ RH_Q = self.RH_Q
+ RH_Q[:,:,:,:h] = Qn[:,:,:,h:]
+ RH_Q[:,:,:,h:] = Qn[:,:,:,:h]
+ torch.neg(RH_Q[:,:,:,:h], out = RH_Q[:,:,:,:h])
+ Qn *= cos
+ Qn.addcmul_(RH_Q, sin)
+
+ RH_K = RH_Q[:,:n_kv_heads,:,:] # torch.empty((n_kv_heads, 1, head_dim), dtype = dtype, device = "cuda")
+ RH_K[:,:,:,:h] = Kn[:,:,:,h:]
+ RH_K[:,:,:,h:] = Kn[:,:,:,:h]
+ torch.neg(RH_K[:,:,:,:h], out = RH_K[:,:,:,:h])
+ Kn *= cos
+ Kn.addcmul_(RH_K, sin)
+
+ # New KV cache
+ # Kn = torch.cat([K1, Kn], dim = 2)
+ # Vn = torch.cat([V1, Vn], dim = 2)
+ self.paged_attention_K[seq_len] = Kn.permute(2, 0, 1, 3)
+ self.paged_attention_V[seq_len] = Vn.permute(2, 0, 1, 3)
+ Kn = self.paged_attention_K[:kv_seq_len].permute(1, 2, 0, 3)
+ Vn = self.paged_attention_V[:kv_seq_len].permute(1, 2, 0, 3)
+
+ # Handle sliding windows
+ sliding_window = getattr(self.config, "sliding_window", None)
+ if sliding_window is not None and kv_seq_len > sliding_window:
+ # From https://github.com/huggingface/transformers/blob/main/src/transformers/models/mistral/modeling_mistral.py#L193
+ slicing_tokens = 1 - sliding_window
+ Knn = Kn[:, :, slicing_tokens:, :]#.contiguous()
+ Vnn = Vn[:, :, slicing_tokens:, :]#.contiguous()
+ else:
+ Knn, Vnn = Kn, Vn
+ pass
+
+ # Grouped query attention
+ _, _, cached_len, _ = Knn.shape
+ if n_groups != 1:
+ Knn = Knn[:, :, None, :, :].expand(bsz, n_kv_heads, n_groups, cached_len, head_dim)
+ Vnn = Vnn[:, :, None, :, :].expand(bsz, n_kv_heads, n_groups, cached_len, head_dim)
+ Knn = Knn.reshape(bsz, n_heads, cached_len, head_dim)
+ Vnn = Vnn.reshape(bsz, n_heads, cached_len, head_dim)
+ pass
+ # else:
+ # Knn, Vnn = Knn, Vnn
+ # pass
+
+ # Attention
+ if bsz == 1:
+ A = torch.matmul(Qn, Knn.transpose(2, 3), out = self.attention[:,:,:,:cached_len])
+ A *= self.scalar
+ # if attention_mask is not None: A += attention_mask # Must add attention_mask for batched
+ A[:] = torch_nn_functional_softmax(A, dim = -1, dtype = torch.float32)#.to(A.dtype)
+ A = torch.matmul(A, Vnn, out = Qn)
+ else:
+ A = scaled_dot_product_attention(Qn, Knn, Vnn, attn_mask = attention_mask, is_causal = False)
+ pass
+ A = A.transpose(1, 2)
+ A = A.reshape(bsz, 1, attention_size)
+ A = fast_linear_forward(self.o_proj, A, out = self.temp_QA[1][:,:,:self.hidden_size])
+ return A, (Kn, Vn)
+pass
+
+
+torch_nn_functional_silu = torch.nn.functional.silu
+def fast_swiglu_inference(self, X):
+ # gate = self.gate_proj(X)
+ # up = self.up_proj(X)
+ bsz, _, hd = X.shape
+ # mlp_size = self.config.intermediate_size
+ # temp = torch.empty((2, bsz, 1, mlp_size), dtype = X.dtype, device = "cuda")
+
+ gate = fast_linear_forward(self.gate_proj, X)#, out = temp[0])
+ up = fast_linear_forward(self. up_proj, X)#, out = temp[1])
+ gate = torch_nn_functional_silu(gate, inplace = True)
+ gate *= up
+
+ # X = self.down_proj(gate)
+ down = fast_linear_forward(self.down_proj, gate, out = up[:,:,:hd])
+ return down
+pass
+
+
+def fast_rms_layernorm_inference(self, X):
+ old_dtype = X.dtype
+ XX = X.to(torch.float32)
+ variance = XX.square().mean(-1, keepdim = True)
+ variance += self.variance_epsilon
+ XX *= variance.rsqrt_()
+ X = XX.to(old_dtype) # Must preserve due to residual
+ X *= self.weight
+ return X
+pass
+
+
+def fast_rms_layernorm_inference_gemma(self, X, out_weight = None):
+ XX = X.to(torch.float32)
+ variance = XX.square().mean(-1, keepdim = True)
+ variance += self.variance_epsilon
+ XX *= variance.rsqrt_()
+
+ if out_weight is None:
+ out_weight = self.weight + 1.0
+ else:
+ out_weight[:] = self.weight
+ out_weight += 1.0
+ pass
+
+ XX *= out_weight
+ return XX.to(X.dtype)
+pass
+
+
+# https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py#L320
+def LlamaAttention_fast_forward(
+ self,
+ hidden_states: torch.Tensor,
+ causal_mask: Optional[xformers.attn_bias.BlockDiagonalCausalMask] = None,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: bool = False,
+ use_cache: bool = False,
+ padding_mask: Optional[torch.LongTensor] = None,
+ *args, **kwargs,
+) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
+
+ # Clear inference
+ if hasattr(self, "paged_attention"):
+ del self.paged_attention_K
+ del self.paged_attention_V
+ del self.paged_attention
+ del self.temp_QA
+ del self.temp_KV
+ del self.RH_Q
+ del self.attention
+ pass
+
+ bsz, q_len, _ = hidden_states.size()
+
+ n_heads = self.num_heads
+ n_groups = self.num_key_value_groups
+ n_kv_heads = self.num_key_value_heads
+ head_dim = self.head_dim
+ assert(n_kv_heads * n_groups == n_heads)
+
+ Q, K, V = self.apply_qkv(self, hidden_states)
+ Q = Q.view(bsz, q_len, n_heads, head_dim).transpose(1, 2)
+ K = K.view(bsz, q_len, n_kv_heads, head_dim).transpose(1, 2)
+ V = V.view(bsz, q_len, n_kv_heads, head_dim).transpose(1, 2)
+
+ kv_seq_len = K.shape[-2]
+ if past_key_value is not None:
+ kv_seq_len += past_key_value[0].shape[-2]
+
+ if position_ids is None:
+ cos = self.rotary_emb.cos_cached
+ sin = self.rotary_emb.sin_cached
+ Q, K = fast_rope_embedding(Q, K, cos, sin)
+ else:
+ cos, sin = self.rotary_emb(V, seq_len = kv_seq_len)
+ Q, K = inplace_rope_embedding(Q, K, cos, sin, position_ids)
+ pass
+
+ if past_key_value is not None:
+ K = torch.cat([past_key_value[0], K], dim = 2)
+ V = torch.cat([past_key_value[1], V], dim = 2)
+ pass
+ past_key_value = (K, V) if use_cache else None
+
+ # Attention module
+ if (not HAS_FLASH_ATTENTION and attention_mask is None):
+ # Xformers memory efficient attention
+ # Also has Flash Attention v2 dispatching
+ Q = Q.transpose(1, 2)
+ K = K.transpose(1, 2)
+ V = V.transpose(1, 2)
+
+ # Group query attention
+ if n_groups != 1:
+ K = K .view(bsz, kv_seq_len, n_kv_heads, 1, head_dim)
+ V = V .view(bsz, kv_seq_len, n_kv_heads, 1, head_dim)
+ K = K.expand(bsz, kv_seq_len, n_kv_heads, n_groups, head_dim)
+ V = V.expand(bsz, kv_seq_len, n_kv_heads, n_groups, head_dim)
+ if hidden_states.requires_grad:
+ K = K.reshape(bsz, kv_seq_len, n_heads, head_dim)
+ V = V.reshape(bsz, kv_seq_len, n_heads, head_dim)
+ else:
+ Q = Q.view(bsz, q_len, n_kv_heads, n_groups, head_dim)
+ pass
+ A = xformers_attention(Q, K, V, attn_bias = causal_mask)
+ A = A.view(bsz, q_len, n_heads, head_dim)
+
+ elif HAS_FLASH_ATTENTION and attention_mask is None:
+ Q = Q.transpose(1, 2)
+ K = K.transpose(1, 2)
+ V = V.transpose(1, 2)
+ A = flash_attn_func(Q, K, V, causal = True)
+ else:
+ # Grouped query attention
+ if n_groups != 1:
+ K = K[:, :, None, :, :].expand(bsz, n_kv_heads, n_groups, kv_seq_len, head_dim)
+ V = V[:, :, None, :, :].expand(bsz, n_kv_heads, n_groups, kv_seq_len, head_dim)
+ K = K.reshape(bsz, n_heads, kv_seq_len, head_dim)
+ V = V.reshape(bsz, n_heads, kv_seq_len, head_dim)
+ pass
+ # Must be contiguous or else results are False!
+ # https://github.com/pytorch/pytorch/issues/112577
+ Q, K, V = Q.contiguous(), K.contiguous(), V.contiguous()
+ # Needs (batch_size, n_heads, seq_len, head_dim)
+ # is_casual and attention_mask must not be both set!
+ A = scaled_dot_product_attention(Q, K, V, attn_mask = attention_mask, is_causal = False)
+ # Go back to (batch_size, seq_len, n_heads, head_dim)
+ A = A.transpose(1, 2).contiguous()
+ pass
+ attn_output = A.reshape(bsz, q_len, n_heads*head_dim)
+ attn_output = self.apply_o(self, attn_output)
+ attn_weights = None
+ return attn_output, attn_weights, past_key_value
+pass
+
+
+# https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py#L590
+def LlamaDecoderLayer_fast_forward(
+ self,
+ hidden_states: torch.Tensor,
+ causal_mask: Optional[xformers.attn_bias.BlockDiagonalCausalMask] = None,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: Optional[bool] = False,
+ use_cache: Optional[bool] = False,
+ padding_mask: Optional[torch.LongTensor] = None,
+ *args, **kwargs,
+) -> Tuple[torch.FloatTensor, Optional[Tuple[torch.FloatTensor, torch.FloatTensor]]]:
+ """
+ Args:
+ hidden_states (`torch.FloatTensor`): input to the layer of shape `(batch, seq_len, embed_dim)`
+ attention_mask (`torch.FloatTensor`, *optional*): attention mask of size
+ `(batch, 1, tgt_len, src_len)` where padding elements are indicated by very large negative values.
+ output_attentions (`bool`, *optional*):
+ Whether or not to return the attentions tensors of all attention layers. See `attentions` under
+ returned tensors for more detail.
+ use_cache (`bool`, *optional*):
+ If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding
+ (see `past_key_values`).
+ past_key_value (`Tuple(torch.FloatTensor)`, *optional*): cached past key and value projection states
+ """
+ if use_cache:
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm_inference(self.input_layernorm, hidden_states)
+ hidden_states, self_attn_weights, present_key_value = self.self_attn(
+ hidden_states=hidden_states,
+ causal_mask=causal_mask,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ padding_mask=padding_mask,
+ )
+ hidden_states += residual
+
+ # Fully Connected
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm_inference(self.post_attention_layernorm, hidden_states)
+ hidden_states = fast_swiglu_inference(self.mlp, hidden_states)
+ hidden_states += residual
+ else:
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm(self.input_layernorm, hidden_states)
+ hidden_states, self_attn_weights, present_key_value = self.self_attn(
+ hidden_states=hidden_states,
+ causal_mask=causal_mask,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ padding_mask=padding_mask,
+ )
+ hidden_states = residual + hidden_states
+
+ # Fully Connected
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm(self.post_attention_layernorm, hidden_states)
+ hidden_states = self.mlp(hidden_states)
+ hidden_states = residual + hidden_states
+ pass
+
+ outputs = (hidden_states,)
+ if output_attentions: outputs += (self_attn_weights,)
+ if use_cache: outputs += (present_key_value,)
+ return outputs
+pass
+
+
+# https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py#L825
+def LlamaModel_fast_forward(
+ self,
+ input_ids: torch.LongTensor,
+ causal_mask: Optional[xformers.attn_bias.BlockDiagonalCausalMask] = None,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_values: Optional[List[torch.FloatTensor]] = None,
+ inputs_embeds: Optional[torch.FloatTensor] = None,
+ use_cache: Optional[bool] = None,
+ output_attentions: Optional[bool] = None,
+ output_hidden_states: Optional[bool] = None,
+ return_dict: Optional[bool] = None,
+ *args, **kwargs,
+) -> Union[Tuple, BaseModelOutputWithPast]:
+
+ output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions
+ assert(output_attentions is False)
+ output_hidden_states = (
+ output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states
+ )
+ use_cache = use_cache if use_cache is not None else self.config.use_cache
+
+ return_dict = return_dict if return_dict is not None else self.config.use_return_dict
+
+ # retrieve input_ids and inputs_embeds
+ if input_ids is not None and inputs_embeds is not None:
+ raise ValueError("Unsloth: You cannot specify both decoder_input_ids and decoder_inputs_embeds at the same time")
+ elif input_ids is not None:
+ batch_size, seq_length = input_ids.shape
+ elif inputs_embeds is not None:
+ batch_size, seq_length, _ = inputs_embeds.shape
+ else:
+ raise ValueError("Unsloth: You have to specify either decoder_input_ids or decoder_inputs_embeds")
+
+ seq_length_with_past = seq_length
+
+ # Fix out of bounds tokenization
+ if hasattr(self, "max_seq_length"):
+ if seq_length > self.max_seq_length:
+ logger.warning_once(
+ f"Unsloth: Input IDs of length {seq_length} > the model's max sequence length of {self.max_seq_length}.\n"\
+ "We shall truncate it ourselves. It's imperative if you correct this issue first."
+ )
+ if input_ids is not None:
+ input_ids = input_ids[:,:self.max_seq_length]
+ elif inputs_embeds is not None:
+ inputs_embeds = inputs_embeds[:,:self.max_seq_length,:]
+ pass
+ pass
+
+ past_key_values_length = 0
+
+ if past_key_values is not None:
+ past_key_values_length = past_key_values[0][0].shape[2]
+ seq_length_with_past = seq_length_with_past + past_key_values_length
+ pass
+
+ # We already handle KV cache position_ids ourselves.
+ if False:#(past_key_values_length != 0):
+ position_ids = torch.arange(
+ past_key_values_length, seq_length + past_key_values_length,
+ dtype = torch.int32,
+ device = "cuda",
+ )
+ position_ids = position_ids.unsqueeze(0).view(-1, seq_length)
+ elif position_ids is not None:
+ position_ids = position_ids.view(-1, seq_length).to(torch.int32)#.long()
+ else:
+ position_ids = None
+ pass
+
+ if position_ids is not None:
+ if position_ids.shape[0] != batch_size:
+ position_ids = position_ids.repeat((batch_size, 1))
+ pass
+
+ # Embed positions
+ if inputs_embeds is None:
+ inputs_embeds = self.embed_tokens(input_ids)
+
+ inputs_embeds = inputs_embeds.to(self.config.torch_dtype)
+
+ # Normalized from Gemma
+ IS_GEMMA = self.config.model_type == "gemma"
+ train_embed_tokens = self.embed_tokens.weight.requires_grad
+
+ if IS_GEMMA:
+ # Match Gemma exactly by casting to bfloat16 / float16
+ # inputs_embeds *= math_sqrt(self.config.hidden_size)
+ # Ie 3072**0.5 = 55.5000 in bfloat16, whilst 55.4256 in float32
+ # & 2048**0.5 = 45.2500 in bfloat16, whilst 45.2548 in float32
+ normalizer = torch.tensor(math_sqrt(self.config.hidden_size), dtype = inputs_embeds.dtype)
+
+ if train_embed_tokens:
+ # Careful we must not do an inplace op!
+ inputs_embeds = inputs_embeds * normalizer
+ else:
+ inputs_requires_grad = inputs_embeds.requires_grad
+ if not inputs_embeds.is_leaf:
+ inputs_embeds = inputs_embeds.detach()
+ inputs_requires_grad = True
+ elif inputs_requires_grad:
+ inputs_embeds.requires_grad_(False)
+ pass
+ inputs_embeds *= normalizer
+ # inputs_embeds *= math_sqrt(self.config.hidden_size)
+ if inputs_requires_grad: inputs_embeds.requires_grad_(True)
+ pass
+ pass
+
+ # Fix up attention mask by setting elements to 0
+ # Specifically for DPO
+ if self._has_no_labels and (attention_mask is not None) and (past_key_values is None) and \
+ (not train_embed_tokens):
+ # Careful for inference the attention_mask is size (1, kv_seq_len)
+ # Whilst the input_embeds is size (1, 1, 4096)
+ inputs_requires_grad = inputs_embeds.requires_grad
+ if not inputs_embeds.is_leaf:
+ inputs_embeds = inputs_embeds.detach()
+ inputs_requires_grad = True
+ elif inputs_requires_grad:
+ inputs_embeds.requires_grad_(False)
+ pass
+ inputs_embeds *= attention_mask.unsqueeze(0).transpose(0, 1).transpose(1, 2)
+ if inputs_requires_grad: inputs_embeds.requires_grad_(True)
+ pass
+
+ # Ignore attention_mask
+ if attention_mask is None:
+ padding_mask = None
+ elif self.training:
+ attention_mask = None
+ padding_mask = None
+ else:
+ # if 0 in attention_mask:
+ # padding_mask = attention_mask
+ # else:
+ padding_mask = None
+
+ attention_mask = _prepare_4d_causal_attention_mask_for_sdpa(
+ attention_mask,
+ (batch_size, seq_length),
+ inputs_embeds,
+ past_key_values_length,
+ sliding_window = getattr(self.config, "sliding_window", None),
+ )
+ pass
+
+ hidden_states = inputs_embeds
+
+ if past_key_values is None and self.training:
+ use_cache = False
+ # if use_cache:
+ # logger.warning_once(
+ # "Unsloth: `use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`"
+ # )
+ # use_cache = False
+ pass
+
+ # decoder layers
+ all_hidden_states = () if output_hidden_states else None
+ all_self_attns = () if output_attentions else None
+ next_decoder_cache = () if use_cache else None
+
+ # Gradient checkpointing methods (ie sqrt)
+ if hasattr(self, "_gradient_checkpointing_boundaries"):
+ boundaries = self._gradient_checkpointing_boundaries
+ else:
+ boundaries = None
+ pass
+
+ # Check checkpointing method
+ gradient_checkpointing = False
+ offloaded_gradient_checkpointing = False
+
+ if (self.gradient_checkpointing and self.training and not use_cache):
+
+ gradient_checkpointing = True
+
+ if output_attentions is False and hasattr(self, "_offloaded_gradient_checkpointing"):
+ offloaded_gradient_checkpointing = True
+ pass
+
+ # Go through every layer!
+ for idx, decoder_layer in enumerate(self.layers):
+
+ if output_hidden_states: all_hidden_states += (hidden_states,)
+ past_key_value = past_key_values[idx] if past_key_values is not None else None
+
+ if offloaded_gradient_checkpointing:
+ hidden_states = Unsloth_Offloaded_Gradient_Checkpointer.apply(
+ decoder_layer,
+ hidden_states,
+ causal_mask,
+ attention_mask,
+ position_ids,
+ past_key_values,
+ output_attentions,
+ use_cache,
+ )
+
+ elif gradient_checkpointing:
+ def create_custom_forward(module):
+ def custom_forward(*inputs):
+ return module(*inputs, past_key_value, output_attentions, padding_mask = padding_mask)
+ return custom_forward
+ pass
+
+ layer_outputs = torch.utils.checkpoint.checkpoint(
+ create_custom_forward(decoder_layer),
+ hidden_states,
+ causal_mask,
+ attention_mask,
+ position_ids,
+ use_reentrant = True,
+ preserve_rng_state = False,
+ )
+ hidden_states = layer_outputs[0]
+
+ else:
+ layer_outputs = decoder_layer(
+ hidden_states,
+ causal_mask=causal_mask,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_value=past_key_value,
+ output_attentions=output_attentions,
+ use_cache=use_cache,
+ padding_mask=padding_mask,
+ )
+ hidden_states = layer_outputs[0]
+ pass
+
+ if use_cache: next_decoder_cache += (layer_outputs[2 if output_attentions else 1],)
+ if output_attentions: all_self_attns += (layer_outputs[1],)
+ pass
+
+ # Final layernorm
+ if use_cache:
+ hidden_states = (fast_rms_layernorm_inference_gemma if IS_GEMMA else fast_rms_layernorm_inference)\
+ (self.norm, hidden_states)
+ else:
+ hidden_states = fast_rms_layernorm(self.norm, hidden_states, gemma = IS_GEMMA)
+ pass
+
+ if output_hidden_states: all_hidden_states += (hidden_states,)
+ next_cache = next_decoder_cache if use_cache else None
+
+ if not return_dict:
+ return tuple(v for v in [hidden_states, next_cache, all_hidden_states, all_self_attns] if v is not None)
+ return BaseModelOutputWithPast(
+ last_hidden_state=hidden_states,
+ past_key_values=next_cache,
+ hidden_states=all_hidden_states,
+ attentions=all_self_attns,
+ )
+pass
+
+
+# https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py#L825
+def LlamaModel_fast_forward_inference(
+ self,
+ input_ids,
+ past_key_values,
+ position_ids,
+ attention_mask = None,
+):
+ input_ids = input_ids[:,:self.max_seq_length]
+ hidden_states = self.model.embed_tokens(input_ids)
+ hidden_states = hidden_states.to(self.config.torch_dtype)
+ bsz, q_len, hd = hidden_states.shape
+ seq_len = past_key_values[0][0].shape[-2]
+ if bsz != 1:
+ attention_mask = _prepare_4d_causal_attention_mask_for_sdpa(
+ attention_mask,
+ (bsz, q_len),
+ hidden_states,
+ seq_len,
+ sliding_window = getattr(self.config, "sliding_window", None),
+ )
+ else:
+ attention_mask = None
+ pass
+
+ next_decoder_cache = []
+ for idx, decoder_layer in enumerate(self.model.layers):
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm_inference(decoder_layer.input_layernorm, hidden_states)
+ hidden_states, present_key_value = LlamaAttention_fast_forward_inference(
+ decoder_layer.self_attn,
+ hidden_states = hidden_states,
+ past_key_value = past_key_values[idx],
+ position_ids = position_ids,
+ attention_mask = attention_mask,
+ do_prefill = not hasattr(decoder_layer.self_attn, "paged_attention"),
+ )
+ hidden_states += residual
+
+ residual = hidden_states
+ hidden_states = fast_rms_layernorm_inference(decoder_layer.post_attention_layernorm, hidden_states)
+ hidden_states = fast_swiglu_inference(decoder_layer.mlp, hidden_states)
+ hidden_states += residual
+
+ next_decoder_cache.append(present_key_value)
+ pass
+ hidden_states = fast_rms_layernorm_inference(self.model.norm, hidden_states)
+
+ return BaseModelOutputWithPast(
+ last_hidden_state = hidden_states,
+ past_key_values = next_decoder_cache,
+ hidden_states = [],
+ attentions = [],
+ )
+pass
+
+
+def CausalLM_fast_forward(fast_forward_inference):
+ def _CausalLM_fast_forward(
+ self,
+ input_ids: torch.LongTensor = None,
+ causal_mask: Optional[xformers.attn_bias.BlockDiagonalCausalMask] = None,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_values: Optional[List[torch.FloatTensor]] = None,
+ inputs_embeds: Optional[torch.FloatTensor] = None,
+ labels: Optional[torch.LongTensor] = None,
+ use_cache: Optional[bool] = None,
+ output_attentions: Optional[bool] = None,
+ output_hidden_states: Optional[bool] = None,
+ return_dict: Optional[bool] = None,
+ *args, **kwargs,
+ ) -> Union[Tuple, CausalLMOutputWithPast]:
+
+ if past_key_values is not None:
+ outputs = fast_forward_inference(
+ self,
+ input_ids,
+ past_key_values,
+ position_ids = position_ids,
+ attention_mask = attention_mask,
+ )
+ else:
+ causal_mask = xformers.attn_bias.LowerTriangularMask()
+
+ output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions
+ output_hidden_states = (
+ output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states
+ )
+ return_dict = return_dict if return_dict is not None else self.config.use_return_dict
+
+ # decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn)
+ self.model._has_no_labels = labels is None
+
+ outputs = self.model(
+ input_ids=input_ids,
+ causal_mask=causal_mask,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_values=past_key_values,
+ inputs_embeds=inputs_embeds,
+ use_cache=use_cache,
+ output_attentions=output_attentions,
+ output_hidden_states=output_hidden_states,
+ return_dict=return_dict,
+ )
+ pass
+
+ hidden_states = outputs[0]
+ bsz, q_len, hd = hidden_states.shape
+ lm_head = self.lm_head.weight
+ if bsz == 1 and q_len == 1:
+ logits = torch.mv(lm_head, hidden_states.ravel().to(lm_head.dtype))
+ logits = logits.unsqueeze(0).unsqueeze(0)
+ else:
+ logits = self.lm_head(hidden_states.to(lm_head.dtype))
+ pass
+ logits = logits.to(self.config.torch_dtype)
+
+ loss = None
+ if labels is not None:
+ shift_logits = logits
+ if not hasattr(self, "extra_ignored_labels"):
+ # Fixes https://github.com/unslothai/unsloth/issues/10
+ self.extra_ignored_labels = torch.full((self.max_seq_length, 1), -100, device = "cuda")
+ pass
+
+ shift_labels = torch.hstack((labels[..., 1:], self.extra_ignored_labels[:labels.shape[0]]))
+ loss = fast_cross_entropy_loss(
+ logits = shift_logits,
+ labels = shift_labels,
+ )
+ pass
+
+ if not return_dict:
+ output = (logits,) + outputs[1:]
+ return (loss,) + output if loss is not None else output
+
+ return CausalLMOutputWithPast(
+ loss=loss,
+ logits=logits,
+ past_key_values=outputs.past_key_values,
+ hidden_states=outputs.hidden_states,
+ attentions=outputs.attentions,
+ )
+ pass
+ return _CausalLM_fast_forward
+pass
+
+
+def PeftModelForCausalLM_fast_forward(
+ self,
+ input_ids=None,
+ causal_mask=None,
+ attention_mask=None,
+ inputs_embeds=None,
+ labels=None,
+ output_attentions=None,
+ output_hidden_states=None,
+ return_dict=None,
+ task_ids=None,
+ **kwargs,
+):
+ return self.base_model(
+ input_ids=input_ids,
+ causal_mask=causal_mask,
+ attention_mask=attention_mask,
+ inputs_embeds=inputs_embeds,
+ labels=labels,
+ output_attentions=output_attentions,
+ output_hidden_states=output_hidden_states,
+ return_dict=return_dict,
+ **kwargs,
+ )
+pass
+
+
+# Solves https://github.com/unslothai/unsloth/issues/168
+# Static KV Cache was introduced in 4.38.0, causing training to be much slower.
+# Inferene can now be CUDAGraphed, but we shall retain the old rotary embeddings.
+# https://github.com/huggingface/transformers/pull/27931
+# https://github.com/huggingface/transformers/blob/v4.37.2/src/transformers/models/llama/modeling_llama.py
+class LlamaRotaryEmbedding(torch.nn.Module):
+ # Fixes https://github.com/huggingface/transformers/pull/28837
+ # https://github.com/microsoft/DeepSpeed/issues/4932
+ # The precision of RoPE buffers is not correct, so we cast to int64.
+ def __init__(self, dim, max_position_embeddings=2048, base=10000, device=None):
+ super().__init__()
+ self.dim = dim
+ self.max_position_embeddings = max_position_embeddings
+ self.base = base
+
+ # Build here to make `torch.jit.trace` work.
+ self._set_cos_sin_cache(seq_len=max_position_embeddings, device=device, dtype=torch.get_default_dtype())
+ pass
+
+ def _set_cos_sin_cache(self, seq_len, device, dtype):
+ # Note: on the original Llama codebase, these tensors are created on the target device (and not on CPU) and
+ # in FP32. They are applied (multiplied) in FP32 as well.
+ self.max_seq_len_cached = seq_len
+ inv_freq = 1.0 / (
+ self.base ** (torch.arange(0, self.dim, 2, dtype=torch.int64, device="cpu").float() / self.dim)
+ )
+ t = torch.arange(self.max_seq_len_cached, device="cpu", dtype=torch.int64).float()
+
+ freqs = torch.outer(t, inv_freq)
+ # Different from paper, but it uses a different permutation in order to obtain the same calculation
+ emb = torch.cat((freqs, freqs), dim=-1)
+ self.register_buffer("cos_cached", emb.cos().to(dtype=dtype, device=device, non_blocking=True), persistent=False)
+ self.register_buffer("sin_cached", emb.sin().to(dtype=dtype, device=device, non_blocking=True), persistent=False)
+ pass
+
+ def forward(self, x, position_ids=None, seq_len=None):
+ # x: [bs, num_attention_heads, seq_len, head_size]
+ if seq_len > self.max_seq_len_cached:
+ self._set_cos_sin_cache(seq_len=seq_len, device=x.device, dtype=x.dtype)
+
+ return (
+ self.cos_cached[:seq_len].to(dtype=x.dtype),
+ self.sin_cached[:seq_len].to(dtype=x.dtype),
+ )
+ pass
+pass
+
+
+class LlamaLinearScalingRotaryEmbedding(LlamaRotaryEmbedding):
+ """LlamaRotaryEmbedding extended with linear scaling. Credits to the Reddit user /u/kaiokendev"""
+ # Fixes https://github.com/huggingface/transformers/pull/28837
+ # https://github.com/microsoft/DeepSpeed/issues/4932
+ # The precision of RoPE buffers is not correct, so we cast to int64.
+ def __init__(self, dim, max_position_embeddings=2048, base=10000, device=None, scaling_factor=1.0):
+ self.scaling_factor = scaling_factor
+ super().__init__(dim, max_position_embeddings, base, device)
+ pass
+
+ def _set_cos_sin_cache(self, seq_len, device, dtype):
+ self.max_seq_len_cached = seq_len
+ inv_freq = 1.0 / (
+ self.base ** (torch.arange(0, self.dim, 2, dtype=torch.int64, device="cpu").float() / self.dim)
+ )
+ t = torch.arange(self.max_seq_len_cached, device="cpu", dtype=torch.int64).float()
+ t = t / self.scaling_factor
+
+ freqs = torch.outer(t, inv_freq)
+ # Different from paper, but it uses a different permutation in order to obtain the same calculation
+ emb = torch.cat((freqs, freqs), dim=-1)
+ self.register_buffer("cos_cached", emb.cos().to(dtype=dtype, device=device, non_blocking=True), persistent=False)
+ self.register_buffer("sin_cached", emb.sin().to(dtype=dtype, device=device, non_blocking=True), persistent=False)
+ pass
+pass
+
+
+def _wrap_fast_inference(generate, device_type, dtype):
+ # Wraps inference with bfloat16 / float16
+ @torch.inference_mode
+ def _fast_generate(*args, **kwargs):
+ with torch.autocast(device_type = device_type, dtype = dtype):
+ return generate(*args, **kwargs)
+ return _fast_generate
+pass
+
+
+class FastLlamaModel:
+
+ @staticmethod
+ def pre_patch():
+ LlamaAttention .forward = LlamaAttention_fast_forward
+ LlamaSdpaAttention .forward = LlamaAttention_fast_forward
+ LlamaFlashAttention2.forward = LlamaAttention_fast_forward
+ LlamaDecoderLayer .forward = LlamaDecoderLayer_fast_forward
+ LlamaModel .forward = LlamaModel_fast_forward
+ LlamaForCausalLM .forward = CausalLM_fast_forward(LlamaModel_fast_forward_inference)
+ PeftModelForCausalLM.forward = PeftModelForCausalLM_fast_forward
+
+ # Solves https://github.com/unslothai/unsloth/issues/168
+ # Static KV Cache was introduced in 4.38.0, causing training to be much slower.
+ # Inferene can now be CUDAGraphed, but we shall retain the old rotary embeddings.
+ # https://github.com/huggingface/transformers/pull/27931
+ # https://github.com/huggingface/transformers/blob/v4.37.2/src/transformers/models/llama/modeling_llama.py
+ import transformers.models.llama.modeling_llama
+ transformers.models.llama.modeling_llama.LlamaRotaryEmbedding = LlamaRotaryEmbedding
+ transformers.models.llama.modeling_llama.LlamaLinearScalingRotaryEmbedding = LlamaLinearScalingRotaryEmbedding
+ return
+ pass
+
+
+ @staticmethod
+ def from_pretrained(
+ model_name = "unsloth/llama-2-7b-bnb-4bit",
+ max_seq_length = None,
+ dtype = None,
+ load_in_4bit = True,
+ token = None,
+ device_map = "sequential",
+ rope_scaling = None,
+ fix_tokenizer = True,
+ model_patcher = None,
+ tokenizer_name = None,
+ trust_remote_code = False,
+ **kwargs,
+ ):
+ if token is None and "HF_TOKEN" in os.environ:
+ token = os.environ["HF_TOKEN"]
+
+ if token is None and "HUGGINGFACE_TOKEN" in os.environ:
+ token = os.environ["HUGGINGFACE_TOKEN"]
+
+ if model_patcher is None: model_patcher = FastLlamaModel
+ SUPPORTS_BFLOAT16 = torch.cuda.is_bf16_supported()
+ gpu_stats = torch.cuda.get_device_properties(0)
+ max_memory = round(gpu_stats.total_memory / 1024 / 1024 / 1024, 3)
+
+ statistics = \
+ f"==((====))== Unsloth: Fast {model_patcher.__name__[4:-5]} patching release {__version__}\n"\
+ f" \\\ /| GPU: {gpu_stats.name}. Max memory: {max_memory} GB. Platform = {platform_system}.\n"\
+ f"O^O/ \_/ \\ Pytorch: {torch.__version__}. CUDA = {gpu_stats.major}.{gpu_stats.minor}. CUDA Toolkit = {torch.version.cuda}.\n"\
+ f"\ / Bfloat16 = {str(SUPPORTS_BFLOAT16).upper()}. Xformers = {xformers_version}. FA = {HAS_FLASH_ATTENTION}.\n"\
+ f' "-____-" Free Apache license: http://github.com/unslothai/unsloth'
+ print(statistics)
+ model_patcher.pre_patch()
+ # get_statistics()
+
+ if dtype is None:
+ dtype = torch.float16 if not SUPPORTS_BFLOAT16 else torch.bfloat16
+ elif dtype == torch.bfloat16 and not SUPPORTS_BFLOAT16:
+ logger.warning_once("Device does not support bfloat16. Will change to float16.")
+ dtype = torch.float16
+
+ assert(dtype == torch.float16 or dtype == torch.bfloat16 or dtype == torch.float32)
+
+ # RoPE scaling
+ model_max_seq_length = \
+ AutoConfig.from_pretrained(model_name, token = token).max_position_embeddings
+
+ # If max_seq_length is not specified, use maximum fron config
+ if max_seq_length is None:
+ max_seq_length = model_max_seq_length
+ pass
+
+ if (rope_scaling is None) and (max_seq_length > model_max_seq_length):
+ rope_scaling = max_seq_length / model_max_seq_length
+ logger.warning_once(
+ f"Unsloth: {model_name} can only handle sequence lengths of at most "\
+ f"{model_max_seq_length}.\nBut with kaiokendev's RoPE scaling of "\
+ f"{round(rope_scaling, 3)}, it can be magically be extended to "\
+ f"{max_seq_length}!"
+ )
+ rope_scaling = {"type": "linear", "factor": rope_scaling,}
+ pass
+
+ bnb_config = None
+ if load_in_4bit:
+ bnb_config = BitsAndBytesConfig(
+ load_in_4bit = True,
+ bnb_4bit_use_double_quant = True,
+ bnb_4bit_quant_type = "nf4",
+ bnb_4bit_compute_dtype = dtype,
+ )
+ pass
+
+ # https://huggingface.co/togethercomputer/LLaMA-2-7B-32K/discussions/12
+ # RoPE Scaling's max_position_embeddings must be updated
+ max_position_embeddings = max(max_seq_length, model_max_seq_length)
+ try:
+ model = AutoModelForCausalLM.from_pretrained(
+ model_name,
+ device_map = device_map,
+ torch_dtype = dtype,
+ quantization_config = bnb_config,
+ token = token,
+ rope_scaling = rope_scaling,
+ max_position_embeddings = max_position_embeddings,
+ trust_remote_code = trust_remote_code,
+ **kwargs,
+ )
+ except Exception as error:
+ if "rope_scaling" in str(error):
+ if rope_scaling is not None:
+ raise TypeError("Unsloth: {model_name} does not support rope_scaling.")
+ pass
+
+ # Counteract missing rope_scaling
+ model = AutoModelForCausalLM.from_pretrained(
+ model_name,
+ device_map = device_map,
+ torch_dtype = dtype,
+ quantization_config = bnb_config,
+ token = token,
+ max_position_embeddings = max_position_embeddings,
+ trust_remote_code = trust_remote_code,
+ **kwargs,
+ )
+ else:
+ raise error
+ pass
+ pass
+
+ # Counteract saved tokenizers
+ tokenizer_name = model_name if tokenizer_name is None else tokenizer_name
+ tokenizer = load_correct_tokenizer(
+ tokenizer_name = tokenizer_name,
+ model_max_length = max_position_embeddings,
+ padding_side = "right",
+ token = token,
+ trust_remote_code = trust_remote_code,
+ )
+
+ model, tokenizer = patch_tokenizer(model, tokenizer)
+ model = model_patcher.post_patch(model)
+
+ # Patch up QKV / O and MLP
+ for idx, layer in enumerate(model.model.layers):
+ layer.self_attn.apply_qkv = original_apply_qkv
+ layer.self_attn.apply_o = original_apply_o
+ pass
+
+ # Patch Trainer
+ from transformers.trainer import Trainer
+ try:
+ if Trainer._inner_training_loop.__name__ != "_fast_inner_training_loop":
+ inner_training_loop = inspect.getsource(Trainer._inner_training_loop)
+ Trainer._original_training_loop = inner_training_loop
+ else:
+ inner_training_loop = Trainer._original_training_loop
+ except:
+ raise RuntimeError(
+ "Our OSS was designed for people with few GPU resources to level the playing field.\n"
+ "The OSS Apache 2 license only supports four GPUs - please obtain a commercial license from our website.\n"
+ "We're a 2 person team, so we still have to fund our development costs - thanks!\n"
+ "If you don't, please consider at least sponsoring us through Ko-fi! Appreciate it!",
+ )
+ pass
+
+ import transformers.trainer
+ items_in_trainer = dir(transformers.trainer)
+ good_items = []
+ for item in items_in_trainer:
+ # TODO: Support Deepspeed
+ if item.startswith(("deepspeed", "xm", "met", "smp")): continue
+ if item in inner_training_loop: good_items.append(item)
+ pass
+ exec("from transformers.trainer import (" + ", ".join(x for x in good_items) + ")", globals())
+
+ start = re.search('logger\.info\([\"\'].+?Running training', inner_training_loop).span(0)[0]
+ end = inner_training_loop.find("\n\n", start)
+ original_debug = inner_training_loop[start:end]
+ spaces = re.search('\n([\s\t]{1,})', original_debug).group(0)[1:]
+ front_spaces = re.match('([\s\t]{1,})', inner_training_loop).group(0)
+
+ debug_info = """debug_info = \\
+ f"==((====))== Unsloth - 2x faster free finetuning | Num GPUs = {args.world_size}\\n"\\
+ f" \\\\\\ /| Num examples = {num_examples:,} | Num Epochs = {num_train_epochs:,}\\n"\\
+ f"O^O/ \\_/ \\ Batch size per device = {self._train_batch_size:,} | Gradient Accumulation steps = {args.gradient_accumulation_steps}\\n"\\
+ f"\\ / Total batch size = {total_train_batch_size:,} | Total steps = {max_steps:,}\\n"\\
+ f' "-____-" Number of trainable parameters = {get_model_param_count(model, trainable_only=True):,}'
+ logger.warning_once(debug_info)"""
+
+ debug_info = debug_info.split('\n')
+ debug_info = "\n".join([debug_info[0]] + [spaces + x[8:] for x in debug_info[1:]])
+ inner_training_loop = inner_training_loop.replace(original_debug, debug_info)
+
+ debug_info = """n_total_devices = total_train_batch_size // \\
+ args.gradient_accumulation_steps // self._train_batch_size
+ if n_total_devices > 2:
+ logger.warning_once(
+ "Our OSS was designed for people with few GPU resources to level the playing field.\\n"
+ "The OSS Apache 2 license only supports four GPUs - please obtain a commercial license from our website.\\n"
+ "We're a 2 person team, so we still have to fund our development costs - thanks!\\n"
+ "If you don't, please consider at least sponsoring us through Ko-fi! Appreciate it!",
+ )
+ debug_info ="""
+ debug_info = debug_info.split('\n')
+ debug_info = "\n".join([debug_info[0]] + [spaces + x[8:] for x in debug_info[1:]])
+ inner_training_loop = inner_training_loop.replace("debug_info =", debug_info, 1)
+
+ front_spaces = re.match(r"[\t\s]{1,}", inner_training_loop).group(0)
+ inner_training_loop = re.sub(r"^" + front_spaces, "", inner_training_loop, flags = re.MULTILINE)
+ inner_training_loop = inner_training_loop.replace(
+ "train_dataloader = tpu_spmd_dataloader(train_dataloader)",
+ "raise RuntimeError('Unsloth: TPUs are not yet supported!')"
+ )
+ inner_training_loop = inner_training_loop.replace(
+ "self.accelerator.free_memory()",
+ "self.accelerator.free_memory()\n" + \
+ front_spaces + "if self.is_deepspeed_enabled:"\
+ "raise RuntimeError('Unsloth: Deepspeed is not yet supported!')\n", 1,
+ )
+
+ check_batches = """train_dataloader = self.get_train_dataloader()
+ ga = args.gradient_accumulation_steps
+ bsz = self._train_batch_size
+ total_batches = bsz * ga * args.world_size
+ n_total_devices = total_batches // ga // bsz
+ if n_total_devices > 2:
+ logger.warning_once(
+ "Please consider a commercial license - Unsloth was designed for the GPU Poor.\\n"
+ "The OSS currently works on 4 GPUs - we're a 2 person team, so please help fund\\n"
+ "our development costs by supporting us through Ko-fi or buying a license! Thanks!",
+ )
+ divisor = n_total_devices / 2
+ bsz = self._train_batch_size = max(int(bsz / divisor), 1)
+ if total_batches // ga // bsz > 2:
+ divisor = n_total_devices / 2
+ ga = args.gradient_accumulation_steps = max(int(ga / divisor), 1)"""
+ check_batches = check_batches.split('\n')
+ check_batches = "\n".join([check_batches[0]] + [front_spaces + x[8:] for x in check_batches[1:]])
+ inner_training_loop = inner_training_loop.replace(
+ "train_dataloader = self.get_train_dataloader()",
+ check_batches, 1,
+ )
+ inner_training_loop = inner_training_loop.replace(
+ "_inner_training_loop",
+ "_fast_inner_training_loop", 1,
+ )
+ exec(inner_training_loop, globals())
+
+ Trainer._inner_training_loop = _fast_inner_training_loop
+ inner_training_loop = inner_training_loop.replace(
+ "is_torch_tpu_available()",
+ "False",
+ )
+ if "n_total_devices >" not in inner_training_loop:
+ raise RuntimeError(
+ "Our OSS was designed for people with few GPU resources to level the playing field.\n"
+ "The OSS Apache 2 license only supports four GPUs - please obtain a commercial license from our website.\n"
+ "We're a 2 person team, so we still have to fund our development costs - thanks!\n"
+ "If you don't, please consider at least sponsoring us through Ko-fi! Appreciate it!",
+ )
+ pass
+ inner_training_loop = inner_training_loop.replace(
+ "is_sagemaker_mp_enabled()",
+ "False",
+ )
+ Trainer._inner_training_loop = _fast_inner_training_loop
+
+ # Save max_seq_length
+ model.max_seq_length = max_position_embeddings
+ internal_model = model
+ while hasattr(internal_model, "model"):
+ internal_model.max_seq_length = max_position_embeddings
+ internal_model = internal_model.model
+ pass
+ internal_model.max_seq_length = max_position_embeddings
+
+ # We check the tokenizer first for errors
+ if fix_tokenizer:
+ tokenizer = check_tokenizer(
+ model = model,
+ tokenizer = tokenizer,
+ model_name = model_name,
+ model_max_length = max_position_embeddings,
+ padding_side = "right",
+ token = token,
+ )
+ pass
+ patch_saving_functions(tokenizer)
+
+ # Fix up config for transformers uploading PEFT
+ # Not necessary anymore since we require transformers>=4.37!
+ if False:
+ name = model.config._name_or_path
+ if name.startswith("unsloth/") and name.endswith("-bnb-4bit"):
+ name = name[:len(name) - len("-bnb-4bit")]
+ model.config.update({"_name_or_path" : name})
+ pass
+ pass
+
+ # Log Unsloth version for future fastpaths for inference
+ model.config.update({"unsloth_version" : __version__})
+
+ # Add save modules
+ patch_saving_functions(model)
+
+ # Save tokenizer for inference purposes
+ tokenizer.padding_side = "left" # Force inference
+ internal_model = model
+ while hasattr(internal_model, "model"):
+ internal_model._saved_temp_tokenizer = tokenizer
+ internal_model = internal_model.model
+ pass
+ internal_model._saved_temp_tokenizer = tokenizer
+
+ return model, tokenizer
+ pass
+
+
+ @staticmethod
+ def post_patch(model):
+ # Patch model
+ layers = model.model.layers
+
+ # Torch.compile fails on embedding matrix??
+ # Workaround randomnly fixes it for torch versions < 2.2
+ model.model.embed_tokens = torch.nn.Embedding.from_pretrained(model.model.embed_tokens.weight)
+ model.config.update({"unsloth_version" : __version__})
+
+ # We also do this for the lm_head
+ lm_head = torch.nn.Linear(1, 1, bias = None)
+ del lm_head.weight
+ lm_head.weight = model.lm_head.weight
+ lm_head.in_features = lm_head.weight.shape[1]
+ lm_head.out_features = lm_head.weight.shape[0]
+ model.lm_head = lm_head
+
+ # Also patch all dtypes - BnB seems to not allocate the correct type?
+ # BnB default dtype seems to be float16!
+ correct_dtype = lm_head.weight.dtype
+
+ for name, module in model.named_modules():
+ if isinstance(module, (Bnb_Linear4bit, Peft_Linear4bit)):
+ weight = module.weight
+ quant_state = weight.quant_state
+
+ if type(quant_state) is list:
+ # BnB seems to have float16 as default!
+ module.weight.quant_state[2] = correct_dtype # Cast to correct dtype
+ else:
+ # https://github.com/TimDettmers/bitsandbytes/pull/763/files
+ quant_state.dtype = correct_dtype
+ pass
+ pass
+ # Downcast RoPE embedding to correct data type
+ if (name.endswith("rotary_emb") or hasattr(module, "cos_cached")) \
+ and (module.cos_cached.dtype != correct_dtype):
+
+ module.cos_cached = module.cos_cached.to(correct_dtype)
+ module.sin_cached = module.sin_cached.to(correct_dtype)
+ pass
+ pass
+ pass
+
+ # Clear deleted GPU items
+ import gc
+ for _ in range(3):
+ gc.collect()
+ torch.cuda.empty_cache()
+ return model
+ pass
+
+
+ @staticmethod
+ def get_peft_model(
+ model,
+ r = 16,
+ target_modules = ["q_proj", "k_proj", "v_proj", "o_proj",
+ "gate_proj", "up_proj", "down_proj"],
+ lora_alpha = 16,
+ lora_dropout = 0,
+ bias = "none",
+ layers_to_transform = None,
+ layers_pattern = None,
+ use_gradient_checkpointing = True,
+ random_state = 3407,
+ max_seq_length = 2048, # not used anymore
+ use_rslora = False,
+ modules_to_save = None,
+ init_lora_weights = True,
+ loftq_config = {},
+ **kwargs,
+ ):
+ transformers_set_seed(random_state)
+
+ if isinstance(model, PeftModelForCausalLM):
+ raise TypeError(
+ "Unsloth: Your model already has LoRA adapters. No need to run this again!"
+ )
+ pass
+
+ if loftq_config is None: loftq_config = {}
+
+ import inspect
+ signature = str(inspect.signature(LoraConfig))
+ SUPPORTS_LOFTQ = "loftq_config" in signature
+ SUPPORTS_RSLORA = "use_rslora" in signature
+
+ assert(max_seq_length <= model.max_seq_length)
+
+ if lora_dropout != 0:
+ logger.warning_once(
+ f"Unsloth: Dropout = 0 is supported for fast patching. You are using dropout = {lora_dropout}.\n"\
+ f"Unsloth will patch all other layers, except LoRA matrices, causing a performance hit."
+ )
+ pass
+
+ if bias != "none":
+ logger.warning_once(
+ f"Unsloth: bias = `none` is supported for fast patching. You are using bias = {bias}.\n"\
+ f"Unsloth will patch all other layers, except LoRA matrices, causing a performance hit."
+ )
+ pass
+
+ if not (type(init_lora_weights) is bool or \
+ init_lora_weights == "gaussian" or init_lora_weights == "loftq"):
+ raise ValueError(
+ 'Unsloth: `init_lora_weights` must be either [True, False, "gaussian", "loftq"].'
+ )
+ pass
+
+ if init_lora_weights == "loftq":
+
+ if not SUPPORTS_LOFTQ:
+ import peft
+ raise RuntimeError(
+ f"Unsloth: Your PEFT version of {peft.__version__} does not support LoftQ init.\n"\
+ "Please install PEFT 0.7.2 or higher.\n"\
+ "You can also install from source: `pip install git+https://github.com/huggingface/peft.git"
+ )
+ pass
+
+ if loftq_config == {}:
+ from peft import LoftQConfig
+ logger.warning_once(
+ f"Unsloth: init_lora_weights = `loftq` is set, but `loftq_config` is None.\n"\
+ f"We shall use `loftq_config = LoftQConfig(loftq_bits = 4, loftq_iter = 1)`."
+ )
+ loftq_config = LoftQConfig(loftq_bits = 4, loftq_iter = 1)
+ pass
+
+ if hasattr(model.config, "quantization_config"):
+ raise ValueError(
+ "Unsloth: You are using `loftq` init, yet `load_in_4bit = True` was set.\n"\
+ "Reload your model without any quantization by setting `load_in_4bit = False`."
+ )
+ pass
+ pass
+
+ assert(type(use_rslora) is bool)
+ if use_rslora:
+ if not SUPPORTS_RSLORA:
+ # We manually check for PEFT
+ import peft
+ raise RuntimeError(
+ f"Unsloth: Your PEFT version of {peft.__version__} does not support `use_rslora`.\n"\
+ "Please install PEFT 0.7.2 or higher.\n"\
+ "You can also install from source: `pip install git+https://github.com/huggingface/peft.git"
+ )
+ pass
+ pass
+
+ accepted_modules = frozenset(("q_proj", "k_proj", "v_proj", "o_proj",
+ "gate_proj", "up_proj", "down_proj",),)
+ model.config.update({"unsloth_version" : __version__})
+
+ if type(modules_to_save) is tuple:
+ modules_to_save = list(modules_to_save)
+ pass
+
+ train_lm_head = False
+ train_embed_tokens = False
+ final_modules = []
+ for module in target_modules:
+ if module == "lm_head":
+ logger.warning_once(
+ "Unsloth: `lm_head` should be placed in `modules_to_save` and not `target_modules`. "\
+ "Luckily, we shall do it for you!"
+ )
+ train_lm_head = True
+ if modules_to_save is None: modules_to_save = ["lm_head"]
+ else: modules_to_save.append("lm_head")
+
+ elif module == "embed_tokens":
+ logger.warning_once(
+ "Unsloth: `embed_tokens` should be placed in `modules_to_save` and not `target_modules`. "\
+ "Luckily, we shall do it for you!"
+ )
+ train_embed_tokens = True
+ if modules_to_save is None: modules_to_save = ["embed_tokens"]
+ else: modules_to_save.append("embed_tokens")
+
+ else:
+ assert(module in accepted_modules)
+ final_modules.append(module)
+ pass
+
+ # Check if we added new tokens!
+ if hasattr(model, "_need_to_train_embeddings"):
+ if not train_lm_head or not train_embed_tokens:
+ print(
+ "Unsloth: You added new tokens but did not specify if you wanted to "\
+ "train the lm_head and embed_tokens.\nWe must turn it on for you."
+ )
+ train_lm_head = True
+ train_embed_tokens = True
+
+ if modules_to_save is None: modules_to_save = ["embed_tokens"]
+ else: modules_to_save.append("embed_tokens")
+
+ if modules_to_save is None: modules_to_save = ["lm_head"]
+ else: modules_to_save.append("lm_head")
+ pass
+ pass
+
+ # Check for Llama-3
+ # if hasattr(model._saved_temp_tokenizer, "_using_llama3_template"):
+ # if not train_embed_tokens and not train_lm_head:
+ # raise RuntimeError("")
+
+ # First fix untrained tokens
+ # Wrong - can cause reserved tokens to pop out!!
+ # if train_embed_tokens or train_lm_head:
+ # fix_untrained_tokens(model, eps = 1e-16)
+ # pass
+
+ # Check modules_to_save
+ if modules_to_save is not None:
+ for module in modules_to_save:
+ if module == "lm_head":
+ train_lm_head = True
+ elif module == "embed_tokens":
+ train_embed_tokens = True
+ else:
+ raise TypeError(
+ f"Unsloth: Module = {module} is not allowed. Only 'lm_head' and 'embed_tokens' is allowed."
+ )
+ pass
+ pass
+ if isinstance(modules_to_save, (tuple, list)):
+ modules_to_save = list(set(modules_to_save))
+ pass
+
+ # Get LoRA
+ arguments = dict(
+ r = r,
+ lora_alpha = lora_alpha,
+ target_modules = final_modules,
+ lora_dropout = lora_dropout,
+ bias = bias,
+ task_type = TaskType.CAUSAL_LM,
+ layers_to_transform = layers_to_transform,
+ init_lora_weights = init_lora_weights,
+ loftq_config = loftq_config,
+ use_rslora = use_rslora,
+ modules_to_save = modules_to_save,
+ **kwargs,
+ )
+ if not SUPPORTS_LOFTQ: del arguments["loftq_config"]
+ if not SUPPORTS_RSLORA: del arguments["use_rslora"]
+
+ _saved_temp_tokenizer = model._saved_temp_tokenizer
+
+ lora_config = LoraConfig(**arguments)
+ model = _get_peft_model(model, lora_config)
+
+ model._saved_temp_tokenizer = _saved_temp_tokenizer
+
+ model = FastLlamaModel.patch_peft_model(model, use_gradient_checkpointing)
+
+ # Now patch lm_head and embed_tokens
+ if train_embed_tokens:
+ print("Unsloth: Casting embed_tokens to float32")
+ assert(hasattr(model.model.model.embed_tokens, "modules_to_save"))
+ model.model.model.embed_tokens.modules_to_save.default.to(torch.float32)
+ model.model.model.embed_tokens.modules_to_save.default.requires_grad_(True)
+ pass
+
+ if train_lm_head:
+ print("Unsloth: Casting lm_head to float32")
+ assert(hasattr(model.model.lm_head, "modules_to_save"))
+ model.model.lm_head.modules_to_save.default.to(torch.float32)
+ model.model.lm_head.modules_to_save.default.requires_grad_(True)
+ pass
+
+ # Patch tokenizer to pad to the right
+ internal_model = model
+ while hasattr(internal_model, "model"):
+ if hasattr(internal_model, "_saved_temp_tokenizer"):
+ internal_model._saved_temp_tokenizer.padding_side = "right"
+ pass
+ internal_model = internal_model.model
+ pass
+ if hasattr(internal_model, "_saved_temp_tokenizer"):
+ internal_model._saved_temp_tokenizer.padding_side = "right"
+ pass
+
+ return model
+ pass
+
+
+ @staticmethod
+ def patch_peft_model(
+ model,
+ use_gradient_checkpointing = True,
+ ):
+ if not isinstance(model, PeftModelForCausalLM):
+ raise TypeError(
+ "Unsloth: Your model needs to call `.get_peft_model` first!"
+ )
+ pass
+
+ # Get activation function
+ model_type = model.config.model_type
+
+ if model_type == "llama": apply_lora_mlp = apply_lora_mlp_swiglu
+ elif model_type == "mistral": apply_lora_mlp = apply_lora_mlp_swiglu
+ elif model_type == "qwen2": apply_lora_mlp = apply_lora_mlp_swiglu
+ elif model_type == "gemma": apply_lora_mlp = apply_lora_mlp_geglu_approx
+ else:
+ raise NotImplementedError(f"Unsloth: {model_type} is not yet implemented!")
+ pass
+
+ model = prepare_model_for_kbit_training(
+ model,
+ use_gradient_checkpointing = use_gradient_checkpointing,
+ use_reentrant = True,
+ )
+
+ # Fix up config for transformers uploading PEFT
+ for active_adapter in model.peft_config.keys():
+ # Not necessary since we requires transformers >= 4.37
+ if False:
+ name = model.peft_config[active_adapter].base_model_name_or_path
+ if name.startswith("unsloth/") and name.endswith("-bnb-4bit"):
+ name = name[:len(name) - len("-bnb-4bit")]
+ model.peft_config[active_adapter].base_model_name_or_path = name
+ pass
+ # Add revision to enable future fast inference paths
+ model.peft_config[active_adapter].revision = f"unsloth"
+ pass
+
+ from transformers.trainer import Trainer
+ if Trainer._inner_training_loop.__name__ != "_fast_inner_training_loop":
+ raise RuntimeError(
+ "Our OSS was designed for people with few GPU resources to level the playing field.\n"
+ "The OSS Apache 2 license only supports four GPUs - please obtain a commercial license from our website.\n"
+ "We're a 2 person team, so we still have to fund our development costs - thanks!\n"
+ "If you don't, please consider at least sponsoring us through Ko-fi! Appreciate it!",
+ )
+ pass
+
+ # Fix loftq issues
+ # loftq_config must not = None, but rather {}
+ all_configs = model.peft_config
+ for key, current_config in all_configs.items():
+ if hasattr(current_config, "loftq_config") and current_config.loftq_config is None:
+ new_args = current_config.__dict__
+ new_args["loftq_config"] = {}
+ current_config = current_config.__class__(**new_args)
+ all_configs[key] = current_config
+ pass
+ pass
+
+ # Do patching
+ n_mlp = 0
+ n_qkv = 0
+ n_o = 0
+ import types
+
+ active_adapter = model.active_adapters[0] if \
+ hasattr(model, "active_adapters") else model.active_adapter
+
+ # Get dropout and bias
+ lora_dropout = model.peft_config[active_adapter].lora_dropout
+ bias = model.peft_config[active_adapter].bias
+
+ if lora_dropout == 0 and bias == "none":
+ for idx, layer in enumerate(model.model.model.layers):
+
+ # MLP patching
+ gate_proj = layer.mlp.gate_proj
+ up_proj = layer.mlp. up_proj
+ down_proj = layer.mlp.down_proj
+
+ if hasattr(gate_proj, "lora_A") and \
+ hasattr( up_proj, "lora_A") and \
+ hasattr(down_proj, "lora_A") and \
+ (getattr(gate_proj, "base_layer", gate_proj).bias is None) and \
+ (getattr( up_proj, "base_layer", up_proj).bias is None) and \
+ (getattr(down_proj, "base_layer", down_proj).bias is None) and \
+ (getattr(gate_proj, "lora_magnitude_vector", None) is None) and \
+ (getattr( up_proj, "lora_magnitude_vector", None) is None) and \
+ (getattr(down_proj, "lora_magnitude_vector", None) is None):
+
+ # https://stackoverflow.com/questions/50599045/python-replacing-a-function-within-a-class-of-a-module
+ layer.mlp.forward = types.MethodType(apply_lora_mlp, layer.mlp)
+ n_mlp += 1
+ else:
+ logger.warning_once(
+ "Unsloth cannot patch MLP layers with our manual autograd engine since either LoRA adapters\n"\
+ "are not enabled or a bias term (like in Qwen) is used."
+ )
+ pass
+
+ # QKV attention patching
+ q_proj = layer.self_attn.q_proj
+ k_proj = layer.self_attn.k_proj
+ v_proj = layer.self_attn.v_proj
+ if hasattr(q_proj, "lora_A") and \
+ hasattr(k_proj, "lora_A") and \
+ hasattr(v_proj, "lora_A") and \
+ (getattr(q_proj, "base_layer", q_proj).bias is None) and \
+ (getattr(q_proj, "base_layer", k_proj).bias is None) and \
+ (getattr(q_proj, "base_layer", v_proj).bias is None) and \
+ (getattr(q_proj, "lora_magnitude_vector", None) is None) and \
+ (getattr(k_proj, "lora_magnitude_vector", None) is None) and \
+ (getattr(v_proj, "lora_magnitude_vector", None) is None):
+
+ layer.self_attn.apply_qkv = apply_lora_qkv
+ n_qkv += 1
+ else:
+ logger.warning_once(
+ "Unsloth cannot patch Attention layers with our manual autograd engine since either LoRA adapters\n"\
+ "are not enabled or a bias term (like in Qwen) is used."
+ )
+ pass
+
+ # O attention patching
+ o_proj = layer.self_attn.o_proj
+ if hasattr(o_proj, "lora_A") and \
+ (getattr(o_proj, "base_layer", o_proj).bias is None) and \
+ (getattr(o_proj, "lora_magnitude_vector", None) is None):
+
+ layer.self_attn.apply_o = apply_lora_o
+ n_o += 1
+ else:
+ logger.warning_once(
+ "Unsloth cannot patch O projection layer with our manual autograd engine since either LoRA adapters\n"\
+ "are not enabled or a bias term (like in Qwen) is used."
+ )
+ pass
+ pass
+ pass
+
+ logger.warning_once(
+ f"Unsloth {__version__} patched {len(model.model.model.layers)} layers with "\
+ f"{n_qkv} QKV layers, {n_o} O layers and {n_mlp} MLP layers.",
+ )
+ patch_saving_functions(model)
+
+ # Patch cross entropy loss labels
+ # Fixes https://github.com/unslothai/unsloth/issues/10
+ max_seq_length = model.max_seq_length
+ extra_ignored_labels = torch.full((max_seq_length, 1), -100, device = "cuda")
+ model.model.extra_ignored_labels = extra_ignored_labels
+ internal_model = model
+ while hasattr(internal_model, "model"):
+ internal_model.max_seq_length = max_seq_length
+ internal_model = internal_model.model
+ pass
+ internal_model.max_seq_length = max_seq_length
+ return model
+ pass
+
+
+ @staticmethod
+ def for_inference(model):
+ internal_model = model
+ internal_model.gradient_checkpointing = False
+ internal_model.training = False
+
+ while hasattr(internal_model, "model"):
+ internal_model = internal_model.model
+ internal_model.gradient_checkpointing = False
+ internal_model.training = False
+ pass
+
+ # Also check if lm_head / embeddings are trained
+ internal_model = model
+ while not hasattr(internal_model, "lm_head"):
+ internal_model = internal_model.model
+ pass
+ lm_head = internal_model.lm_head.weight
+ device_type = lm_head.device.type
+ dtype = model.config.torch_dtype
+
+ if type(dtype) is str:
+ if dtype == "float16": dtype = torch.float16
+ elif dtype == "bfloat16": dtype = torch.bfloat16
+ pass
+
+ # Wrap model.generate
+ model._unwrapped_old_generate = model.generate
+ model.generate = _wrap_fast_inference(model.generate, device_type, dtype)
+
+ # Patch tokenizer to pad to the left
+ internal_model = model
+ while hasattr(internal_model, "model"):
+ if hasattr(internal_model, "_saved_temp_tokenizer"):
+ internal_model._saved_temp_tokenizer.padding_side = "left"
+ pass
+ internal_model = internal_model.model
+ pass
+ if hasattr(internal_model, "_saved_temp_tokenizer"):
+ internal_model._saved_temp_tokenizer.padding_side = "left"
+ pass
+ pass
+
+
+ @staticmethod
+ def for_training(model, use_gradient_checkpointing = True):
+ internal_model = model
+ internal_model.gradient_checkpointing = use_gradient_checkpointing
+ internal_model.training = True
+
+ # Delete all fast inference loras
+ for param in model.parameters():
+ if hasattr(param, "_fast_lora"):
+ del param._fast_lora
+ pass
+
+ while hasattr(internal_model, "model"):
+ internal_model = internal_model.model
+ internal_model.gradient_checkpointing = use_gradient_checkpointing
+ internal_model.training = True
+ pass
+
+ # Also revert model.generate
+ if hasattr(model, "_unwrapped_old_generate"):
+ model.generate = model._unwrapped_old_generate
+ del model._unwrapped_old_generate
+ pass
+
+ # Patch tokenizer to pad to the right
+ internal_model = model
+ while hasattr(internal_model, "model"):
+ if hasattr(internal_model, "_saved_temp_tokenizer"):
+ internal_model._saved_temp_tokenizer.padding_side = "right"
+ pass
+ internal_model = internal_model.model
+ pass
+ if hasattr(internal_model, "_saved_temp_tokenizer"):
+ internal_model._saved_temp_tokenizer.padding_side = "right"
+ pass
+ pass
+pass
+
diff --git a/unsloth/unsloth/models/loader.py b/unsloth/unsloth/models/loader.py
new file mode 100644
index 0000000000000000000000000000000000000000..86a0f5d7bf378df890d0cd0488cd2a441717431a
--- /dev/null
+++ b/unsloth/unsloth/models/loader.py
@@ -0,0 +1,195 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+from .llama import FastLlamaModel, logger
+from .mistral import FastMistralModel
+from .qwen2 import FastQwen2Model
+from transformers import AutoConfig
+from transformers import __version__ as transformers_version
+from peft import PeftConfig, PeftModel
+from .mapper import INT_TO_FLOAT_MAPPER, FLOAT_TO_INT_MAPPER
+import os
+
+# https://github.com/huggingface/transformers/pull/26037 allows 4 bit loading!
+major, minor = transformers_version.split(".")[:2]
+major, minor = int(major), int(minor)
+SUPPORTS_FOURBIT = (major > 4) or (major == 4 and minor >= 37)
+SUPPORTS_GEMMA = (major > 4) or (major == 4 and minor >= 38)
+if SUPPORTS_GEMMA:
+ from .gemma import FastGemmaModel
+del major, minor
+
+
+def _get_model_name(model_name, load_in_4bit = True):
+
+ if not SUPPORTS_FOURBIT and model_name in INT_TO_FLOAT_MAPPER:
+ model_name = INT_TO_FLOAT_MAPPER[model_name]
+ logger.warning_once(
+ f"Unsloth: Your transformers version of {transformers_version} does not support native "\
+ f"4bit loading.\nThe minimum required version is 4.37.\n"\
+ f'Try `pip install --upgrade "transformers>=4.37"`\n'\
+ f"to obtain the latest transformers build, then restart this session.\n"\
+ f"For now, we shall load `{model_name}` instead (still 4bit, just slower downloading)."
+ )
+
+ elif not load_in_4bit and model_name in INT_TO_FLOAT_MAPPER:
+ new_model_name = INT_TO_FLOAT_MAPPER[model_name]
+ logger.warning_once(
+ f"Unsloth: You passed in `{model_name}` which is a 4bit model, yet you set\n"\
+ f"`load_in_4bit = False`. We shall load `{new_model_name}` instead."
+ )
+ model_name = new_model_name
+
+ elif load_in_4bit and SUPPORTS_FOURBIT and model_name in FLOAT_TO_INT_MAPPER:
+ new_model_name = FLOAT_TO_INT_MAPPER[model_name]
+ logger.warning_once(
+ f"Unsloth: You passed in `{model_name}` and `load_in_4bit = True`.\n"\
+ f"We shall load `{new_model_name}` for 4x faster loading."
+ )
+ model_name = new_model_name
+ pass
+
+ return model_name
+pass
+
+
+class FastLanguageModel(FastLlamaModel):
+ @staticmethod
+ def from_pretrained(
+ model_name = "unsloth/llama-3-8b-bnb-4bit",
+ max_seq_length = None,
+ dtype = None,
+ load_in_4bit = True,
+ token = None,
+ device_map = "sequential",
+ rope_scaling = None,
+ fix_tokenizer = True,
+ trust_remote_code = False,
+ use_gradient_checkpointing = True,
+ resize_model_vocab = None,
+ *args, **kwargs,
+ ):
+ if token is None and "HF_TOKEN" in os.environ:
+ token = os.environ["HF_TOKEN"]
+
+ if token is None and "HUGGINGFACE_TOKEN" in os.environ:
+ token = os.environ["HUGGINGFACE_TOKEN"]
+
+ old_model_name = model_name
+ model_name = _get_model_name(model_name, load_in_4bit)
+
+ # First check if it's a normal model via AutoConfig
+ is_peft = False
+ try:
+ model_config = AutoConfig.from_pretrained(model_name, token = token)
+ is_peft = False
+ except:
+ try:
+ # Most likely a PEFT model
+ peft_config = PeftConfig.from_pretrained(model_name, token = token)
+ except:
+ raise RuntimeError(f"Unsloth: `{model_name}` is not a full model or a PEFT model.")
+
+ # Check base model again for PEFT
+ model_name = _get_model_name(peft_config.base_model_name_or_path, load_in_4bit)
+ model_config = AutoConfig.from_pretrained(model_name, token = token)
+ is_peft = True
+ pass
+
+ model_type = model_config.model_type
+
+ if model_type == "llama": dispatch_model = FastLlamaModel
+ elif model_type == "mistral": dispatch_model = FastMistralModel
+ elif model_type == "gemma":
+ if not SUPPORTS_GEMMA:
+ raise RuntimeError(
+ f"Unsloth: Your transformers version of {transformers_version} does not support Gemma.\n"\
+ f"The minimum required version is 4.38.\n"\
+ f'Try `pip install --upgrade "transformers>=4.38"`\n'\
+ f"to obtain the latest transformers build, then restart this session."\
+ )
+ dispatch_model = FastGemmaModel
+ elif model_type == "qwen2":
+ dispatch_model = FastQwen2Model
+ else:
+ raise NotImplementedError(
+ f"Unsloth: {model_name} not supported yet!\n"\
+ "Make an issue to https://github.com/unslothai/unsloth!",
+ )
+ pass
+
+ # Check if this is local model since the tokenizer gets overwritten
+ if os.path.exists(os.path.join(old_model_name, "tokenizer_config.json")) and \
+ os.path.exists(os.path.join(old_model_name, "tokenizer.json")) and \
+ os.path.exists(os.path.join(old_model_name, "special_tokens_map.json")):
+
+ tokenizer_name = old_model_name
+ else:
+ tokenizer_name = None
+ pass
+
+ model, tokenizer = dispatch_model.from_pretrained(
+ model_name = model_name,
+ max_seq_length = max_seq_length,
+ dtype = dtype,
+ load_in_4bit = load_in_4bit,
+ token = token,
+ device_map = device_map,
+ rope_scaling = rope_scaling,
+ fix_tokenizer = fix_tokenizer,
+ model_patcher = dispatch_model,
+ tokenizer_name = tokenizer_name,
+ trust_remote_code = trust_remote_code,
+ *args, **kwargs,
+ )
+
+ if resize_model_vocab is not None:
+ model.resize_token_embeddings(resize_model_vocab)
+
+ # In case the model supports tagging, add the unsloth tag.
+ if hasattr(model, "add_model_tags"):
+ model.add_model_tags(["unsloth",])
+ pass
+ if hasattr(tokenizer, "add_model_tags"):
+ tokenizer.add_model_tags(["unsloth",])
+ pass
+
+ if load_in_4bit:
+ # Fix up bitsandbytes config
+ quantization_config = \
+ {
+ # Sometimes torch_dtype is not a string!!
+ "bnb_4bit_compute_dtype" : model.config.to_dict()["torch_dtype"],
+ "bnb_4bit_quant_type" : "nf4",
+ "bnb_4bit_use_double_quant" : True,
+ "llm_int8_enable_fp32_cpu_offload" : False,
+ "llm_int8_has_fp16_weight" : False,
+ "llm_int8_skip_modules" : None,
+ "llm_int8_threshold" : 6.0,
+ "load_in_4bit" : True,
+ "load_in_8bit" : False,
+ "quant_method" : "bitsandbytes",
+ }
+ model.config.update({"quantization_config" : quantization_config})
+ pass
+
+ if is_peft:
+ # Now add PEFT adapters
+ model = PeftModel.from_pretrained(model, old_model_name, token = token)
+ # Patch it as well!
+ model = dispatch_model.patch_peft_model(model, use_gradient_checkpointing)
+ pass
+ return model, tokenizer
+ pass
+pass
diff --git a/unsloth/unsloth/models/mapper.py b/unsloth/unsloth/models/mapper.py
new file mode 100644
index 0000000000000000000000000000000000000000..b4fbe573872d23e1a6c4a70a4764fe3c4d53b0e9
--- /dev/null
+++ b/unsloth/unsloth/models/mapper.py
@@ -0,0 +1,158 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+__all__ = [
+ "INT_TO_FLOAT_MAPPER",
+ "FLOAT_TO_INT_MAPPER",
+]
+
+__INT_TO_FLOAT_MAPPER = \
+{
+ "unsloth/mistral-7b-bnb-4bit" : (
+ "unsloth/mistral-7b",
+ "mistralai/Mistral-7B-v0.1",
+ ),
+ "unsloth/llama-2-7b-bnb-4bit" : (
+ "unsloth/llama-2-7b",
+ "meta-llama/Llama-2-7b-hf",
+ ),
+ "unsloth/llama-2-13b-bnb-4bit" : (
+ "unsloth/llama-2-13b",
+ "meta-llama/Llama-2-13b-hf",
+ ),
+ "unsloth/codellama-34b-bnb-4bit" : (
+ "codellama/CodeLlama-34b-hf",
+ ),
+ "unsloth/zephyr-sft-bnb-4bit" : (
+ "unsloth/zephyr-sft",
+ "HuggingFaceH4/mistral-7b-sft-beta",
+ ),
+ "unsloth/tinyllama-bnb-4bit" : (
+ "unsloth/tinyllama",
+ "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T",
+ ),
+ "unsloth/tinyllama-chat-bnb-4bit" : (
+ "unsloth/tinyllama-chat",
+ "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
+ ),
+ "unsloth/mistral-7b-instruct-v0.1-bnb-4bit" : (
+ "mistralai/Mistral-7B-Instruct-v0.1",
+ ),
+ "unsloth/mistral-7b-instruct-v0.2-bnb-4bit" : (
+ "mistralai/Mistral-7B-Instruct-v0.2",
+ ),
+ "unsloth/llama-2-7b-chat-bnb-4bit" : (
+ "unsloth/llama-2-7b-chat",
+ "meta-llama/Llama-2-7b-chat-hf",
+ ),
+ "unsloth/llama-2-7b-chat-bnb-4bit" : (
+ "unsloth/llama-2-7b-chat",
+ "meta-llama/Llama-2-7b-chat-hf",
+ ),
+ "unsloth/codellama-7b-bnb-4bit" : (
+ "unsloth/codellama-7b",
+ "codellama/CodeLlama-7b-hf",
+ ),
+ "unsloth/codellama-13b-bnb-4bit" : (
+ "codellama/CodeLlama-13b-hf",
+ ),
+ "unsloth/yi-6b-bnb-4bit" : (
+ "unsloth/yi-6b",
+ "01-ai/Yi-6B",
+ ),
+ "unsloth/solar-10.7b-bnb-4bit" : (
+ "upstage/SOLAR-10.7B-v1.0",
+ ),
+ "unsloth/gemma-7b-bnb-4bit" : (
+ "unsloth/gemma-7b",
+ "google/gemma-7b",
+ ),
+ "unsloth/gemma-2b-bnb-4bit" : (
+ "unsloth/gemma-2b",
+ "google/gemma-2b",
+ ),
+ "unsloth/gemma-7b-it-bnb-4bit" : (
+ "unsloth/gemma-7b-it",
+ "google/gemma-7b-it",
+ ),
+ "unsloth/gemma-2b-bnb-4bit" : (
+ "unsloth/gemma-2b-it",
+ "google/gemma-2b-it",
+ ),
+ "unsloth/mistral-7b-v0.2-bnb-4bit" : (
+ "unsloth/mistral-7b-v0.2",
+ "alpindale/Mistral-7B-v0.2-hf",
+ ),
+ "unsloth/gemma-1.1-2b-it-bnb-4bit" : (
+ "unsloth/gemma-1.1-2b-it",
+ "google/gemma-1.1-2b-it",
+ ),
+ "unsloth/gemma-1.1-7b-it-bnb-4bit" : (
+ "unsloth/gemma-1.1-7b-it",
+ "google/gemma-1.1-7b-it",
+ ),
+ "unsloth/Starling-LM-7B-beta-bnb-4bit" : (
+ "unsloth/Starling-LM-7B-beta",
+ "Nexusflow/Starling-LM-7B-beta",
+ ),
+ "unsloth/Hermes-2-Pro-Mistral-7B-bnb-4bit" : (
+ "unsloth/Hermes-2-Pro-Mistral-7B",
+ "NousResearch/Hermes-2-Pro-Mistral-7B",
+ ),
+ "unsloth/OpenHermes-2.5-Mistral-7B-bnb-4bit" : (
+ "unsloth/OpenHermes-2.5-Mistral-7B",
+ "teknium/OpenHermes-2.5-Mistral-7B",
+ ),
+ "unsloth/codegemma-2b-bnb-4bit" : (
+ "unsloth/codegemma-2b",
+ "google/codegemma-2b",
+ ),
+ "unsloth/codegemma-7b-bnb-4bit" : (
+ "unsloth/codegemma-7b",
+ "google/codegemma-7b",
+ ),
+ "unsloth/codegemma-7b-it-bnb-4bit" : (
+ "unsloth/codegemma-7b-it",
+ "google/codegemma-7b-it",
+ ),
+ "unsloth/llama-3-8b-bnb-4bit" : (
+ "unsloth/llama-3-8b",
+ "meta-llama/Meta-Llama-3-8B",
+ ),
+ "unsloth/llama-3-8b-Instruct-bnb-4bit" : (
+ "unsloth/llama-3-8b-Instruct",
+ "meta-llama/Meta-Llama-3-8B-Instruct",
+ ),
+ "unsloth/llama-3-70b-bnb-4bit" : (
+ "meta-llama/Meta-Llama-3-70B",
+ ),
+ "unsloth/llama-3-70b-Instruct-bnb-4bit" : (
+ "meta-llama/Meta-Llama-3-70B-Instruct",
+ ),
+ "unsloth/Phi-3-mini-4k-instruct-bnb-4bit" : (
+ "unsloth/Phi-3-mini-4k-instruct",
+ "microsoft/Phi-3-mini-4k-instruct",
+ ),
+}
+
+INT_TO_FLOAT_MAPPER = {}
+FLOAT_TO_INT_MAPPER = {}
+
+for key, values in __INT_TO_FLOAT_MAPPER.items():
+ INT_TO_FLOAT_MAPPER[key] = values[0]
+
+ for value in values:
+ FLOAT_TO_INT_MAPPER[value] = key
+ pass
+pass
diff --git a/unsloth/unsloth/models/mistral.py b/unsloth/unsloth/models/mistral.py
new file mode 100644
index 0000000000000000000000000000000000000000..4594919b38ec258042534592ebab5989478548bf
--- /dev/null
+++ b/unsloth/unsloth/models/mistral.py
@@ -0,0 +1,579 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+from .llama import *
+import os
+from ._utils import __version__
+
+from transformers.models.mistral.modeling_mistral import (
+ MistralAttention,
+ MistralDecoderLayer,
+ MistralModel,
+ MistralForCausalLM,
+)
+# For Pytorch 2.1.1
+try:
+ from transformers.models.mistral.modeling_mistral import (
+ MistralSdpaAttention,
+ MistralFlashAttention2,
+ )
+except:
+ MistralSdpaAttention = MistralAttention
+ MistralFlashAttention2 = MistralAttention
+pass
+
+
+def MistralAttention_fast_forward(
+ self,
+ hidden_states: torch.Tensor,
+ causal_mask: Optional[xformers.attn_bias.BlockDiagonalCausalMask] = None,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_value: Optional[Tuple[torch.Tensor]] = None,
+ output_attentions: bool = False,
+ use_cache: bool = False,
+ padding_mask: Optional[torch.LongTensor] = None,
+ *args, **kwargs,
+) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
+
+ # Clear inference
+ if hasattr(self, "paged_attention"):
+ del self.paged_attention_K
+ del self.paged_attention_V
+ del self.paged_attention
+ del self.temp_QA
+ del self.temp_KV
+ del self.RH_Q
+ del self.attention
+ pass
+
+ bsz, q_len, _ = hidden_states.size()
+
+ n_heads = self.num_heads
+ n_groups = self.num_key_value_groups
+ n_kv_heads = self.num_key_value_heads
+ head_dim = self.head_dim
+ assert(n_kv_heads * n_groups == n_heads)
+
+ Q, K, V = self.apply_qkv(self, hidden_states)
+ Q = Q.view(bsz, q_len, n_heads, head_dim).transpose(1, 2)
+ K = K.view(bsz, q_len, n_kv_heads, head_dim).transpose(1, 2)
+ V = V.view(bsz, q_len, n_kv_heads, head_dim).transpose(1, 2)
+
+ kv_seq_len = K.shape[-2]
+ if past_key_value is not None:
+ kv_seq_len += past_key_value[0].shape[-2]
+
+ if position_ids is None:
+ cos = self.rotary_emb.cos_cached
+ sin = self.rotary_emb.sin_cached
+ Q, K = fast_rope_embedding(Q, K, cos, sin)
+ else:
+ cos, sin = self.rotary_emb(V, seq_len = kv_seq_len)
+ Q, K = inplace_rope_embedding(Q, K, cos, sin, position_ids)
+ pass
+
+ if past_key_value is not None:
+ K = torch.cat([past_key_value[0], K], dim = 2)
+ V = torch.cat([past_key_value[1], V], dim = 2)
+ pass
+ past_key_value = (K, V) if use_cache else None
+
+ # Attention module
+ if (not HAS_FLASH_ATTENTION and attention_mask is None):
+ # Xformers memory efficient attention
+ Q = Q.transpose(1, 2)
+ K = K.transpose(1, 2)
+ V = V.transpose(1, 2)
+ K_M = V_M = bsz * kv_seq_len
+ Q_M = bsz * q_len
+
+ has_swa = isinstance(causal_mask, xformers.attn_bias.BlockDiagonalCausalMask)
+
+ # Group query attention
+ K = K .view(bsz, kv_seq_len, n_kv_heads, 1, head_dim)
+ V = V .view(bsz, kv_seq_len, n_kv_heads, 1, head_dim)
+ K = K.expand(bsz, kv_seq_len, n_kv_heads, n_groups, head_dim)
+ V = V.expand(bsz, kv_seq_len, n_kv_heads, n_groups, head_dim)
+ if hidden_states.requires_grad:
+ K = K.reshape(bsz, kv_seq_len, n_heads, head_dim)
+ V = V.reshape(bsz, kv_seq_len, n_heads, head_dim)
+
+ if has_swa:
+ Q = Q.view(1, Q_M, n_heads, head_dim)
+ K = K.view(1, K_M, n_heads, head_dim)
+ V = V.view(1, V_M, n_heads, head_dim)
+ pass
+ else:
+ # Xformers does support the forward pass though
+ Q = Q.view(bsz, q_len, n_kv_heads, n_groups, head_dim)
+
+ if has_swa:
+ Q = Q.view(1, Q_M, n_kv_heads, n_groups, head_dim)
+ K = K.view(1, K_M, n_kv_heads, n_groups, head_dim)
+ V = V.view(1, V_M, n_kv_heads, n_groups, head_dim)
+ pass
+ pass
+
+ A = xformers_attention(Q, K, V, attn_bias = causal_mask)
+ A = A.view(bsz, q_len, n_heads, head_dim)
+
+ elif HAS_FLASH_ATTENTION and attention_mask is None:
+ Q = Q.transpose(1, 2)
+ K = K.transpose(1, 2)
+ V = V.transpose(1, 2)
+ sw = getattr(self.config, "sliding_window", None)
+ sw = kv_seq_len if (sw is None or sw == "null") else sw
+ window = (-1, -1) if (kv_seq_len <= sw) else (sw, sw)
+ A = flash_attn_func(Q, K, V, causal = True, window_size = window)
+ else:
+ # Grouped query attention
+ # if n_groups != 1:
+ K = K[:, :, None, :, :].expand(bsz, n_kv_heads, n_groups, kv_seq_len, head_dim)
+ V = V[:, :, None, :, :].expand(bsz, n_kv_heads, n_groups, kv_seq_len, head_dim)
+ K = K.reshape(bsz, n_heads, kv_seq_len, head_dim)
+ V = V.reshape(bsz, n_heads, kv_seq_len, head_dim)
+ # pass
+ # Must be contiguous or else results are False!
+ # https://github.com/pytorch/pytorch/issues/112577
+ Q, K, V = Q.contiguous(), K.contiguous(), V.contiguous()
+ # Needs (batch_size, n_heads, seq_len, head_dim)
+ # is_casual and attention_mask must not be both set!
+ A = scaled_dot_product_attention(Q, K, V, attn_mask = attention_mask, is_causal = False)
+ # Go back to (batch_size, seq_len, n_heads, head_dim)
+ A = A.transpose(1, 2).contiguous()
+ pass
+
+ attn_output = A.reshape(bsz, q_len, self.hidden_size)
+ attn_output = self.apply_o(self, attn_output)
+ attn_weights = None
+ return attn_output, attn_weights, past_key_value
+pass
+
+
+def MistralForCausalLM_fast_forward(
+ self,
+ input_ids: torch.LongTensor = None,
+ causal_mask: Optional[xformers.attn_bias.BlockDiagonalCausalMask] = None,
+ attention_mask: Optional[torch.Tensor] = None,
+ position_ids: Optional[torch.LongTensor] = None,
+ past_key_values: Optional[List[torch.FloatTensor]] = None,
+ inputs_embeds: Optional[torch.FloatTensor] = None,
+ labels: Optional[torch.LongTensor] = None,
+ use_cache: Optional[bool] = None,
+ output_attentions: Optional[bool] = None,
+ output_hidden_states: Optional[bool] = None,
+ return_dict: Optional[bool] = None,
+ *args, **kwargs,
+) -> Union[Tuple, CausalLMOutputWithPast]:
+
+ if causal_mask is None and past_key_values is None:
+ bsz, q_len = input_ids.shape
+ sliding_window = getattr(self.config, "sliding_window", None)
+ if sliding_window is None or sliding_window == "null" or sliding_window <= 0:
+ causal_mask = xformers.attn_bias.LowerTriangularMask()
+ elif q_len <= sliding_window:
+ causal_mask = xformers.attn_bias.LowerTriangularMask()
+ else:
+ # Fix from https://github.com/Rypo
+ causal_mask = xformers.attn_bias.BlockDiagonalCausalMask\
+ .from_seqlens([q_len]*bsz)\
+ .make_local_attention(window_size = sliding_window)
+ pass
+
+ output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions
+ output_hidden_states = (
+ output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states
+ )
+ return_dict = return_dict if return_dict is not None else self.config.use_return_dict
+
+ # decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn)
+ self.model._has_no_labels = labels is None
+
+ if past_key_values is not None:
+ outputs = LlamaModel_fast_forward_inference(
+ self,
+ input_ids,
+ past_key_values,
+ position_ids = position_ids,
+ attention_mask = attention_mask,
+ )
+ else:
+ outputs = self.model(
+ input_ids=input_ids,
+ causal_mask=causal_mask,
+ attention_mask=attention_mask,
+ position_ids=position_ids,
+ past_key_values=past_key_values,
+ inputs_embeds=inputs_embeds,
+ use_cache=use_cache,
+ output_attentions=output_attentions,
+ output_hidden_states=output_hidden_states,
+ return_dict=return_dict,
+ )
+ pass
+
+ hidden_states = outputs[0]
+ bsz, q_len, hd = hidden_states.shape
+ lm_head = self.lm_head.weight
+ if bsz == 1 and q_len == 1:
+ logits = torch.mv(lm_head, hidden_states.ravel().to(lm_head.dtype))
+ logits = logits.unsqueeze(0).unsqueeze(0)
+ else:
+ logits = self.lm_head(hidden_states.to(lm_head.dtype))
+ pass
+ logits = logits.to(self.config.torch_dtype)
+
+ loss = None
+ if labels is not None:
+ shift_logits = logits
+ if not hasattr(self, "extra_ignored_labels"):
+ # Fixes https://github.com/unslothai/unsloth/issues/10
+ self.extra_ignored_labels = torch.full((self.max_seq_length, 1), -100, device = "cuda")
+ pass
+
+ shift_labels = torch.hstack((labels[..., 1:], self.extra_ignored_labels[:labels.shape[0]]))
+ loss = fast_cross_entropy_loss(
+ logits = shift_logits,
+ labels = shift_labels,
+ )
+ pass
+
+ if not return_dict:
+ output = (logits,) + outputs[1:]
+ return (loss,) + output if loss is not None else output
+
+ return CausalLMOutputWithPast(
+ loss=loss,
+ logits=logits,
+ past_key_values=outputs.past_key_values,
+ hidden_states=outputs.hidden_states,
+ attentions=outputs.attentions,
+ )
+pass
+
+
+class FastMistralModel(FastLlamaModel):
+
+ @staticmethod
+ def pre_patch():
+ MistralAttention .forward = MistralAttention_fast_forward
+ MistralSdpaAttention .forward = MistralAttention_fast_forward
+ MistralFlashAttention2.forward = MistralAttention_fast_forward
+ MistralDecoderLayer .forward = LlamaDecoderLayer_fast_forward
+ MistralModel .forward = LlamaModel_fast_forward
+ MistralForCausalLM .forward = MistralForCausalLM_fast_forward
+ PeftModelForCausalLM .forward = PeftModelForCausalLM_fast_forward
+
+ # Solves https://github.com/unslothai/unsloth/issues/168
+ # Static KV Cache was introduced in 4.38.0, causing training to be much slower.
+ # Inferene can now be CUDAGraphed, but we shall retain the old rotary embeddings.
+ # https://github.com/huggingface/transformers/pull/27931
+ # https://github.com/huggingface/transformers/blob/v4.37.2/src/transformers/models/llama/modeling_llama.py
+ import transformers.models.mistral.modeling_mistral
+ transformers.models.mistral.modeling_mistral.MistralRotaryEmbedding = LlamaRotaryEmbedding
+ return
+ pass
+
+
+ @staticmethod
+ def from_pretrained(
+ model_name = "unsloth/mistral-7b-bnb-4bit",
+ max_seq_length = None,
+ dtype = None,
+ load_in_4bit = True,
+ token = None,
+ device_map = "sequential",
+ rope_scaling = None, # Mistral does not support RoPE scaling
+ fix_tokenizer = True,
+ model_patcher = None,
+ tokenizer_name = None,
+ trust_remote_code = False,
+ **kwargs,
+ ):
+ if token is None and "HF_TOKEN" in os.environ:
+ token = os.environ["HF_TOKEN"]
+
+ if token is None and "HUGGINGFACE_TOKEN" in os.environ:
+ token = os.environ["HUGGINGFACE_TOKEN"]
+
+ if model_patcher is None: model_patcher = FastMistralModel
+ # Mistral does NOT support RoPE Scaling!
+ if rope_scaling is not None:
+ logger.warning_once("Unsloth: Mistral models do not support RoPE scaling.")
+ pass
+
+ SUPPORTS_BFLOAT16 = torch.cuda.is_bf16_supported()
+ gpu_stats = torch.cuda.get_device_properties(0)
+ max_memory = round(gpu_stats.total_memory / 1024 / 1024 / 1024, 3)
+
+ statistics = \
+ f"==((====))== Unsloth: Fast {model_patcher.__name__[4:-5]} patching release {__version__}\n"\
+ f" \\\ /| GPU: {gpu_stats.name}. Max memory: {max_memory} GB. Platform = {platform_system}.\n"\
+ f"O^O/ \_/ \\ Pytorch: {torch.__version__}. CUDA = {gpu_stats.major}.{gpu_stats.minor}. CUDA Toolkit = {torch.version.cuda}.\n"\
+ f"\ / Bfloat16 = {str(SUPPORTS_BFLOAT16).upper()}. Xformers = {xformers_version}. FA = {HAS_FLASH_ATTENTION}.\n"\
+ f' "-____-" Free Apache license: http://github.com/unslothai/unsloth'
+ print(statistics)
+ model_patcher.pre_patch()
+ # get_statistics()
+
+ if dtype is None:
+ dtype = torch.float16 if not SUPPORTS_BFLOAT16 else torch.bfloat16
+ elif dtype == torch.bfloat16 and not SUPPORTS_BFLOAT16:
+ logger.warning_once("Device does not support bfloat16. Will change to float16.")
+ dtype = torch.float16
+
+ assert(dtype == torch.float16 or dtype == torch.bfloat16 or dtype == torch.float32)
+
+ # Check max sequence length
+ model_config = AutoConfig.from_pretrained(model_name, token = token)
+ model_max_seq_length = model_config.max_position_embeddings
+
+ # If max_seq_length is not specified, use maximum fron config
+ if max_seq_length is None:
+ max_seq_length = model_max_seq_length
+ pass
+
+ # Mistral does NOT support RoPE Scaling sadly so we have to error out.
+ if max_seq_length > model_max_seq_length:
+ raise RuntimeError(
+ f"Unsloth: Unfortunately {model_patcher.__name__[4:-5]} type models do not support RoPE scaling!\n"\
+ f"The maximum sequence length supported is {model_max_seq_length}.",
+ )
+ pass
+
+ bnb_config = None
+ if load_in_4bit:
+ bnb_config = BitsAndBytesConfig(
+ load_in_4bit = True,
+ bnb_4bit_use_double_quant = True,
+ bnb_4bit_quant_type = "nf4",
+ bnb_4bit_compute_dtype = dtype,
+ )
+
+ max_position_embeddings = max(max_seq_length, model_max_seq_length)
+ model = AutoModelForCausalLM.from_pretrained(
+ model_name,
+ device_map = device_map,
+ torch_dtype = dtype,
+ quantization_config = bnb_config,
+ token = token,
+ # rope_scaling = rope_scaling,
+ trust_remote_code = trust_remote_code,
+ **kwargs,
+ )
+
+ # Counteract saved tokenizers
+ tokenizer_name = model_name if tokenizer_name is None else tokenizer_name
+ tokenizer = load_correct_tokenizer(
+ tokenizer_name,
+ model_max_length = max_position_embeddings,
+ padding_side = "right",
+ token = token,
+ trust_remote_code = trust_remote_code,
+ )
+
+ model, tokenizer = patch_tokenizer(model, tokenizer)
+ model = model_patcher.post_patch(model)
+
+ # Patch up QKV / O and MLP
+ for idx, layer in enumerate(model.model.layers):
+ layer.self_attn.apply_qkv = original_apply_qkv
+ layer.self_attn.apply_o = original_apply_o
+ pass
+
+ # Patch Trainer
+ from transformers.trainer import Trainer
+ if Trainer._inner_training_loop.__name__ != "_fast_inner_training_loop":
+ try:
+ inner_training_loop = inspect.getsource(Trainer._inner_training_loop)
+ except:
+ raise RuntimeError(
+ "Our OSS was designed for people with few GPU resources to level the playing field.\n"
+ "The OSS Apache 2 license only supports four GPUs - please obtain a commercial license from our website.\n"
+ "We're a 2 person team, so we still have to fund our development costs - thanks!\n"
+ "If you don't, please consider at least sponsoring us through Ko-fi! Appreciate it!",
+ )
+ pass
+ pass
+
+ # Patch Trainer
+ from transformers.trainer import Trainer
+ try:
+ if Trainer._inner_training_loop.__name__ != "_fast_inner_training_loop":
+ inner_training_loop = inspect.getsource(Trainer._inner_training_loop)
+ Trainer._original_training_loop = inner_training_loop
+ else:
+ inner_training_loop = Trainer._original_training_loop
+ except:
+ raise RuntimeError(
+ "Our OSS was designed for people with few GPU resources to level the playing field.\n"
+ "The OSS Apache 2 license only supports four GPUs - please obtain a commercial license from our website.\n"
+ "We're a 2 person team, so we still have to fund our development costs - thanks!\n"
+ "If you don't, please consider at least sponsoring us through Ko-fi! Appreciate it!",
+ )
+ pass
+
+ import transformers.trainer
+ items_in_trainer = dir(transformers.trainer)
+ good_items = []
+ for item in items_in_trainer:
+ # TODO: Support Deepspeed
+ if item.startswith(("deepspeed", "xm", "met", "smp")): continue
+ if item in inner_training_loop: good_items.append(item)
+ pass
+ exec("from transformers.trainer import (" + ", ".join(x for x in good_items) + ")", globals())
+
+ start = re.search('logger\.info\([\"\'].+?Running training', inner_training_loop).span(0)[0]
+ end = inner_training_loop.find("\n\n", start)
+ original_debug = inner_training_loop[start:end]
+ spaces = re.search('\n([\s\t]{1,})', original_debug).group(0)[1:]
+ front_spaces = re.match('([\s\t]{1,})', inner_training_loop).group(0)
+
+ debug_info = """debug_info = \\
+ f"==((====))== Unsloth - 2x faster free finetuning | Num GPUs = {args.world_size}\\n"\\
+ f" \\\\\\ /| Num examples = {num_examples:,} | Num Epochs = {num_train_epochs:,}\\n"\\
+ f"O^O/ \\_/ \\ Batch size per device = {self._train_batch_size:,} | Gradient Accumulation steps = {args.gradient_accumulation_steps}\\n"\\
+ f"\\ / Total batch size = {total_train_batch_size:,} | Total steps = {max_steps:,}\\n"\\
+ f' "-____-" Number of trainable parameters = {get_model_param_count(model, trainable_only=True):,}'
+ logger.warning_once(debug_info)"""
+
+ debug_info = debug_info.split('\n')
+ debug_info = "\n".join([debug_info[0]] + [spaces + x[8:] for x in debug_info[1:]])
+ inner_training_loop = inner_training_loop.replace(original_debug, debug_info)
+
+ debug_info = """n_total_devices = total_train_batch_size // \\
+ args.gradient_accumulation_steps // self._train_batch_size
+ if n_total_devices > 2:
+ logger.warning_once(
+ "Our OSS was designed for people with few GPU resources to level the playing field.\\n"
+ "The OSS Apache 2 license only supports four GPUs - please obtain a commercial license from our website.\\n"
+ "We're a 2 person team, so we still have to fund our development costs - thanks!\\n"
+ "If you don't, please consider at least sponsoring us through Ko-fi! Appreciate it!",
+ )
+ debug_info ="""
+ debug_info = debug_info.split('\n')
+ debug_info = "\n".join([debug_info[0]] + [spaces + x[8:] for x in debug_info[1:]])
+ inner_training_loop = inner_training_loop.replace("debug_info =", debug_info, 1)
+
+ front_spaces = re.match(r"[\t\s]{1,}", inner_training_loop).group(0)
+ inner_training_loop = re.sub(r"^" + front_spaces, "", inner_training_loop, flags = re.MULTILINE)
+ inner_training_loop = inner_training_loop.replace(
+ "train_dataloader = tpu_spmd_dataloader(train_dataloader)",
+ "raise RuntimeError('Unsloth: TPUs are not yet supported!')"
+ )
+ inner_training_loop = inner_training_loop.replace(
+ "self.accelerator.free_memory()",
+ "self.accelerator.free_memory()\n" + \
+ front_spaces + "if self.is_deepspeed_enabled:"\
+ "raise RuntimeError('Unsloth: Deepspeed is not yet supported!')\n", 1,
+ )
+
+ check_batches = """train_dataloader = self.get_train_dataloader()
+ ga = args.gradient_accumulation_steps
+ bsz = self._train_batch_size
+ total_batches = bsz * ga * args.world_size
+ n_total_devices = total_batches // ga // bsz
+ if n_total_devices > 2:
+ logger.warning_once(
+ "Please consider a commercial license - Unsloth was designed for the GPU Poor.\\n"
+ "The OSS currently works on 4 GPUs - we're a 2 person team, so please help fund\\n"
+ "our development costs by supporting us through Ko-fi or buying a license! Thanks!",
+ )
+ divisor = n_total_devices / 2
+ bsz = self._train_batch_size = max(int(bsz / divisor), 1)
+ if total_batches // ga // bsz > 2:
+ divisor = n_total_devices / 2
+ ga = args.gradient_accumulation_steps = max(int(ga / divisor), 1)"""
+ check_batches = check_batches.split('\n')
+ check_batches = "\n".join([check_batches[0]] + [front_spaces + x[8:] for x in check_batches[1:]])
+ inner_training_loop = inner_training_loop.replace(
+ "train_dataloader = self.get_train_dataloader()",
+ check_batches, 1,
+ )
+ inner_training_loop = inner_training_loop.replace(
+ "_inner_training_loop",
+ "_fast_inner_training_loop", 1,
+ )
+ exec(inner_training_loop, globals())
+
+ Trainer._inner_training_loop = _fast_inner_training_loop
+ inner_training_loop = inner_training_loop.replace(
+ "is_torch_tpu_available()",
+ "False",
+ )
+ if "n_total_devices >" not in inner_training_loop:
+ raise RuntimeError(
+ "Our OSS was designed for people with few GPU resources to level the playing field.\n"
+ "The OSS Apache 2 license only supports four GPUs - please obtain a commercial license from our website.\n"
+ "We're a 2 person team, so we still have to fund our development costs - thanks!\n"
+ "If you don't, please consider at least sponsoring us through Ko-fi! Appreciate it!",
+ )
+ pass
+ inner_training_loop = inner_training_loop.replace(
+ "is_sagemaker_mp_enabled()",
+ "False",
+ )
+ Trainer._inner_training_loop = _fast_inner_training_loop
+
+ # Save max_seq_length
+ max_position_embeddings = max(max_seq_length, model.config.max_position_embeddings)
+ model.max_seq_length = max_position_embeddings
+ internal_model = model
+ while hasattr(internal_model, "model"):
+ internal_model.max_seq_length = max_position_embeddings
+ internal_model = internal_model.model
+ pass
+ internal_model.max_seq_length = max_position_embeddings
+
+ # We check the tokenizer first for errors
+ if fix_tokenizer:
+ tokenizer = check_tokenizer(
+ model = model,
+ tokenizer = tokenizer,
+ model_name = model_name,
+ model_max_length = max_position_embeddings,
+ padding_side = "right",
+ token = token,
+ )
+ pass
+ patch_saving_functions(tokenizer)
+
+ # Fix up config for transformers uploading PEFT
+ # Not necessary anymore since we require transformers>=4.37
+ if False:
+ name = model.config._name_or_path
+ if name.startswith("unsloth/") and name.endswith("-bnb-4bit"):
+ name = name[:len(name) - len("-bnb-4bit")]
+ model.config.update({"_name_or_path" : name})
+ pass
+
+ # Log Unsloth version for future fastpaths for inference
+ model.config.update({"unsloth_version" : __version__})
+
+ # Add save modules
+ patch_saving_functions(model)
+
+ # Save tokenizer for inference purposes
+ tokenizer.padding_side = "left" # Force inference
+ internal_model = model
+ while hasattr(internal_model, "model"):
+ internal_model._saved_temp_tokenizer = tokenizer
+ internal_model = internal_model.model
+ pass
+ internal_model._saved_temp_tokenizer = tokenizer
+
+ return model, tokenizer
+ pass
+pass
diff --git a/unsloth/unsloth/models/qwen2.py b/unsloth/unsloth/models/qwen2.py
new file mode 100644
index 0000000000000000000000000000000000000000..76fe31a6d156ec570f04fc15e05bb00fe14d693c
--- /dev/null
+++ b/unsloth/unsloth/models/qwen2.py
@@ -0,0 +1,91 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+from .llama import *
+from .mistral import FastMistralModel
+import os
+from ._utils import __version__
+
+from transformers.models.qwen2.modeling_qwen2 import (
+ Qwen2Attention,
+ Qwen2DecoderLayer,
+ Qwen2Model,
+ Qwen2ForCausalLM,
+)
+# For Pytorch 2.1.1
+try:
+ from transformers.models.qwen2.modeling_qwen2 import (
+ Qwen2SdpaAttention,
+ Qwen2FlashAttention2,
+ )
+except:
+ Qwen2SdpaAttention = Qwen2Attention
+ Qwen2FlashAttention2 = Qwen2Attention
+pass
+
+
+class FastQwen2Model(FastLlamaModel):
+
+ @staticmethod
+ def pre_patch():
+ Qwen2Attention .forward = LlamaAttention_fast_forward
+ Qwen2SdpaAttention .forward = LlamaAttention_fast_forward
+ Qwen2FlashAttention2.forward = LlamaAttention_fast_forward
+ Qwen2DecoderLayer .forward = LlamaDecoderLayer_fast_forward
+ Qwen2Model .forward = LlamaModel_fast_forward
+ Qwen2ForCausalLM .forward = CausalLM_fast_forward(LlamaModel_fast_forward_inference)
+ PeftModelForCausalLM.forward = PeftModelForCausalLM_fast_forward
+
+ # Solves https://github.com/unslothai/unsloth/issues/168
+ # Static KV Cache was introduced in 4.38.0, causing training to be much slower.
+ # Inferene can now be CUDAGraphed, but we shall retain the old rotary embeddings.
+ # https://github.com/huggingface/transformers/pull/27931
+ # https://github.com/huggingface/transformers/blob/v4.37.2/src/transformers/models/llama/modeling_llama.py
+ import transformers.models.qwen2.modeling_qwen2
+ transformers.models.qwen2.modeling_qwen2.Qwen2RotaryEmbedding = LlamaRotaryEmbedding
+ return
+ pass
+
+
+ @staticmethod
+ def from_pretrained(
+ model_name = "Qwen/Qwen1.5-7B",
+ max_seq_length = 4096,
+ dtype = None,
+ load_in_4bit = True,
+ token = None,
+ device_map = "sequential",
+ rope_scaling = None, # Qwen2 does not support RoPE scaling
+ fix_tokenizer = True,
+ model_patcher = None,
+ tokenizer_name = None,
+ trust_remote_code = False,
+ **kwargs,
+ ):
+ return FastMistralModel.from_pretrained(
+ model_name = model_name,
+ max_seq_length = max_seq_length,
+ dtype = dtype,
+ load_in_4bit = load_in_4bit,
+ token = token,
+ device_map = device_map,
+ rope_scaling = rope_scaling,
+ fix_tokenizer = fix_tokenizer,
+ model_patcher = FastQwen2Model,
+ tokenizer_name = tokenizer_name,
+ trust_remote_code = trust_remote_code,
+ **kwargs,
+ )
+ pass
+pass
diff --git a/unsloth/unsloth/save.py b/unsloth/unsloth/save.py
new file mode 100644
index 0000000000000000000000000000000000000000..39b18d0dd95880f173d7e0fe933a2677fcded3d1
--- /dev/null
+++ b/unsloth/unsloth/save.py
@@ -0,0 +1,1618 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+from bitsandbytes.nn import Linear4bit as Bnb_Linear4bit
+from peft.tuners.lora import Linear4bit as Peft_Linear4bit
+from peft.tuners.lora import Linear as Peft_Linear
+from typing import Optional, Callable, Union, List
+import torch
+import os
+import shutil
+import pickle
+import gc
+from transformers.models.llama.modeling_llama import logger
+from .kernels import fast_dequantize, QUANT_STATE, get_lora_parameters
+import subprocess
+import psutil
+import re
+from transformers.models.llama.modeling_llama import logger
+from .tokenizer_utils import fix_sentencepiece_gguf
+
+__all__ = [
+ "print_quantization_methods",
+ "unsloth_save_model",
+ "save_to_gguf",
+ "patch_saving_functions",
+]
+
+# Check environments
+keynames = "\n" + "\n".join(os.environ.keys())
+IS_COLAB_ENVIRONMENT = "\nCOLAB_" in keynames
+IS_KAGGLE_ENVIRONMENT = "\nKAGGLE_" in keynames
+del keynames
+
+# Weights
+LLAMA_WEIGHTS = (
+ "self_attn.q_proj", "self_attn.k_proj", "self_attn.v_proj", "self_attn.o_proj",
+ "mlp.gate_proj", "mlp.up_proj", "mlp.down_proj",
+)
+LLAMA_LAYERNORMS = (
+ "input_layernorm", "post_attention_layernorm",
+)
+
+# https://github.com/ggerganov/llama.cpp/blob/master/examples/quantize/quantize.cpp#L19
+# From https://mlabonne.github.io/blog/posts/Quantize_Llama_2_models_using_ggml.html
+ALLOWED_QUANTS = \
+{
+ "not_quantized" : "Recommended. Fast conversion. Slow inference, big files.",
+ "fast_quantized" : "Recommended. Fast conversion. OK inference, OK file size.",
+ "quantized" : "Recommended. Slow conversion. Fast inference, small files.",
+ "f32" : "Not recommended. Retains 100% accuracy, but super slow and memory hungry.",
+ "f16" : "Fastest conversion + retains 100% accuracy. Slow and memory hungry.",
+ "q8_0" : "Fast conversion. High resource use, but generally acceptable.",
+ "q4_k_m" : "Recommended. Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q4_K",
+ "q5_k_m" : "Recommended. Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q5_K",
+ "q2_k" : "Uses Q4_K for the attention.vw and feed_forward.w2 tensors, Q2_K for the other tensors.",
+ "q3_k_l" : "Uses Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K",
+ "q3_k_m" : "Uses Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K",
+ "q3_k_s" : "Uses Q3_K for all tensors",
+ "q4_0" : "Original quant method, 4-bit.",
+ "q4_1" : "Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models.",
+ "q4_k_s" : "Uses Q4_K for all tensors",
+ "q4_k" : "alias for q4_k_m",
+ "q5_k" : "alias for q5_k_m",
+ "q5_0" : "Higher accuracy, higher resource usage and slower inference.",
+ "q5_1" : "Even higher accuracy, resource usage and slower inference.",
+ "q5_k_s" : "Uses Q5_K for all tensors",
+ "q6_k" : "Uses Q8_K for all tensors",
+ # "iq2_xxs" : "2.06 bpw quantization", # Not supported sadly
+ # "iq2_xs" : "2.31 bpw quantization",
+ # "iq3_xxs" : "3.06 bpw quantization",
+ "q3_k_xs" : "3-bit extra small quantization",
+}
+
+def print_quantization_methods():
+ for key, value in ALLOWED_QUANTS.items():
+ print(f'"{key}" ==> {value}')
+ pass
+pass
+
+
+def check_if_sentencepiece_model(model, temporary_location = "_unsloth_sentencepiece_temp"):
+ if not hasattr(model, "_saved_temp_tokenizer"): return False
+
+ temp_tokenizer = model._saved_temp_tokenizer
+ sentencepiece_model = False
+ file_location = f"{temporary_location}/{temp_tokenizer.name_or_path}"
+ if not os.path.exists(file_location):
+ os.makedirs(file_location)
+ pass
+ temp_tokenizer.save_pretrained(file_location)
+ if os.path.isfile(f"{file_location}/tokenizer.model"):
+ sentencepiece_model = True
+ pass
+ shutil.rmtree(file_location)
+ return sentencepiece_model
+pass
+
+
+def _free_cached_model(model):
+ from huggingface_hub import scan_cache_dir
+ cached_repos = list(scan_cache_dir().repos)
+
+ # Go through every cached repo, and delete the one that matches the model we want to save.
+ # Can save 4GB of disk space - useful for Kaggle systems.
+ for cached_repo in cached_repos:
+ if cached_repo.repo_id == model.config._name_or_path:
+ remove_cache_commit = list(cached_repo.revisions)[0].commit_hash
+ delete_strategy = scan_cache_dir().delete_revisions(remove_cache_commit,)
+
+ logger.warning_once(
+ "Unsloth: Will remove a cached repo with size " + \
+ delete_strategy.expected_freed_size_str,
+ )
+
+ delete_strategy.execute()
+ pass
+ pass
+pass
+
+
+def _merge_lora(layer, name):
+
+ if isinstance(layer, (Bnb_Linear4bit, Peft_Linear4bit, Peft_Linear)):
+ # Is LoRA so we need to merge!
+ W, quant_state, A, B, s = get_lora_parameters(layer)
+ if quant_state is not None:
+ dtype = quant_state.dtype if type(quant_state) is not list else quant_state[2]
+ W = fast_dequantize(W, quant_state)
+ else:
+ dtype = W.dtype
+ W = W.to(torch.float32).t()
+ # W = W.t()
+
+ if A is not None:
+ # sAB = (A.t().to(torch.float32) @ (s * B.t().to(torch.float32)))
+ # W += sAB
+ W.addmm_(A.t().to(torch.float32), B.t().to(torch.float32), alpha = s)
+ # W.addmm_(A.t().to(W.dtype), B.t().to(W.dtype), alpha = s)
+ # if not torch.isfinite(W).all():
+ maximum_element = torch.max(W.min().abs(), W.max())
+ if not torch.isfinite(maximum_element).item():
+ raise ValueError(f"Unsloth: Merge failed.\n{name} has some elements = infinity.")
+ pass
+ W = W.t().to(dtype)
+ else:
+ W = layer.weight
+ return W
+pass
+
+
+def fast_save_pickle(shard, name):
+ # Use this if # CPUs is <= 2
+ print(f"Unsloth: Saving {name}...")
+ torch.save(
+ shard,
+ name,
+ # HIGHEST_PROTOCOL seems to not work with Pytorch!
+ # pickle_module = pickle,
+ # pickle_protocol = pickle.HIGHEST_PROTOCOL,
+ )
+ return
+pass
+
+
+@torch.inference_mode
+def unsloth_save_model(
+ model,
+ tokenizer,
+ save_directory : Union[str, os.PathLike],
+ save_method : str = "lora", # ["lora", "merged_16bit", "merged_4bit"]
+ push_to_hub : bool = False,
+ token : Optional[Union[str, bool]] = None,
+ is_main_process : bool = True,
+ state_dict : Optional[dict] = None,
+ save_function : Callable = torch.save,
+ max_shard_size : Union[int, str] = "5GB",
+ safe_serialization : bool = True,
+ variant : Optional[str] = None,
+ save_peft_format : bool = True,
+
+ # Push to hub
+ use_temp_dir : Optional[bool] = None,
+ commit_message : Optional[str] = "Trained with Unsloth",
+ private : Optional[bool] = None,
+ create_pr : bool = False,
+ revision : str = None,
+ commit_description : str = "Upload model trained with Unsloth 2x faster",
+ tags : List[str] = None,
+
+ # Our functions
+ temporary_location : str = "_unsloth_temporary_saved_buffers",
+ maximum_memory_usage : float = 0.9,
+):
+ if token is None and "HF_TOKEN" in os.environ:
+ token = os.environ["HF_TOKEN"]
+
+ if token is None and "HUGGINGFACE_TOKEN" in os.environ:
+ token = os.environ["HUGGINGFACE_TOKEN"]
+
+ if commit_message is None: commit_message = ""
+ if "Unsloth" not in commit_message:
+ commit_message += " (Trained with Unsloth)"
+ commit_message = commit_message.lstrip()
+
+ if commit_description is None:
+ commit_description = "Upload model trained with Unsloth 2x faster"
+ elif "Unsloth 2x faster" not in commit_description:
+ commit_description += " (Trained with Unsloth 2x faster)"
+ pass
+
+ if save_method == "merged_4bit":
+ raise RuntimeError(
+ "Unsloth: Merging into 4bit will cause your model to lose accuracy if you plan\n"\
+ "to merge to GGUF or others later on. I suggest you to do this as a final step\n"\
+ "if you're planning to do multiple saves.\n"\
+ "If you are certain, change `save_method` to `merged_4bit_forced`."
+ )
+ elif save_method == "merged_4bit_forced":
+ save_method = "merged_4bit"
+ pass
+
+ save_pretrained_settings = dict(locals())
+ for deletion in ("model", "tokenizer", "save_method", "temporary_location", "maximum_memory_usage"):
+ del save_pretrained_settings[deletion]
+ pass
+
+ # First check for a token!
+ if push_to_hub:
+ from huggingface_hub import whoami
+ try:
+ username = whoami(token = token)["name"]
+ except:
+ raise RuntimeError(
+ "Unsloth: Please supply a token!\n"\
+ "Go to https://huggingface.co/settings/tokens"
+ )
+ pass
+ pass
+
+ assert(maximum_memory_usage > 0 and maximum_memory_usage <= 0.95)
+
+ # Clean memory up first
+ for _ in range(3):
+ torch.cuda.empty_cache()
+ gc.collect()
+ pass
+
+ save_method = save_method.lower().replace(" ", "_")
+ if save_method != "lora" and save_method != "merged_16bit" and save_method != "merged_4bit":
+ raise RuntimeError(
+ "Unsloth: You must select one of 3 options when saving models:\n"\
+ '"lora" ==> This is the fastest and easiet. Just saves LoRA modules.\n'\
+ '"merged_16bit" ==> This merges LoRA weights and saves to float16. Needed for llama.cpp / GGUF.\n'\
+ '"merged_4bit" ==> This merges LoRA weights and saves to 4bit. Useful for DPO / inference.'
+ )
+ pass
+
+ if save_method == "merged_4bit":
+
+ print("Unsloth: Merging 4bit and LoRA weights to 4bit...")
+ print("This might take 5 minutes...")
+
+ # Counteract no LoRA adapters!
+ if hasattr(model, "merge_and_unload"):
+ model = model.merge_and_unload()
+ pass
+ print("Done.")
+ pass
+
+ if tags is not None:
+ assert(isinstance(tags, (list, tuple)))
+ tags = list(tags) + ["unsloth",]
+ else:
+ tags = ["unsloth",]
+ pass
+ save_pretrained_settings["tags"] = tags
+
+ if ((save_method == "lora") or (save_method == "merged_4bit")) and push_to_hub:
+ if token is None:
+ raise RuntimeError(
+ "Unsloth: Pushing to HF requires a token. Pass `token = 'hf_....'`\n"\
+ "Go to https://huggingface.co/settings/tokens."
+ )
+ pass
+
+ if save_method == "lora":
+ print("Unsloth: Saving LoRA adapters. Please wait...")
+ elif save_method == "merged_4bit":
+ print("Unsloth: Saving 4bit Bitsandbytes model. Please wait...")
+ pass
+
+ # Update model tag
+ _ = upload_to_huggingface(
+ model, save_directory, token,
+ "finetuned", "trl", file_location = None,
+ old_username = None, private = private,
+ )
+
+ getattr(model, "original_push_to_hub", tokenizer.push_to_hub)\
+ (
+ repo_id = save_directory,
+ use_temp_dir = use_temp_dir,
+ commit_message = commit_message,
+ private = private,
+ token = token,
+ max_shard_size = max_shard_size,
+ create_pr = create_pr,
+ safe_serialization = safe_serialization,
+ revision = revision,
+ commit_description = commit_description,
+ tags = tags,
+ )
+ if tokenizer is not None:
+ # Set padding side to left for inference
+ old_padding_side = tokenizer.padding_side
+ tokenizer.padding_side = "left"
+
+ getattr(tokenizer, "original_push_to_hub", tokenizer.push_to_hub)\
+ (
+ repo_id = save_directory,
+ use_temp_dir = use_temp_dir,
+ commit_message = commit_message,
+ private = private,
+ token = token,
+ max_shard_size = max_shard_size,
+ create_pr = create_pr,
+ safe_serialization = safe_serialization,
+ revision = revision,
+ commit_description = commit_description,
+ tags = tags,
+ )
+
+ # Revert back padding side
+ tokenizer.padding_side = old_padding_side
+ pass
+
+ if hasattr(model, "config"):
+ print(f"Saved {save_method} model to https://huggingface.co/" + save_directory)
+ pass
+ return save_directory, None
+ pass
+
+ # Tokenizer has different saving arguments
+ tokenizer_save_settings = \
+ {
+ "save_directory" : save_pretrained_settings["save_directory"],
+ "legacy_format" : None,
+ "filename_prefix" : None,
+ "push_to_hub" : save_pretrained_settings["push_to_hub"],
+ "private" : save_pretrained_settings["private"],
+ "token" : save_pretrained_settings["token"],
+ }
+
+ # Check if PEFT Model or not - if yes, 3 levels. If not 2 levels.
+ from peft import PeftModelForCausalLM
+ if isinstance(model, PeftModelForCausalLM):
+ internal_model = model.model
+ else:
+ internal_model = model
+ pass
+
+ # Cannot be converted properly!
+ if (save_method == "merged_4bit") or (save_method == "lora") or (
+ not hasattr(model, "model") or \
+ not hasattr(internal_model.model, "layers")
+ ):
+ # Do general saving
+ # Edit save_pretrained_settings
+ # [TODO] _create_repo has errors due to **kwargs getting accepted
+ # commit_description does not seem to work?
+ what_to_delete = ("use_temp_dir", "commit_message", "create_pr", "revision", "commit_description", "tags",) \
+ if save_pretrained_settings["push_to_hub"] is False else \
+ ("use_temp_dir", "create_pr", "revision", "tags", "commit_description",)
+ for deletion in what_to_delete:
+ del save_pretrained_settings[deletion]
+ pass
+ if hasattr(model, "add_model_tags"):
+ model.add_model_tags(["unsloth",])
+
+ # Update model tag
+ if push_to_hub:
+ _ = upload_to_huggingface(
+ model, save_pretrained_settings["save_directory"], token,
+ "finetuned", "trl", file_location = None,
+ old_username = None, private = private,
+ )
+ pass
+
+ if tokenizer is not None:
+ print("Unsloth: Saving tokenizer...", end = "")
+
+ # Set padding side to left for inference
+ old_padding_side = tokenizer.padding_side
+ tokenizer.padding_side = "left"
+
+ tokenizer.save_pretrained(**tokenizer_save_settings)
+
+ # Revert back padding side
+ tokenizer.padding_side = old_padding_side
+
+ print(" Done.")
+ else:
+ print()
+
+ print("Unsloth: Saving model...", end = "")
+ if save_method != "lora": print(" This might take 10 minutes for Llama-7b...", end = "")
+
+ model.save_pretrained(**save_pretrained_settings)
+
+ if push_to_hub and hasattr(model, "config"):
+ print("Saved to https://huggingface.co/" + save_pretrained_settings["save_directory"])
+ pass
+
+ print(" Done.")
+ return save_directory, None
+ pass
+
+ # If push_to_hub, we must remove the .../ part of a repo
+ username = None
+ if push_to_hub and "/" in save_directory:
+
+ # +1 solves absolute path issues
+ username = save_directory[:save_directory.find("/")]
+ new_save_directory = save_directory[save_directory.find("/")+1:]
+
+ logger.warning_once(
+ f"Unsloth: You are pushing to hub, but you passed your HF username = {username}.\n"\
+ f"We shall truncate {save_directory} to {new_save_directory}"
+ )
+
+ save_pretrained_settings["save_directory"] = new_save_directory
+ tokenizer_save_settings ["save_directory"] = new_save_directory
+ save_directory = new_save_directory
+ pass
+
+ print("Unsloth: Merging 4bit and LoRA weights to 16bit...")
+
+ # Determine max RAM usage minus sharding
+ max_ram = psutil.virtual_memory().available
+ sharded_ram_usage = 5 * 1024 * 1024 * 1024
+ if type(max_shard_size) is str:
+ gb_found = re.match("([0-9]{1,})[\s]{0,}GB", max_shard_size, flags = re.IGNORECASE)
+ mb_found = re.match("([0-9]{1,})[\s]{0,}MB", max_shard_size, flags = re.IGNORECASE)
+ if gb_found: sharded_ram_usage = int(gb_found.group(1)) * 1024 * 1024 * 1024
+ elif mb_found: sharded_ram_usage = int(mb_found.group(1)) * 1024 * 1024
+ elif type(max_shard_size) is int:
+ sharded_ram_usage = sharded_ram_usage
+ pass
+
+ # Switch to our fast saving modules if it's a slow PC!
+ n_cpus = psutil.cpu_count(logical = False)
+ if n_cpus is None: n_cpus = psutil.cpu_count()
+ if n_cpus is None: n_cpus = 1
+
+ if safe_serialization is None:
+ safe_serialization = True
+ save_pretrained_settings["safe_serialization"] = safe_serialization
+
+ elif safe_serialization and (n_cpus <= 2):
+ logger.warning_once(
+ f"Unsloth: You have {n_cpus} CPUs. Using `safe_serialization` is 10x slower.\n"\
+ f"We shall switch to Pytorch saving, which will take 3 minutes and not 30 minutes.\n"\
+ f"To force `safe_serialization`, set it to `None` instead.",
+ )
+ safe_serialization = False
+ save_function = fast_save_pickle
+ save_pretrained_settings["safe_serialization"] = safe_serialization
+ save_pretrained_settings["save_function"] = save_function
+ pass
+
+ # Only safe_serialization uses more RAM
+ if safe_serialization:
+ max_ram -= sharded_ram_usage
+ else:
+ max_ram -= sharded_ram_usage*0.25 # Uses much less
+ pass
+
+ max_ram = int(max(0, max_ram) * maximum_memory_usage)
+ print(f"Unsloth: Will use up to "\
+ f"{round(max_ram/1024/1024/1024, 2)} out of "\
+ f"{round(psutil.virtual_memory().total/1024/1024/1024, 2)} RAM for saving.")
+
+ # Max directory for disk saving
+ if not os.path.exists(temporary_location):
+ os.makedirs(temporary_location)
+ pass
+
+ # Check if Kaggle or Colab, since only 20GB of Disk space allowed.
+ if IS_KAGGLE_ENVIRONMENT or IS_COLAB_ENVIRONMENT:
+ # We free up 4GB of space
+ logger.warning_once(
+ "Unsloth: Kaggle/Colab has limited disk space. We need to delete the downloaded\n"\
+ "model which will save 4-16GB of disk space, allowing you to save on Kaggle/Colab."
+ )
+ _free_cached_model(internal_model)
+ pass
+
+ # HF also uses a OrderedDict
+ from collections import OrderedDict
+ state_dict = OrderedDict()
+
+ torch_dtype = internal_model.config.torch_dtype
+ if type(torch_dtype) is str:
+ if torch_dtype == "float16": torch_dtype = torch.float16
+ elif torch_dtype == "bfloat16": torch_dtype = torch.bfloat16
+ pass
+
+ # Check modules to save float32 dtype
+ state_dict["model.embed_tokens.weight"] = internal_model.model.embed_tokens.weight.data.to(torch_dtype)
+
+ max_vram = int(torch.cuda.get_device_properties(0).total_memory * maximum_memory_usage)
+
+ from tqdm import tqdm as ProgressBar
+ for j, layer in enumerate(ProgressBar(internal_model.model.layers)):
+ for item in LLAMA_WEIGHTS:
+ proj = eval(f"layer.{item}")
+ name = f"model.layers.{j}.{item}.weight"
+ W = _merge_lora(proj, name)
+
+ if (torch.cuda.memory_allocated() + W.nbytes) < max_vram:
+ # Save to GPU memory
+ state_dict[name] = W
+ # [TODO] Saving to RAM seems to leak memory???
+ # elif (max_ram - W.nbytes) > 0:
+ # # Save to CPU memory
+ # logger.warning_once(f"We will save to RAM and not VRAM now.")
+ # state_dict[name] = W.to("cpu", non_blocking = True, copy = True)
+ # max_ram = max(max_ram - W.nbytes, 0)
+ else:
+ # Save to Disk
+ logger.warning_once(f"We will save to Disk and not RAM now.")
+ filename = os.path.join(temporary_location, f"{name}.pt")
+ torch.save(W, filename, pickle_module = pickle, pickle_protocol = pickle.HIGHEST_PROTOCOL,)
+ state_dict[name] = torch.load(filename, map_location = "cpu", mmap = True)
+ pass
+ for item in LLAMA_LAYERNORMS:
+ state_dict[f"model.layers.{j}.{item}.weight"] = eval(f"layer.{item}.weight.data")
+ pass
+ pass
+
+ state_dict["model.norm.weight"] = internal_model.model.norm.weight.data
+ # Check for modules_to_save float32 dtype
+
+ # Check for tied weights
+ if internal_model.model.embed_tokens.weight.data_ptr() != internal_model.lm_head.weight.data_ptr():
+ state_dict["lm_head.weight"] = internal_model.lm_head.weight.data.to(torch_dtype)
+ pass
+
+ # All tensors MUST be type torch.Tensor and not torch.nn.parameter.Parameter
+ for key, value in state_dict.items():
+ if hasattr(value, "data"): state_dict[key] = value = value.data
+ if type(value) is not torch.Tensor:
+ logger.warning_once(f"Unsloth: {key} is not a Tensor but a {type(value)}.")
+ pass
+ pass
+
+ # Edit save_pretrained_settings
+ # [TODO] _create_repo has errors due to **kwargs getting accepted
+ save_pretrained_settings["state_dict"] = state_dict
+
+ # commit_description does not seem to work?
+ what_to_delete = ("use_temp_dir", "commit_message", "create_pr", "revision", "commit_description", "tags",) \
+ if not push_to_hub else \
+ ("use_temp_dir", "create_pr", "revision", "tags", "commit_description",)
+ for deletion in what_to_delete:
+ del save_pretrained_settings[deletion]
+ pass
+ if hasattr(model, "add_model_tags"):
+ model.add_model_tags(["unsloth",])
+
+ # Update model tag
+ if push_to_hub:
+ _ = upload_to_huggingface(
+ model, save_pretrained_settings["save_directory"], token,
+ "finetuned", "trl", file_location = None,
+ old_username = username, private = private,
+ )
+ pass
+
+ # First check if we're pushing to an organization!
+ save_directory = save_pretrained_settings["save_directory"]
+
+ if save_pretrained_settings["push_to_hub"]:
+ new_save_directory, new_username = _determine_username(save_directory, username, token)
+
+ if token is not None:
+ from huggingface_hub import whoami
+ actual_username = whoami(token = token)["name"]
+ else:
+ actual_username = username
+ pass
+
+ # Check if pushing to an organization
+ if save_pretrained_settings["push_to_hub"] and (username != actual_username):
+ print(f"Unsloth: Saving to organization with address {new_save_directory}")
+ # We upload everything at the end!
+ tokenizer_save_settings["push_to_hub"] = False
+ tokenizer_save_settings["save_directory"] = new_save_directory
+ pass
+
+ # Save tokenizer
+ if tokenizer is not None:
+ print("Unsloth: Saving tokenizer...", end = "")
+
+ # Set padding side to left for inference
+ old_padding_side = tokenizer.padding_side
+ tokenizer.padding_side = "left"
+
+ tokenizer.save_pretrained(**tokenizer_save_settings)
+
+ # Revert back padding side
+ tokenizer.padding_side = old_padding_side
+
+ print(" Done.")
+ else:
+ print()
+ pass
+
+ print("Unsloth: Saving model... This might take 5 minutes for Llama-7b...")
+
+ # Since merged, edit quantization_config
+ old_config = model.config
+ new_config = model.config.to_dict()
+ if "quantization_config" in new_config:
+ del new_config["quantization_config"]
+ original_model = model
+ new_config = type(model.config).from_dict(new_config)
+ while hasattr(original_model, "model"):
+ original_model = original_model.model
+ original_model.config = new_config
+ model.config = new_config
+
+ # Save!
+
+ # Check if pushing to an organization
+ if save_pretrained_settings["push_to_hub"] and (username != actual_username):
+ print(f"Unsloth: Saving to organization with address {new_save_directory}")
+ # Pushing to organization!
+ # Sadly .save_pretrained doesn't work :(
+ # We first save it via .save_pretrained, then upload manually!
+ save_pretrained_settings["save_directory"] = new_save_directory
+ save_pretrained_settings["push_to_hub"] = False
+ internal_model.save_pretrained(**save_pretrained_settings)
+
+ # Now manually go through each file and upload them manually!
+ filenames = os.listdir(new_save_directory)
+
+ from huggingface_hub import HfApi
+ hf_api = HfApi(token = save_pretrained_settings["token"])
+
+ print("Unsloth: Uploading all files... Please wait...")
+ hf_api.upload_folder(
+ folder_path = new_save_directory,
+ path_in_repo = ".",
+ repo_id = new_save_directory,
+ repo_type = "model",
+ commit_message = "(Trained with Unsloth)",
+ ignore_patterns = "*.md",
+ )
+ else:
+ internal_model.save_pretrained(**save_pretrained_settings)
+ pass
+
+ # Revert config back
+ original_model = model
+ while hasattr(original_model, "model"):
+ original_model = original_model.model
+ original_model.config = old_config
+ model.config = old_config
+ print("Done.")
+
+ if push_to_hub and hasattr(model, "config"):
+ print(f"Saved merged model to https://huggingface.co/{username}/{save_directory.lstrip('/')}")
+ pass
+
+ save_pretrained_settings["state_dict"] = None
+
+ for j, (key, value) in enumerate(state_dict.items()):
+ state_dict[key] = None
+ if j % 10 == 0:
+ torch.cuda.empty_cache()
+ gc.collect()
+ pass
+ pass
+ state_dict = None
+ del state_dict
+ torch.cuda.empty_cache()
+ gc.collect()
+
+ # Remove temporary location
+ import shutil
+ shutil.rmtree(temporary_location)
+
+ for _ in range(3):
+ torch.cuda.empty_cache()
+ gc.collect()
+ return save_directory, username
+pass
+
+
+def install_llama_cpp_clone_non_blocking():
+ full_command = ["git", "clone", "--recursive", "https://github.com/ggerganov/llama.cpp"]
+ run_installer = subprocess.Popen(full_command, stdout = subprocess.DEVNULL, stderr = subprocess.STDOUT)
+ return run_installer
+pass
+
+
+def install_llama_cpp_make_non_blocking():
+ # https://github.com/ggerganov/llama.cpp/issues/7062
+ # Weirdly GPU conversion for GGUF breaks??
+ # env = { **os.environ, "LLAMA_CUDA": "1", }
+ n_jobs = max(int(psutil.cpu_count()*1.5), 1)
+ # Force make clean
+ os.system("make clean -C llama.cpp")
+ full_command = ["make", "all", "-j"+str(n_jobs), "-C", "llama.cpp"]
+
+ # https://github.com/ggerganov/llama.cpp/issues/7062
+ # Weirdly GPU conversion for GGUF breaks??
+ # run_installer = subprocess.Popen(full_command, env = env, stdout = subprocess.DEVNULL, stderr = subprocess.STDOUT)
+ run_installer = subprocess.Popen(full_command, stdout = subprocess.DEVNULL, stderr = subprocess.STDOUT)
+ return run_installer
+pass
+
+
+def install_python_non_blocking(packages = []):
+ full_command = ["pip", "install"] + packages
+ run_installer = subprocess.Popen(full_command, stdout = subprocess.DEVNULL, stderr = subprocess.STDOUT)
+ return run_installer
+pass
+
+
+def install_llama_cpp_old(version = -10):
+ # Download the 10th latest release since the latest might be broken!
+ # FALLBACK mechanism
+ releases = subprocess.check_output(["git", "ls-remote", "--tags", "https://github.com/ggerganov/llama.cpp.git"])
+ releases = releases.decode("utf-8").replace("\t", " ").split("\n")
+ for i, x in enumerate(releases):
+ if "refs/tags/b" not in x: break
+ releases = releases[:i]
+ latest = releases[-1]
+ version = releases[version].split(" ")[0]
+
+ # Check if the llama.cpp exists
+ if os.path.exists("llama.cpp"):
+ print(
+ "**[WARNING]** You have a llama.cpp old directory which is broken.\n"\
+ "Unsloth will DELETE the broken directory and install a new one.\n"\
+ "Press CTRL + C / cancel this if this is wrong. We shall wait 10 seconds.\n"
+ )
+ import time
+ for i in range(10):
+ print(f"**[WARNING]** Deleting llama.cpp directory... {10-i} seconds left.")
+ time.sleep(1)
+ import shutil
+ shutil.rmtree("llama.cpp")
+ pass
+
+ # Clone a specific commit
+ # Also don't use the GPU!
+ commands = [
+ "git clone --recursive https://github.com/ggerganov/llama.cpp",
+ f"cd llama.cpp && git reset --hard {version} && git clean -df",
+ "make clean -C llama.cpp",
+ f"make all -j{psutil.cpu_count()*2} -C llama.cpp",
+ ]
+ for command in commands:
+ with subprocess.Popen(command, shell = True, stdout = subprocess.PIPE, stderr = subprocess.STDOUT, bufsize = 1) as sp:
+ for line in sp.stdout:
+ print(line.decode("utf-8", errors = "replace"), flush = True, end = "")
+ pass
+ pass
+ # Check if successful
+ if not os.path.exists("llama.cpp/quantize"):
+ raise RuntimeError(
+ "Unsloth: llama.cpp GGUF seems to be too buggy to install.\n"\
+ "File a report to llama.cpp's main repo since this is not an Unsloth issue."
+ )
+ pass
+pass
+
+
+def install_llama_cpp_blocking(use_cuda = True):
+ # https://github.com/ggerganov/llama.cpp/issues/7062
+ # Weirdly GPU conversion for GGUF breaks??
+ # use_cuda = "LLAMA_CUDA=1" if use_cuda else ""
+
+ commands = [
+ "git clone --recursive https://github.com/ggerganov/llama.cpp",
+ "make clean -C llama.cpp",
+ # https://github.com/ggerganov/llama.cpp/issues/7062
+ # Weirdly GPU conversion for GGUF breaks??
+ # f"{use_cuda} make all -j{psutil.cpu_count()*2} -C llama.cpp",
+ f"make all -j{psutil.cpu_count()*2} -C llama.cpp",
+ "pip install gguf protobuf",
+ ]
+ if os.path.exists("llama.cpp"): return
+
+ for command in commands:
+ with subprocess.Popen(command, shell = True, stdout = subprocess.PIPE, stderr = subprocess.STDOUT, bufsize = 1) as sp:
+ for line in sp.stdout:
+ print(line.decode("utf-8", errors = "replace"), flush = True, end = "")
+ pass
+ pass
+pass
+
+
+def _fix_gemma_gguf():
+ # Fixes Gemma saving to GGUF to float32 instead of float16!
+ with open("llama.cpp/convert-hf-to-gguf.py", "rb") as file:
+ text = file.read()
+ pass
+
+ gemma_start = text.find(b"class GemmaModel(Model):")
+ if gemma_start == -1: return
+
+ gemma_end = text.find(b"self.gguf_writer.add_tensor(new_name, data)", gemma_start)
+ if gemma_end == -1: return
+
+ gemma_text = text[gemma_start : gemma_end]
+ bad_text = \
+b""" data = data.astype(np.float32)
+
+ # if f16 desired, convert any float32 2-dim weight tensors to float16
+ if self.ftype == 1 and data_dtype == np.float32 and name.endswith(".weight") and n_dims == 2:
+ data = data.astype(np.float16)"""
+ good_text = \
+b""" # if f32 desired, convert any float16 to float32
+ if self.ftype == 0 and data_dtype == np.float16:
+ data = data.astype(np.float32)
+
+ # TODO: Why cant we use these float16 as-is? There should be not reason to store float16 as float32
+ if self.ftype == 1 and data_dtype == np.float16 and n_dims == 1:
+ data = data.astype(np.float32)
+
+ # if f16 desired, convert any float32 2-dim weight tensors to float16
+ if self.ftype == 1 and data_dtype == np.float32 and name.endswith(".weight") and n_dims == 2:
+ data = data.astype(np.float16)"""
+ find_bad = gemma_text.find(bad_text)
+ if find_bad == -1: return
+
+ gemma_text = gemma_text[:find_bad] + good_text + gemma_text[find_bad + len(bad_text):]
+ text = text[:gemma_start] + gemma_text + text[gemma_end:]
+
+ with open("llama.cpp/convert-hf-to-gguf.py", "w+b") as file:
+ file.write(text)
+ pass
+pass
+
+
+def save_to_gguf(
+ model_type : str,
+ is_sentencepiece : bool = False,
+ model_directory : str = "unsloth_finetuned_model",
+ quantization_method : str = "fast_quantized",
+ first_conversion : str = "f16",
+ _run_installer = None, # Non blocking install of llama.cpp
+):
+ # logger.warning(
+ # "NOTICE: llama.cpp GGUF conversion is currently unstable, since llama.cpp is\n"\
+ # "undergoing some major bug fixes as at 5th of May 2024. This is not an Unsloth issue.\n"\
+ # "Please be patient - GGUF saving should still work, but might not work as well."
+ # )
+
+ if quantization_method.startswith("iq2"):
+ raise RuntimeError("Unsloth: Currently iq2 type quantizations aren't supported yet - sorry!")
+
+ # Careful convert.py is only for Llama / Mistral based archs
+ use_fast_convert = False
+ if not is_sentencepiece: use_fast_convert = False # Llama-3
+ elif model_type == "llama": use_fast_convert = True
+ elif model_type == "mistral": use_fast_convert = True
+ pass
+ logger.warning_once(f"Unsloth: Converting {model_type} model. Can use fast conversion = {use_fast_convert}.")
+
+ if quantization_method == "not_quantized": quantization_method = "f16"
+ elif quantization_method == "fast_quantized": quantization_method = "q8_0"
+ elif quantization_method == "quantized": quantization_method = "q4_k_m"
+ elif quantization_method is None: quantization_method = "q8_0"
+ pass
+
+ if quantization_method not in ALLOWED_QUANTS.keys():
+ error = f"Unsloth: Quant method = [{quantization_method}] not supported. Choose from below:\n"
+ for key, value in ALLOWED_QUANTS.items():
+ error += f"[{key}] => {value}\n"
+ raise RuntimeError(error)
+ pass
+
+ print_info = \
+ f"==((====))== Unsloth: Conversion from QLoRA to GGUF information\n"\
+ f" \\\ /| [0] Installing llama.cpp will take 3 minutes.\n"\
+ f"O^O/ \_/ \\ [1] Converting HF to GUUF 16bits will take 3 minutes.\n"\
+ f"\ / [2] Converting GGUF 16bits to {quantization_method} will take 20 minutes.\n"\
+ f' "-____-" In total, you will have to wait around 26 minutes.\n'
+ print(print_info)
+
+ # Check first_conversion format
+ if first_conversion == "f16" : pass
+ elif first_conversion == "f32" : pass
+ elif first_conversion == "q8_0": pass
+ else:
+ raise RuntimeError(
+ f"Unsloth: `first_conversion` can only be one of ['f16', 'f32', 'q8_0'] and not `{first_conversion}`."
+ )
+ pass
+
+ print("Unsloth: [0] Installing llama.cpp. This will take 3 minutes...")
+ if _run_installer is not None:
+ error = _run_installer.wait()
+ else:
+ error = 0
+ install_llama_cpp_blocking()
+ pass
+ # Check if successful. If not install 10th latest release
+ if error != 0 or not os.path.exists("llama.cpp/quantize"):
+ print(f"Unsloth: llama.cpp error code = {error}.")
+ install_llama_cpp_old(-10)
+ pass
+
+ if quantization_method == "f32": first_conversion = "f32"
+ elif quantization_method == "f16": first_conversion = "f16"
+ elif quantization_method == "q8_0": first_conversion = "q8_0"
+ else:
+ # Quantized models must have f16 as the default argument
+ if first_conversion == "f32" : pass
+ elif first_conversion == "f16" : pass
+ elif first_conversion == "q8_0":
+ logger.warning_once(
+ "Unsloth: Using q8_0 for the `first_conversion` will lose a bit of accuracy, "\
+ "but saves disk space!"
+ )
+ # first_conversion = "f16"
+ pass
+ pass
+
+ # Non llama/mistral needs can only use f32 or f16
+ if not use_fast_convert and (first_conversion != "f16" or first_conversion != "f32"):
+ logger.warning_once("Unsloth: We must use f16 for non Llama and Mistral models.")
+ first_conversion = "f16"
+ pass
+
+ n_cpus = psutil.cpu_count()
+ if n_cpus is None: n_cpus = 1
+ n_cpus *= 2
+ # Concurrency from https://rentry.org/llama-cpp-conversions#merging-loras-into-a-model
+
+ final_location = f"./{model_directory}-unsloth.{first_conversion.upper()}.gguf"
+
+ print(f"Unsloth: [1] Converting model at {model_directory} into {first_conversion} GGUF format.\n"\
+ f"The output location will be {final_location}\n"\
+ "This will take 3 minutes...")
+
+ # We first check if tokenizer.model exists in the model_directory
+ if os.path.exists(f"{model_directory}/tokenizer.model"):
+ vocab_type = "spm,hfft,bpe"
+ # Fix Sentencepiece model as well!
+ fix_sentencepiece_gguf(model_directory)
+ else:
+ vocab_type = "bpe"
+ pass
+
+ if use_fast_convert:
+ command = f"python llama.cpp/convert.py {model_directory} "\
+ f"--outfile {final_location} --vocab-type {vocab_type} "\
+ f"--outtype {first_conversion} --concurrency {n_cpus} --pad-vocab"
+ else:
+ # Need to fix convert-hf-to-gguf.py for some models!
+ # _fix_gemma_gguf()
+
+ command = f"python llama.cpp/convert-hf-to-gguf.py {model_directory} "\
+ f"--outfile {final_location} "\
+ f"--outtype {first_conversion}"
+ pass
+
+ with subprocess.Popen(command, shell = True, stdout = subprocess.PIPE, stderr = subprocess.STDOUT, bufsize = 1) as sp:
+ for line in sp.stdout:
+ print(line.decode("utf-8", errors = "replace"), flush = True, end = "")
+ if sp.returncode is not None and sp.returncode != 0:
+ raise subprocess.CalledProcessError(sp.returncode, sp.args)
+ pass
+
+ # Check if quantization succeeded!
+ if not os.path.isfile(final_location):
+ if IS_KAGGLE_ENVIRONMENT:
+ raise RuntimeError(
+ f"Unsloth: Quantization failed for {final_location}\n"\
+ "You are in a Kaggle environment, which might be the reason this is failing.\n"\
+ "Kaggle only provides 20GB of disk space. Merging to 16bit for 7b models use 16GB of space.\n"\
+ "This means using `model.{save_pretrained/push_to_hub}_merged` works, but\n"\
+ "`model.{save_pretrained/push_to_hub}_gguf will use too much disk space.\n"\
+ "I suggest you to save the 16bit model first, then use manual llama.cpp conversion."
+ )
+ else:
+ raise RuntimeError(
+ f"Unsloth: Quantization failed for {final_location}\n"\
+ "You might have to compile llama.cpp yourself, then run this again.\n"\
+ "You do not need to close this Python program. Run the following commands in a new terminal:\n"\
+ "You must run this in the same folder as you're saving your model.\n"\
+ "git clone --recursive https://github.com/ggerganov/llama.cpp\n"\
+ "cd llama.cpp && make clean && make all -j\n"\
+ "Once that's done, redo the quantization."
+ )
+ pass
+ pass
+ print(f"Unsloth: Conversion completed! Output location: {final_location}")
+
+ if quantization_method != first_conversion:
+ old_location = final_location
+ print(f"Unsloth: [2] Converting GGUF 16bit into {quantization_method}. This will take 20 minutes...")
+ final_location = f"./{model_directory}-unsloth.{quantization_method.upper()}.gguf"
+
+ command = f"./llama.cpp/quantize {old_location} "\
+ f"{final_location} {quantization_method} {n_cpus}"
+
+ # quantize uses stderr
+ with subprocess.Popen(command, shell = True, stdout = subprocess.PIPE, stderr = subprocess.STDOUT, bufsize = 1) as sp:
+ for line in sp.stdout:
+ print(line.decode("utf-8", errors = "replace"), flush = True, end = "")
+ if sp.returncode is not None and sp.returncode != 0:
+ raise subprocess.CalledProcessError(sp.returncode, sp.args)
+ pass
+
+ # Check if quantization succeeded!
+ if not os.path.isfile(final_location):
+ if IS_KAGGLE_ENVIRONMENT:
+ raise RuntimeError(
+ f"Unsloth: Quantization failed for {final_location}\n"\
+ "You are in a Kaggle environment, which might be the reason this is failing.\n"\
+ "Kaggle only provides 20GB of disk space. Merging to 16bit for 7b models use 16GB of space.\n"\
+ "This means using `model.{save_pretrained/push_to_hub}_merged` works, but\n"\
+ "`model.{save_pretrained/push_to_hub}_gguf will use too much disk space.\n"\
+ "I suggest you to save the 16bit model first, then use manual llama.cpp conversion."
+ )
+ else:
+ raise RuntimeError(
+ "Unsloth: Quantization failed! You might have to compile llama.cpp yourself, then run this again.\n"\
+ "You do not need to close this Python program. Run the following commands in a new terminal:\n"\
+ "You must run this in the same folder as you're saving your model.\n"\
+ "git clone --recursive https://github.com/ggerganov/llama.cpp\n"\
+ "cd llama.cpp && make clean && make all -j\n"\
+ "Once that's done, redo the quantization."
+ )
+ pass
+ pass
+
+ print(f"Unsloth: Conversion completed! Output location: {final_location}")
+ pass
+
+ return final_location
+pass
+
+
+def unsloth_save_pretrained_merged(
+ self,
+ save_directory : Union[str, os.PathLike],
+ tokenizer = None,
+ save_method : str = "merged_16bit", # ["lora", "merged_16bit", "merged_4bit"]
+ push_to_hub : bool = False,
+ token : Optional[Union[str, bool]] = None,
+ is_main_process : bool = True,
+ state_dict : Optional[dict] = None,
+ save_function : Callable = torch.save,
+ max_shard_size : Union[int, str] = "5GB",
+ safe_serialization : bool = True,
+ variant : Optional[str] = None,
+ save_peft_format : bool = True,
+ tags : List[str] = None,
+ temporary_location : str = "_unsloth_temporary_saved_buffers",
+ maximum_memory_usage : float = 0.75,
+):
+ """
+ Same as .save_pretrained(...) except 4bit weights are auto
+ converted to float16 with as few overhead as possible.
+
+ Choose for `save_method` to be either:
+ 1. `16bit`: Merge LoRA into float16 weights. Useful for GGUF / llama.cpp.
+ 2. `4bit`: Merge LoRA into int4 weights. Useful for DPO / HF inference.
+ 3. `lora`: Save LoRA adapters with no merging. Useful for HF inference.
+ """
+ if tokenizer is None:
+ logger.warning_once(
+ "Unsloth: You're not saving a tokenizer as well?\n"\
+ "You can do it separately via `tokenizer.save_pretrained(...)`"
+ )
+ pass
+
+ arguments = dict(locals())
+ arguments["model"] = self
+ del arguments["self"]
+ unsloth_save_model(**arguments)
+ for _ in range(3):
+ gc.collect()
+pass
+
+
+def unsloth_push_to_hub_merged(
+ self,
+ repo_id : str,
+ tokenizer = None,
+ save_method : str = "merged_16bit", # ["lora", "merged_16bit", "merged_4bit"]
+ use_temp_dir : Optional[bool] = None,
+ commit_message : Optional[str] = "Trained with Unsloth",
+ private : Optional[bool] = None,
+ token : Union[bool, str, None] = None,
+ max_shard_size : Union[int, str, None] = "5GB",
+ create_pr : bool = False,
+ safe_serialization : bool = True,
+ revision : str = None,
+ commit_description : str = "Upload model trained with Unsloth 2x faster",
+ tags : Optional[List[str]] = None,
+ temporary_location : str = "_unsloth_temporary_saved_buffers",
+ maximum_memory_usage : float = 0.75,
+):
+ """
+ Same as .push_to_hub(...) except 4bit weights are auto
+ converted to float16 with as few overhead as possible.
+
+ Choose for `save_method` to be either:
+ 1. `16bit`: Merge LoRA into float16 weights. Useful for GGUF / llama.cpp.
+ 2. `4bit`: Merge LoRA into int4 weights. Useful for DPO / HF inference.
+ 3. `lora`: Save LoRA adapters with no merging. Useful for HF inference.
+ """
+ if tokenizer is None:
+ logger.warning_once(
+ "Unsloth: You're not saving a tokenizer as well?\n"\
+ "You can do it separately via `tokenizer.push_to_hub(...)`"
+ )
+ pass
+
+ arguments = dict(locals())
+ arguments["model"] = self
+ arguments["save_directory"] = repo_id
+ arguments["push_to_hub"] = True
+ del arguments["self"]
+ del arguments["repo_id"]
+ unsloth_save_model(**arguments)
+ for _ in range(3):
+ gc.collect()
+pass
+
+
+MODEL_CARD = \
+"""---
+base_model: {base_model}
+tags:
+- text-generation-inference
+- transformers
+- unsloth
+- {model_type}
+- {extra}
+license: apache-2.0
+language:
+- en
+---
+
+# Uploaded {method} model
+
+- **Developed by:** {username}
+- **License:** apache-2.0
+- **Finetuned from model :** {base_model}
+
+This {model_type} model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
+
+[ ](https://github.com/unslothai/unsloth)
+"""
+
+
+def _determine_username(save_directory, old_username, token):
+ username = ""
+ save_directory = save_directory.lstrip("./")
+ if "/" not in save_directory:
+ from huggingface_hub import whoami
+ try:
+ username = whoami(token = token)["name"]
+ if type(old_username) is str and username != old_username:
+ username = old_username
+ pass
+ save_directory = f"{username}/{save_directory}"
+ except:
+ raise RuntimeError(f"Unsloth: {save_directory} is not a Huggingface directory.")
+ else:
+ username = save_directory.split("/")[0]
+ pass
+ return save_directory, username
+pass
+
+
+def upload_to_huggingface(
+ model,
+ save_directory,
+ token,
+ method,
+ extra = "",
+ file_location = None,
+ old_username = None,
+ private = None,
+):
+ save_directory, username = _determine_username(save_directory, old_username, token)
+
+ from huggingface_hub import create_repo
+ try:
+ create_repo(
+ repo_id = save_directory,
+ token = token,
+ repo_type = "model",
+ exist_ok = False,
+ private = private,
+ )
+
+ # Create model card
+ from huggingface_hub import ModelCard
+ content = MODEL_CARD.format(
+ username = username,
+ base_model = model.config._name_or_path,
+ model_type = model.config.model_type,
+ method = "",
+ extra = extra,
+ )
+ card = ModelCard(content)
+ card.push_to_hub(save_directory, token = token)
+ except:
+ pass
+
+ if file_location is not None:
+ # Now upload file
+ from huggingface_hub import HfApi
+ hf_api = HfApi(token = token)
+
+ if "/" in file_location:
+ uploaded_location = file_location[file_location.rfind("/")+1:]
+ else:
+ uploaded_location = file_location
+ pass
+
+ hf_api.upload_file(
+ path_or_fileobj = file_location,
+ path_in_repo = uploaded_location,
+ repo_id = save_directory,
+ repo_type = "model",
+ commit_message = "(Trained with Unsloth)",
+ )
+
+ # We also upload a config.json file
+ import json
+ with open("_temporary_unsloth_config.json", "w") as file:
+ json.dump({"model_type" : model.config.model_type}, file, indent = 4)
+ pass
+ hf_api.upload_file(
+ path_or_fileobj = "_temporary_unsloth_config.json",
+ path_in_repo = "config.json",
+ repo_id = save_directory,
+ repo_type = "model",
+ commit_message = "(Trained with Unsloth)",
+ )
+ os.remove("_temporary_unsloth_config.json")
+ pass
+ return username
+pass
+
+
+def unsloth_save_pretrained_gguf(
+ self,
+ save_directory : Union[str, os.PathLike],
+ tokenizer = None,
+ quantization_method : str = "fast_quantized",
+ first_conversion : str = "f16",
+ push_to_hub : bool = False,
+ token : Optional[Union[str, bool]] = None,
+ private : Optional[bool] = None,
+ is_main_process : bool = True,
+ state_dict : Optional[dict] = None,
+ save_function : Callable = torch.save,
+ max_shard_size : Union[int, str] = "5GB",
+ safe_serialization : bool = True,
+ variant : Optional[str] = None,
+ save_peft_format : bool = True,
+ tags : List[str] = None,
+ temporary_location : str = "_unsloth_temporary_saved_buffers",
+ maximum_memory_usage : float = 0.85,
+):
+ """
+ Same as .save_pretrained(...) except 4bit weights are auto
+ converted to float16 then converted to GGUF / llama.cpp format.
+
+ Choose for `quantization_method` to be:
+ "not_quantized" : "Recommended. Fast conversion. Slow inference, big files.",
+ "fast_quantized" : "Recommended. Fast conversion. OK inference, OK file size.",
+ "quantized" : "Recommended. Slow conversion. Fast inference, small files.",
+ "f32" : "Not recommended. Retains 100% accuracy, but super slow and memory hungry.",
+ "f16" : "Fastest conversion + retains 100% accuracy. Slow and memory hungry.",
+ "q8_0" : "Fast conversion. High resource use, but generally acceptable.",
+ "q4_k_m" : "Recommended. Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q4_K",
+ "q5_k_m" : "Recommended. Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q5_K",
+ "q2_k" : "Uses Q4_K for the attention.vw and feed_forward.w2 tensors, Q2_K for the other tensors.",
+ "q3_k_l" : "Uses Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K",
+ "q3_k_m" : "Uses Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K",
+ "q3_k_s" : "Uses Q3_K for all tensors",
+ "q4_0" : "Original quant method, 4-bit.",
+ "q4_1" : "Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models.",
+ "q4_k_s" : "Uses Q4_K for all tensors",
+ "q4_k" : "alias for q4_k_m",
+ "q5_k" : "alias for q5_k_m",
+ "q5_0" : "Higher accuracy, higher resource usage and slower inference.",
+ "q5_1" : "Even higher accuracy, resource usage and slower inference.",
+ "q5_k_s" : "Uses Q5_K for all tensors",
+ "q6_k" : "Uses Q8_K for all tensors",
+ "iq2_xxs" : "2.06 bpw quantization",
+ "iq2_xs" : "2.31 bpw quantization",
+ "iq3_xxs" : "3.06 bpw quantization",
+ "q3_k_xs" : "3-bit extra small quantization",
+ """
+ if tokenizer is None:
+ raise ValueError("Unsloth: Saving to GGUF must have a tokenizer.")
+
+ arguments = dict(locals())
+ arguments["model"] = self
+ arguments["tokenizer"] = tokenizer
+ arguments["push_to_hub"] = False # We save ourselves
+ arguments["save_method"] = "merged_16bit" # Must be 16bit
+ del arguments["self"]
+ del arguments["quantization_method"]
+ del arguments["first_conversion"]
+
+ # Non blocking install GGUF first
+ if not os.path.exists("llama.cpp"):
+
+ if IS_KAGGLE_ENVIRONMENT:
+ # Kaggle is weird - no blocking installs, and no CUDA?
+ python_install = install_python_non_blocking(["gguf", "protobuf"])
+ python_install.wait()
+ install_llama_cpp_blocking(use_cuda = False)
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ makefile = None
+ else:
+ git_clone = install_llama_cpp_clone_non_blocking()
+ python_install = install_python_non_blocking(["gguf", "protobuf"])
+ git_clone.wait()
+ makefile = install_llama_cpp_make_non_blocking()
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ python_install.wait()
+ pass
+ else:
+ try:
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ makefile = None
+ except:
+ # Retry by recloning llama.cpp
+ if IS_KAGGLE_ENVIRONMENT:
+ # Kaggle is weird - no blocking installs, and no CUDA?
+ python_install = install_python_non_blocking(["gguf", "protobuf"])
+ python_install.wait()
+ install_llama_cpp_blocking(use_cuda = False)
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ makefile = None
+ else:
+ git_clone = install_llama_cpp_clone_non_blocking()
+ python_install = install_python_non_blocking(["gguf", "protobuf"])
+ git_clone.wait()
+ makefile = install_llama_cpp_make_non_blocking()
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ python_install.wait()
+ pass
+ pass
+ pass
+
+ for _ in range(3):
+ gc.collect()
+
+ model_type = self.config.model_type
+ is_sentencepiece_model = check_if_sentencepiece_model(self)
+ file_location = save_to_gguf(model_type, is_sentencepiece_model,
+ new_save_directory, quantization_method, first_conversion, makefile,
+ )
+
+ if push_to_hub:
+ print("Unsloth: Uploading GGUF to Huggingface Hub...")
+ username = upload_to_huggingface(
+ self, save_directory, token,
+ "GGUF converted", "gguf", file_location, old_username, private,
+ )
+ link = f"{username}/{new_save_directory.lstrip('/.')}" \
+ if username not in new_save_directory else \
+ new_save_directory.lstrip('/.')
+ print(f"Saved GGUF to https://huggingface.co/{link}")
+ pass
+pass
+
+
+def unsloth_push_to_hub_gguf(
+ self,
+ repo_id : str,
+ tokenizer = None,
+ quantization_method : str = "fast_quantized",
+ first_conversion : str = "f16",
+ use_temp_dir : Optional[bool] = None,
+ commit_message : Optional[str] = "Trained with Unsloth",
+ private : Optional[bool] = None,
+ token : Union[bool, str, None] = None,
+ max_shard_size : Union[int, str, None] = "5GB",
+ create_pr : bool = False,
+ safe_serialization : bool = True,
+ revision : str = None,
+ commit_description : str = "Upload model trained with Unsloth 2x faster",
+ tags : Optional[List[str]] = None,
+ temporary_location : str = "_unsloth_temporary_saved_buffers",
+ maximum_memory_usage : float = 0.85,
+):
+ """
+ Same as .push_to_hub(...) except 4bit weights are auto
+ converted to float16 then converted to GGUF / llama.cpp format.
+
+ Choose for `quantization_method` to be:
+ "not_quantized" : "Recommended. Fast conversion. Slow inference, big files.",
+ "fast_quantized" : "Recommended. Fast conversion. OK inference, OK file size.",
+ "quantized" : "Recommended. Slow conversion. Fast inference, small files.",
+ "f32" : "Not recommended. Retains 100% accuracy, but super slow and memory hungry.",
+ "f16" : "Fastest conversion + retains 100% accuracy. Slow and memory hungry.",
+ "q8_0" : "Fast conversion. High resource use, but generally acceptable.",
+ "q4_k_m" : "Recommended. Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q4_K",
+ "q5_k_m" : "Recommended. Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q5_K",
+ "q2_k" : "Uses Q4_K for the attention.vw and feed_forward.w2 tensors, Q2_K for the other tensors.",
+ "q3_k_l" : "Uses Q5_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K",
+ "q3_k_m" : "Uses Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K",
+ "q3_k_s" : "Uses Q3_K for all tensors",
+ "q4_0" : "Original quant method, 4-bit.",
+ "q4_1" : "Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models.",
+ "q4_k_s" : "Uses Q4_K for all tensors",
+ "q5_0" : "Higher accuracy, higher resource usage and slower inference.",
+ "q5_1" : "Even higher accuracy, resource usage and slower inference.",
+ "q5_k_s" : "Uses Q5_K for all tensors",
+ "q6_k" : "Uses Q8_K for all tensors",
+ """
+ if tokenizer is None:
+ raise ValueError("Unsloth: Saving to GGUF must have a tokenizer.")
+
+ arguments = dict(locals())
+ arguments["model"] = self
+ arguments["tokenizer"] = tokenizer
+ arguments["save_directory"] = repo_id
+ arguments["push_to_hub"] = False # We save ourselves
+ arguments["save_method"] = "merged_16bit" # Must be 16bit
+ del arguments["self"]
+ del arguments["repo_id"]
+ del arguments["quantization_method"]
+ del arguments["first_conversion"]
+
+ # Non blocking install GGUF first
+ if not os.path.exists("llama.cpp"):
+
+ if IS_KAGGLE_ENVIRONMENT:
+ # Kaggle is weird - no blocking installs, and no CUDA?
+ python_install = install_python_non_blocking(["gguf", "protobuf"])
+ python_install.wait()
+ install_llama_cpp_blocking(use_cuda = False)
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ makefile = None
+ else:
+ git_clone = install_llama_cpp_clone_non_blocking()
+ python_install = install_python_non_blocking(["gguf", "protobuf"])
+ git_clone.wait()
+ makefile = install_llama_cpp_make_non_blocking()
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ python_install.wait()
+ pass
+ else:
+ try:
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ makefile = None
+ except:
+ # Retry by recloning llama.cpp
+ if IS_KAGGLE_ENVIRONMENT:
+ # Kaggle is weird - no blocking installs, and no CUDA?
+ python_install = install_python_non_blocking(["gguf", "protobuf"])
+ python_install.wait()
+ install_llama_cpp_blocking(use_cuda = False)
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ makefile = None
+ else:
+ git_clone = install_llama_cpp_clone_non_blocking()
+ python_install = install_python_non_blocking(["gguf", "protobuf"])
+ git_clone.wait()
+ makefile = install_llama_cpp_make_non_blocking()
+ new_save_directory, old_username = unsloth_save_model(**arguments)
+ python_install.wait()
+ pass
+ pass
+ pass
+
+ for _ in range(3):
+ gc.collect()
+
+ model_type = self.config.model_type
+ is_sentencepiece_model = check_if_sentencepiece_model(self)
+ file_location = save_to_gguf(model_type, is_sentencepiece_model,
+ new_save_directory, quantization_method, first_conversion, makefile,
+ )
+
+ print("Unsloth: Uploading GGUF to Huggingface Hub...")
+ username = upload_to_huggingface(
+ self, repo_id, token,
+ "GGUF converted", "gguf", file_location, old_username, private,
+ )
+ link = f"{username}/{new_save_directory.lstrip('/.')}" \
+ if username not in new_save_directory else \
+ new_save_directory.lstrip('/.')
+ print(f"Saved GGUF to https://huggingface.co/{link}")
+pass
+
+
+def patch_saving_functions(model):
+ import inspect
+ import re
+ import types
+ from typing import Callable, Optional, Union, List
+
+ # And now re add our saving methods!
+ if model.push_to_hub.__name__ == "unsloth_push_to_hub":
+ original_push_to_hub = model.original_push_to_hub
+ else:
+ original_push_to_hub = model.push_to_hub
+ pass
+
+ signature = str(inspect.signature(original_push_to_hub)).replace("NoneType", "None")
+ signature = signature[1:]
+ signature = re.sub("", "torch.save", signature)
+ docs = original_push_to_hub.__doc__.encode("utf-8").decode("utf-8")
+
+ push_to_hub_text = f'''def unsloth_push_to_hub(self, {signature}:
+ """
+ {docs}
+ """
+ arguments = dict(locals())
+ del arguments["self"]
+ if "tags" in arguments and arguments["tags"] is not None:
+ assert(isinstance(arguments["tags"], (list, tuple)))
+ arguments["tags"] = list(arguments["tags"]) + ["unsloth",]
+ elif "tags" in arguments:
+ arguments["tags"] = ["unsloth",]
+ elif hasattr(self, "add_model_tags"):
+ self.add_model_tags(["unsloth",])
+
+ if "commit_message" in arguments:
+ commit_message = arguments["commit_message"]
+ if commit_message is not None:
+ if not commit_message.endswith(" "): commit_message += " "
+ if "Unsloth" not in commit_message:
+ commit_message += "(Trained with Unsloth)"
+ else:
+ commit_message = "Upload model trained with Unsloth"
+ arguments["commit_message"] = commit_message
+
+ if "commit_description" in arguments:
+ commit_description = arguments["commit_description"]
+ if commit_description is not None:
+ if not commit_description.endswith(" "): commit_description += " "
+ if "Unsloth" not in commit_description:
+ commit_description += "(Trained with Unsloth 2x faster)"
+ else:
+ commit_description = "Upload model trained with Unsloth 2x faster"
+ arguments["commit_description"] = commit_description
+
+ # Update model tag
+ if hasattr(self, "config"):
+ _ = upload_to_huggingface(
+ self, arguments["repo_id"], arguments["token"],
+ "finetuned", "trl", file_location = None,
+ old_username = None, private = arguments["private"],
+ )
+ pass
+
+ try:
+ self.original_push_to_hub(**arguments)
+ except:
+ del arguments["tags"]
+ self.original_push_to_hub(**arguments)
+ pass
+
+ if hasattr(self, "config"):
+ print("Saved model to https://huggingface.co/" + arguments["repo_id"])
+ pass
+ '''
+ exec(push_to_hub_text, globals())
+
+ original_model = model
+ while True:
+
+ if original_model.push_to_hub.__name__ != "unsloth_push_to_hub":
+ original_model.original_push_to_hub = original_model.push_to_hub
+ original_model.push_to_hub = types.MethodType(unsloth_push_to_hub, original_model)
+ if hasattr(original_model, "add_model_tags"):
+ original_model.add_model_tags(["unsloth",])
+ pass
+ pass
+
+ if hasattr(original_model, "model"): original_model = original_model.model
+ else: break
+ pass
+
+ # Add saving methods to top level model
+ if hasattr(model, "config"):
+ # Counteract tokenizers
+ model.push_to_hub_merged = types.MethodType(unsloth_push_to_hub_merged, model)
+ model.save_pretrained_merged = types.MethodType(unsloth_save_pretrained_merged, model)
+ model.push_to_hub_gguf = types.MethodType(unsloth_push_to_hub_gguf, model)
+ model.save_pretrained_gguf = types.MethodType(unsloth_save_pretrained_gguf, model)
+ pass
+ return model
+pass
diff --git a/unsloth/unsloth/tokenizer_utils.py b/unsloth/unsloth/tokenizer_utils.py
new file mode 100644
index 0000000000000000000000000000000000000000..7a1d2ddaccecbaba46e9b019eb8f16083e7e3092
--- /dev/null
+++ b/unsloth/unsloth/tokenizer_utils.py
@@ -0,0 +1,755 @@
+# Copyright 2023-present Daniel Han-Chen & the Unsloth team. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+from transformers import AutoTokenizer
+from transformers.convert_slow_tokenizer import convert_slow_tokenizer
+from transformers import PreTrainedTokenizerFast
+import re
+import os
+from transformers.models.llama.modeling_llama import logger
+from peft import PeftModelForCausalLM
+import torch
+
+__all__ = [
+ "load_correct_tokenizer",
+ "fix_sentencepiece_tokenizer",
+ "check_tokenizer",
+ "add_new_tokens",
+ "fix_sentencepiece_gguf",
+]
+
+
+IGNORED_TOKENIZER_CHECKING = frozenset((
+ "CodeLlamaTokenizerFast",
+ "CodeLlamaTokenizer",
+))
+
+# Check environments
+keynames = "\n" + "\n".join(os.environ.keys())
+IS_COLAB_ENVIRONMENT = "\nCOLAB_" in keynames
+IS_KAGGLE_ENVIRONMENT = "\nKAGGLE_" in keynames
+del keynames
+
+
+def try_fix_tokenizer(tokenizer, prepend = True):
+
+ if hasattr(tokenizer, "_tokenizer"):
+ converted_tokenizer = tokenizer._tokenizer
+ else:
+ converted_tokenizer = convert_slow_tokenizer(tokenizer)
+ pass
+
+ tokenizer_string = converted_tokenizer.to_str()
+
+ # Llama does _apple. Sometimes this is wrong!!
+ prepend_text = '{"type":"Prepend","prepend":"▁"},'
+ if not prepend and prepend_text in tokenizer_string:
+ tokenizer_string = tokenizer_string.replace(prepend_text, "", 1)
+ pass
+
+ dir_names = dir(tokenizer)
+ # Get eos_token, bos_token etc
+ token_names = [x for x in dir_names if x.endswith("_token") and x.count("_") == 1]
+
+ for token_name in token_names:
+ token = getattr(tokenizer, token_name, None)
+ if token is None: continue
+ token_id = getattr(tokenizer, token_name + "_id", None)
+
+ # Locate the token's id mapping in the string
+ find_text = f'"id":{token_id},"content":"'
+ start = tokenizer_string.find(find_text) + len(find_text)
+ if start == -1: continue
+ end = tokenizer_string.find('",', start)
+
+ bad_token = tokenizer_string[start : end]
+ # Check if token is the actual same one - if not, edit it
+ if bad_token != token:
+ bad_text = f'{find_text}{bad_token}",'
+ good_text = f'{find_text}{token}",'
+ tokenizer_string = tokenizer_string.replace(bad_text, good_text, 1)
+
+ # And replace vocab section
+ bad_text = f'"{bad_token}":{token_id},'
+ good_text = f'"{token}":{token_id},'
+ tokenizer_string = tokenizer_string.replace(bad_text, good_text, 1)
+ pass
+ pass
+
+ fixed_tokenizer = converted_tokenizer.from_str(tokenizer_string)
+ return fixed_tokenizer
+pass
+
+
+def get_sorted_dict(dictionary):
+ sorted_keys = sorted(dictionary.values())
+ inverted_dictionary = { value : key for key, value in dictionary.items() }
+
+ sorted_dictionary = {}
+ for key in sorted_keys:
+ value = inverted_dictionary[key]
+ sorted_dictionary[value] = key
+ return sorted_dictionary
+pass
+
+
+def convert_to_fast_tokenizer(
+ slow_tokenizer,
+ temporary_location = "_unsloth_sentencepiece_temp",
+):
+ is_fast = getattr(slow_tokenizer, "is_fast", False)
+ if is_fast: return slow_tokenizer
+
+ try:
+ tokenizer_name = slow_tokenizer.__class__.__name__
+ lowered_tokenizer_name = tokenizer_name.lower()
+ if lowered_tokenizer_name.endswith("tokenizer"):
+ class_name = lowered_tokenizer_name[:-len("tokenizer")]
+ FastTokenizer = eval(
+ f'__import__(f"transformers.models.{class_name}").{tokenizer_name}Fast'
+ )
+ else:
+ FastTokenizer = PreTrainedTokenizerFast
+ except:
+ FastTokenizer = PreTrainedTokenizerFast
+ pass
+
+ # Get all arguments (bos_token, etc)
+ docs = FastTokenizer.__doc__
+ docs = docs[docs.find("Args:"):]
+ args = re.findall(r"\n[\s]+([^\s]{1,}) \(", docs, flags = re.MULTILINE)
+ args = [x for x in args if not x.endswith("_file")]
+
+ # Also some missing maybe!
+ docs = PreTrainedTokenizerFast.__doc__
+ docs = docs[docs.find("Args:"):]
+ args2 = re.findall(r"\n[\s]+([^\s]{1,}) \(", docs, flags = re.MULTILINE)
+ args2 = [x for x in args2 if not x.endswith("_file")]
+ args = list(set(args + args2))
+
+ kwargs = {}
+ for arg in args: kwargs[arg] = getattr(slow_tokenizer, arg, None)
+ kwargs["tokenizer_object"] = try_fix_tokenizer(slow_tokenizer, prepend = True)
+ fast_tokenizer = FastTokenizer( **kwargs )
+
+ # Check if they're similar!
+ sorted_slow_tokenizer = get_sorted_dict(slow_tokenizer.get_vocab())
+ sorted_fast_tokenizer = get_sorted_dict(fast_tokenizer.get_vocab())
+
+ check_vocab = (sorted_slow_tokenizer == sorted_fast_tokenizer)
+ check_special = (slow_tokenizer.all_special_tokens == fast_tokenizer.all_special_tokens)
+
+ # Failure so return slow_tokenizer
+ if not check_vocab or not check_special: return slow_tokenizer
+
+ # Now confirm if they match
+ if not assert_same_tokenization(slow_tokenizer, fast_tokenizer):
+ # Maybe remove prepending of __apple?
+ kwargs["tokenizer_object"] = try_fix_tokenizer(slow_tokenizer, prepend = False)
+ fast_tokenizer = FastTokenizer( **kwargs )
+ if not assert_same_tokenization(slow_tokenizer, fast_tokenizer):
+ # Failure :(
+ return slow_tokenizer
+ pass
+ pass
+
+ # Also tokenizer.model is missing!
+ name = slow_tokenizer.name_or_path.replace("/", "_")
+ if not os.path.exists(temporary_location):
+ os.makedirs(temporary_location)
+ pass
+ new_location = f"{temporary_location}/{name}"
+ slow_tokenizer.save_pretrained(new_location)
+ fast_tokenizer.save_pretrained(new_location)
+
+ # Now load it!
+ fast_tokenizer = AutoTokenizer.from_pretrained(new_location)
+ if assert_same_tokenization(slow_tokenizer, fast_tokenizer):
+ return fast_tokenizer
+ return slow_tokenizer
+pass
+
+
+def assert_same_tokenization(slow_tokenizer, fast_tokenizer):
+ # Get eos_token, bos_token etc
+ dir_names = dir(slow_tokenizer)
+ special_tokens = list(filter(None, (
+ getattr(slow_tokenizer, x) for x in dir_names
+ if x.endswith("_token") and x.count("_") == 1
+ )))
+ all_special_tokens = list(set(special_tokens + slow_tokenizer.all_special_tokens))
+ try:
+ string = "\n".join(all_special_tokens) + \
+ "A quick brown fox jumps over the lazy dog!!\n\nHi \n\n" + \
+ "".join(all_special_tokens)
+ return slow_tokenizer(string).input_ids == fast_tokenizer(string).input_ids
+ except:
+ # For eg see https://github.com/unslothai/unsloth/issues/292
+ # Sometimes tokenizer has weird tokens, causing a combined tokenization to fail.
+ # [TODO] We temporarily disable this for CodeLlama tokenizers
+ if slow_tokenizer.__repr__().split("(", 1)[0] in IGNORED_TOKENIZER_CHECKING:
+ return True
+ else:
+ return False
+pass
+
+
+def fix_sentencepiece_tokenizer(
+ old_tokenizer,
+ new_tokenizer,
+ token_mapping,
+ temporary_location = "_unsloth_sentencepiece_temp",
+):
+ # From https://github.com/google/sentencepiece/issues/121
+ # We need to manually edit the sentencepiece tokenizer!
+ from transformers.utils import sentencepiece_model_pb2
+
+ if not os.path.exists(temporary_location):
+ os.makedirs(temporary_location)
+ pass
+
+ # Check if tokenizer.model exists
+ if not os.path.isfile(f"{temporary_location}/tokenizer.model"):
+ return new_tokenizer
+ pass
+
+ # First save the old tokenizer
+ old_tokenizer.save_pretrained(temporary_location)
+
+ tokenizer_file = sentencepiece_model_pb2.ModelProto()
+ tokenizer_file.ParseFromString(open(f"{temporary_location}/tokenizer.model", "rb").read())
+
+ # Now save the new tokenizer
+ new_tokenizer.save_pretrained(temporary_location)
+
+ # Now correct the old tokenizer's .model file
+ for old_token, new_token in token_mapping.items():
+ ids = old_tokenizer([old_token], add_special_tokens = False).input_ids
+ ids = ids[0]
+ if (len(ids) != 1):
+ # Skip this token!
+ print(f"Skip mapping {old_token} to {new_token} since {new_token} is already in the tokenizer!")
+ continue
+ pass
+ ids = ids[0]
+ # [TODO] Hack for Starling - try except
+ try:
+ tokenizer_piece = tokenizer_file.pieces[ids]
+ except:
+ continue
+ assert(tokenizer_piece.piece == old_token)
+ tokenizer_piece.piece = new_token
+ pass
+
+ # And now write it
+ with open(f"{temporary_location}/tokenizer.model", "wb") as file:
+ file.write(tokenizer_file.SerializeToString())
+ pass
+
+ # And load it!
+ from transformers import AutoTokenizer
+ tokenizer = AutoTokenizer.from_pretrained(
+ temporary_location,
+ eos_token = new_tokenizer.eos_token,
+ pad_token = new_tokenizer.pad_token,
+ )
+ return tokenizer
+pass
+
+
+def fix_sentencepiece_gguf(saved_location):
+ """
+ Fixes sentencepiece tokenizers which did not extend the vocabulary with
+ user defined tokens.
+ Inspiration from https://github.com/ggerganov/llama.cpp/blob/master/convert-hf-to-gguf.py
+ """
+ import numpy as np
+ from copy import deepcopy
+ from transformers.utils import sentencepiece_model_pb2
+ import json
+ from enum import IntEnum
+ import os
+
+ class SentencePieceTokenTypes(IntEnum):
+ NORMAL = 1
+ UNKNOWN = 2
+ CONTROL = 3
+ USER_DEFINED = 4
+ UNUSED = 5
+ BYTE = 6
+ pass
+
+ # Load tokenizer.model
+ tokenizer_file = sentencepiece_model_pb2.ModelProto()
+ if not os.path.isfile(f"{saved_location}/tokenizer.model"): return
+ tokenizer_file.ParseFromString(open(f"{saved_location}/tokenizer.model", "rb").read())
+ sentence_piece_size = len(tokenizer_file.pieces)
+
+ # Load added_tokens_json
+ if not os.path.isfile(f"{saved_location}/added_tokens.json"): return
+ with open(f"{saved_location}/added_tokens.json", "r", encoding = "utf-8") as file:
+ added_tokens_json = json.load(file)
+ pass
+ if len(added_tokens_json) == 0: return
+
+ added_tokens_json = dict(sorted(added_tokens_json.items(), key = lambda item: item[1]))
+ new_size = sentence_piece_size + len(added_tokens_json)
+
+ # Confirm added_tokens_json is correct
+ added_tokens_ids = np.array(list(added_tokens_json.values()))
+ diff = np.diff(added_tokens_ids)
+ if (diff.min() != 1 or diff.max() != 1): return
+ if (added_tokens_ids.min() != sentence_piece_size): return
+
+ # Edit sentence piece tokens with added_tokens_json
+ logger.warning(
+ f"Unsloth: Extending {saved_location}/tokenizer.model with added_tokens.json.\n"\
+ f"Originally tokenizer.model is of size ({sentence_piece_size}).\n"\
+ f"But we need to extend to sentencepiece vocab size ({new_size})."
+ )
+ new_tokens = deepcopy(tokenizer_file.pieces[-len(added_tokens_ids):])
+ for new_token, added_token in zip(new_tokens, added_tokens_json.keys()):
+ new_token.piece = added_token.encode("utf-8")
+ new_token.score = -1000.0
+ new_token.type = SentencePieceTokenTypes.USER_DEFINED
+ pass
+
+ tokenizer_file.pieces.extend(new_tokens)
+
+ with open(f"{saved_location}/tokenizer.model", "wb") as file:
+ file.write(tokenizer_file.SerializeToString())
+ pass
+
+ # Add padding tokens
+ # actual_vocab_size = model.config.vocab_size
+ # padding = actual_vocab_size - len(tokenizer_file.pieces)
+ return
+pass
+
+
+def load_correct_tokenizer(
+ tokenizer_name,
+ model_max_length = None,
+ padding_side = "right",
+ token = None,
+ trust_remote_code = False,
+ cache_dir = "huggingface_tokenizers_cache",
+):
+ if IS_COLAB_ENVIRONMENT or IS_KAGGLE_ENVIRONMENT:
+ cache_dir = cache_dir
+ else:
+ cache_dir = None
+ pass
+
+ # Try loading the slow tokenizer. If it fails, then try Fast only
+ # Mainly to solve Deepseek models with no tokenizer.model file
+ slow_tokenizer = None
+ try:
+ slow_tokenizer = AutoTokenizer.from_pretrained(
+ tokenizer_name,
+ model_max_length = model_max_length,
+ padding_side = padding_side,
+ token = token,
+ trust_remote_code = trust_remote_code,
+ # Cannot just use use_fast = False as per https://twitter.com/danielhanchen/status/1789659394302718373
+ use_fast = False,
+ legacy = False,
+ from_slow = True,
+ cache_dir = cache_dir,
+ )
+ except:
+ print(
+ f"Unsloth: {tokenizer_name} has no tokenizer.model file.\n"\
+ "Just informing you about this - this is not a critical error."
+ )
+ pass
+
+ fast_tokenizer = AutoTokenizer.from_pretrained(
+ tokenizer_name,
+ model_max_length = model_max_length,
+ padding_side = padding_side,
+ token = token,
+ trust_remote_code = trust_remote_code,
+ cache_dir = cache_dir,
+ )
+
+ if slow_tokenizer is not None:
+ if hasattr(fast_tokenizer, "add_bos_token") and hasattr(slow_tokenizer, "add_bos_token"):
+ fast_tokenizer.add_bos_token = slow_tokenizer.add_bos_token
+ if hasattr(fast_tokenizer, "add_eos_token") and hasattr(slow_tokenizer, "add_eos_token"):
+ fast_tokenizer.add_eos_token = slow_tokenizer.add_eos_token
+
+ # Confirm if slow and fast are equivalent!
+ if assert_same_tokenization(slow_tokenizer, fast_tokenizer):
+ return fast_tokenizer
+ else:
+ return convert_to_fast_tokenizer(slow_tokenizer)
+ pass
+ else:
+ return fast_tokenizer
+ pass
+pass
+
+
+def check_tokenizer(
+ model,
+ tokenizer,
+ model_name = "unsloth/llama-2-7b-bnb-4bit",
+ model_max_length = 4096,
+ padding_side = "right",
+ token = None,
+ _reload = True,
+):
+ # Checks tokenizer for out of bounds ids.
+ # Mainly a fix for https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha
+ # where had token id=32002.
+ # See https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha/discussions/25
+ # Seems like the Fast tokenizer in Rust breaks things!
+
+ # We ignore some of them!
+ if tokenizer.__repr__().split("(", 1)[0] in IGNORED_TOKENIZER_CHECKING:
+ return tokenizer
+ pass
+
+ max_embedding_size = model.model.embed_tokens.weight.shape[0]
+ added_tokens_fast = tokenizer.added_tokens_decoder
+ added_tokens_fast = {index : str(value) for index, value in added_tokens_fast.items()}
+ sorted_keys = sorted(added_tokens_fast)
+ added_tokens_fast = {key : added_tokens_fast[key] for key in sorted_keys}
+
+ for j, index in enumerate(added_tokens_fast.keys()):
+ if index >= max_embedding_size:
+ bad_indices = list(added_tokens_fast.keys ())[j:]
+ bad_tokens = list(added_tokens_fast.values())[j:]
+ if not _reload:
+ # Try removing the token
+ added_tokens = [str(x) for x in tokenizer.added_tokens_decoder.values()]
+ special_tokens = tokenizer.special_tokens_map
+ import itertools
+ special_tokens = frozenset(
+ itertools.chain.from_iterable(
+ [x] if type(x) is str else x for x in special_tokens.values()
+ )
+ )
+ can_be_removed1 = [x for x in bad_tokens if x not in special_tokens]
+ can_be_removed2 = [x for x in can_be_removed1 if x in tokenizer._added_tokens_encoder.keys()]
+
+ # Check of extra tokens can in fact we removed!
+ can_be_removed = \
+ (len(can_be_removed1) == len(bad_tokens)) and \
+ (len(can_be_removed2) == len(bad_tokens))
+
+ # Check if sep_token or other generic types
+ remove_generic = False
+ try_mapper = []
+ if not can_be_removed:
+ names = dir(tokenizer)
+ names = (x for x in names if x.endswith("_token") and x.count("_") == 1)
+ generic_tokens = [(x, getattr(tokenizer, x, None)) for x in names]
+
+ try_removal = []
+ for token in bad_tokens:
+ for (name_token, check_token) in generic_tokens:
+ if check_token == token:
+ try_removal.append(token)
+ try_mapper.append(name_token)
+ pass
+ pass
+ pass
+
+ # Recheck!
+ can_be_removed = (len(try_removal) == len(bad_tokens))
+ if can_be_removed: remove_generic = True
+ can_be_removed1 = bad_tokens
+ pass
+
+ if can_be_removed:
+ # Yes it can be fixed!
+ for j, bad_token in enumerate(can_be_removed1):
+ remove_id = tokenizer._added_tokens_encoder[bad_token]
+ del tokenizer._added_tokens_decoder[remove_id]
+ del tokenizer._added_tokens_encoder[bad_token]
+
+ if remove_generic and (try_removal[j] == bad_token):
+ # Remove sep token for example
+ setattr(tokenizer, try_mapper[j], None)
+ setattr(tokenizer, try_mapper[j] + "_id", None)
+ pass
+ pass
+ # Confirm 1 more time!
+ if max(tokenizer.added_tokens_decoder.keys()) < max_embedding_size:
+ logger.warning_once(
+ f"Unsloth loaded a broken tokenizer `{model_name}`, but managed to repair it!\n"\
+ f"Tokens {bad_tokens} with ids {bad_indices} exceeds the max vocab size of {max_embedding_size}.\n"\
+ "We removed these bad tokens. If you think this is incorrect, fix your tokenizer first."
+ )
+ return convert_to_fast_tokenizer(tokenizer)
+ pass
+ pass
+
+ # :( Failure
+ raise RuntimeError(
+ f"Unsloth tried to load `{model_name}`, but cannot succeed.\n"\
+ f"Tokens {bad_tokens} with ids {bad_indices} exceeds the max vocab size of {max_embedding_size}.\n"\
+ f"Fix your tokenizer since it'll perform out of bounds memory accesses."
+ )
+ pass
+
+ if IS_COLAB_ENVIRONMENT or IS_KAGGLE_ENVIRONMENT:
+ cache_dir = "huggingface_tokenizers_cache"
+ else:
+ cache_dir = None
+ pass
+
+ # Sometimes slow tokenizer does not work like Deepseek
+ try:
+ # Try slow tokenizer which can fix things!
+ tokenizer = AutoTokenizer.from_pretrained(
+ model_name,
+ model_max_length = model_max_length,
+ padding_side = padding_side,
+ token = token,
+ # Cannot just use use_fast = False as per https://twitter.com/danielhanchen/status/1789659394302718373
+ use_fast = False,
+ legacy = False,
+ from_slow = True,
+ cache_dir = cache_dir,
+ )
+ return check_tokenizer(
+ model = model,
+ tokenizer = tokenizer,
+ model_name = model_name,
+ model_max_length = model_max_length,
+ padding_side = padding_side,
+ token = token,
+ _reload = False,
+ )
+ break
+ except:
+ # Tokenizer has out of bounds issues and we can't
+ # load the slow tokenizer version :(
+ logger.warning_once(
+ "Unsloth: Tokenizer is most likely buggy, and Unsloth failed to repair it.\n"\
+ "It will still work, but beware of out of bounds memory accesses.\n"\
+ "Please file an issue on the model owner's repo about this issue."
+ )
+ return tokenizer
+ pass
+ pass
+ pass
+ return convert_to_fast_tokenizer(tokenizer)
+pass
+
+
+@torch.inference_mode
+def fix_untrained_tokens(model, eps = 1e-16):
+ """
+ Llama-3 for eg has untrained vectors in the base model.
+ These include <|eot_id|>, <|start_header_id|>, <|end_header_id|>
+ We reset them to the mean of the rest of the tokens
+ """
+ embedding_matrix = model.get_input_embeddings ().weight.data
+ lm_head_matrix = model.get_output_embeddings().weight.data
+
+ # Get untrained tokens
+ indicator_untrained = torch.amax(embedding_matrix, axis = 1) <= eps
+ where_untrained = torch.where(indicator_untrained)[0]
+ n_untrained = where_untrained.shape[0]
+ n_trained = embedding_matrix.shape[0] - n_untrained
+ if n_untrained != 0:
+ print(
+ f"Unsloth: Not an error, but your model has {n_untrained} untrained tokens.\n"\
+ "We shall set them to the mean of the other trained tokens."
+ )
+ pass
+
+ # First set untrained to all 0s - sometimes it's not! 1e-23 for bfloat16
+ embedding_matrix[where_untrained] = 0
+ lm_head_matrix [where_untrained] = 0
+
+ # Find sum
+ sum_embedding = torch.sum(embedding_matrix, dtype = torch.float32, axis = 0)
+ sum_lm_head = torch.sum(lm_head_matrix, dtype = torch.float32, axis = 0)
+
+ # Find correct average by dividing by sum of trained tokens
+ mean_embedding = (sum_embedding / n_trained).to(embedding_matrix.dtype)
+ mean_lm_head = (sum_lm_head / n_trained).to(lm_head_matrix .dtype)
+
+ # Set them to the mean
+ embedding_matrix[where_untrained] = mean_embedding
+ lm_head_matrix [where_untrained] = mean_lm_head
+
+ return mean_embedding, mean_lm_head
+pass
+
+
+@torch.inference_mode
+def mean_of_trained_tokens(model, eps = 1e-16):
+ """
+ Llama-3 for eg has untrained vectors in the base model.
+ These include <|eot_id|>, <|start_header_id|>, <|end_header_id|>
+ We reset them to the mean of the rest of the tokens
+ """
+ embedding_matrix = model.get_input_embeddings ().weight.data.clone()
+ lm_head_matrix = model.get_output_embeddings().weight.data.clone()
+
+ # Get untrained tokens
+ indicator_untrained = torch.amax(embedding_matrix, axis = 1) <= eps
+ where_untrained = torch.where(indicator_untrained)[0]
+ n_untrained = where_untrained.shape[0]
+ n_trained = embedding_matrix.shape[0] - n_untrained
+ if n_untrained != 0:
+ print(
+ f"Unsloth: Not an error, but your model has {n_untrained} untrained tokens.\n"\
+ "We shall set them to the mean of the other trained tokens."
+ )
+ pass
+
+ # First set untrained to all 0s - sometimes it's not! 1e-23 for bfloat16
+ embedding_matrix[where_untrained] = 0
+ lm_head_matrix [where_untrained] = 0
+
+ # Find sum
+ sum_embedding = torch.sum(embedding_matrix, dtype = torch.float32, axis = 0)
+ sum_lm_head = torch.sum(lm_head_matrix, dtype = torch.float32, axis = 0)
+
+ # Find correct average by dividing by sum of trained tokens
+ mean_embedding = (sum_embedding / n_trained).to(embedding_matrix.dtype)
+ mean_lm_head = (sum_lm_head / n_trained).to(lm_head_matrix .dtype)
+
+ return mean_embedding, mean_lm_head
+pass
+
+
+@torch.inference_mode
+def add_new_tokens(
+ model,
+ tokenizer,
+ new_tokens = [],
+ method = "mean",
+ interpolation = 0.5,
+):
+ """
+ Smartly resizes the tokenizer and adds new tokens to the model.
+ We also disregard untrained tokens by removing them from the mean calculation.
+ """
+ assert(isinstance(new_tokens, (list, tuple)))
+ assert(len(new_tokens) > 0)
+ assert(method == "mean" or method == "interpolation")
+ assert(interpolation >= 0 and interpolation <= 1)
+
+ # Check if tokens already exist
+ overlapping_tokens = set(new_tokens) & set(tokenizer.vocab.keys())
+ if len(overlapping_tokens) != 0:
+ print(
+ f"Unsloth: You're adding new_tokens = {new_tokens}\n"\
+ f"There are tokens which are overlapping = {list(overlapping_tokens)}\n"\
+ f"We shall safely ignore these overlapping tokens."
+ )
+ new_tokens = [x for x in new_tokens if x not in overlapping_tokens]
+ pass
+
+ # Get mean of trained tokens
+ # mean_embedding, mean_lm_head = fix_untrained_tokens(model)
+
+ # Weirdly be careful reserved tokens can pop out
+ mean_embedding, mean_lm_head = mean_of_trained_tokens(model)
+ mean_embedding = mean_embedding.to(torch.float32)
+ mean_lm_head = mean_lm_head .to(torch.float32)
+
+ # Add tokens!
+ old_length = len(tokenizer)
+ tokenizer.add_tokens(new_tokens)
+ model.resize_token_embeddings(len(tokenizer))
+
+ # If we use interpolation, we interpolate between the mean embeddings and
+ # the Word2Vec sum of the other vectors
+ embedding_matrix = model.get_input_embeddings ().weight.data
+ lm_head_matrix = model.get_output_embeddings().weight.data
+
+ if method == "interpolation":
+ print(
+ "Unsloth: You are using interpolation to add new tokens.\n"\
+ f"We shall set new tokens = mean(embeddings)*{1-interpolation} + mean(new_tokens)*{interpolation}"
+ )
+ for j, token in enumerate(new_tokens):
+ input_ids = tokenizer(token, add_special_tokens = False).input_ids
+ mean_embedding_token = embedding_matrix[input_ids].mean(axis = 0, dtype = torch.float32)
+ mean_lm_head_token = lm_head_matrix [input_ids].mean(axis = 0, dtype = torch.float32)
+
+ # Interpolate
+ mean_embedding_token = mean_embedding*(1-interpolation) + mean_embedding_token*interpolation
+ mean_lm_head_token = mean_lm_head *(1-interpolation) + mean_lm_head_token *interpolation
+
+ # Set the new vector
+ embedding_matrix[old_length+j] = mean_embedding_token
+ lm_head_matrix [old_length+j] = mean_lm_head_token
+ pass
+ else:
+ # Now set the new tokens to the mean!
+ embedding_matrix[old_length:] = mean_embedding
+ lm_head_matrix [old_length:] = mean_lm_head
+ pass
+
+ # We set a flag to say we need to train embeddings
+ internal_model = model
+ while hasattr(internal_model, "model"):
+ internal_model._need_to_train_embeddings = True
+ internal_model = internal_model.model
+ pass
+ internal_model._need_to_train_embeddings = True
+
+ return
+pass
+
+
+from inspect import getsource
+import trl.trainer.sft_trainer
+from trl.trainer.sft_trainer import *
+
+def fix_sft_trainer_tokenizer():
+ """
+ Fixes double adding BOS tokens like in llama-3
+ """
+ for function_name, replacer in (
+ ("_prepare_non_packed_dataloader", "def tokenize(element):",),
+ # ("_prepare_packed_dataloader", "if dataset_text_field is not None",),
+ ):
+ function = getsource(eval(f"trl.trainer.sft_trainer.SFTTrainer.{function_name}"))
+ where = function.find("def")
+ function = function.split("\n")
+ function = "\n".join(x[where:] for x in function)
+
+ check_text = \
+ "\n"\
+ "test_text = dataset[0][dataset_text_field] if (formatting_func is None or not use_formatting_func) else formatting_func(dataset[0])\n"\
+ "chat_template = getattr(tokenizer, 'chat_template', None)\n"\
+ "chat_template = '' if chat_template is None else chat_template\n"\
+ "has_bos_token_already = (test_text.startswith(tokenizer.bos_token) or tokenizer.bos_token in chat_template) "\
+ "if getattr(tokenizer, 'bos_token', None) is not None else False\n"\
+ "add_special_tokens = False if has_bos_token_already else add_special_tokens\n\n"
+
+ check_text = check_text.split("\n")
+ check_text = "\n".join(" "*where + x for x in check_text)
+
+ function = function.replace(replacer, check_text + replacer)
+ exec(function, globals())
+
+ # Replace TRL's SFTTrainer
+ exec(f"trl.trainer.sft_trainer.SFTTrainer.{function_name} = {function_name}", globals())
+ pass
+pass
+
+# Fixes double adding BOS tokens like in llama-3
+fix_sft_trainer_tokenizer()
diff --git a/wandb/debug-cli.root.log b/wandb/debug-cli.root.log
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/wandb/debug-internal.log b/wandb/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..fe979cae789bc1173bd85436f162519bfc46cc46
--- /dev/null
+++ b/wandb/debug-internal.log
@@ -0,0 +1,179 @@
+2024-05-16 03:27:39,985 INFO StreamThr :25247 [internal.py:wandb_internal():85] W&B internal server running at pid: 25247, started at: 2024-05-16 03:27:39.984560
+2024-05-16 03:27:39,987 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status
+2024-05-16 03:27:39,992 INFO WriterThread:25247 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/run-kmszoyl0.wandb
+2024-05-16 03:27:39,993 DEBUG SenderThread:25247 [sender.py:send():378] send: header
+2024-05-16 03:27:40,003 DEBUG SenderThread:25247 [sender.py:send():378] send: run
+2024-05-16 03:27:40,392 INFO SenderThread:25247 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files
+2024-05-16 03:27:40,393 INFO SenderThread:25247 [sender.py:_start_run_threads():1123] run started: kmszoyl0 with start time 1715830059.989096
+2024-05-16 03:27:40,425 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: check_version
+2024-05-16 03:27:40,426 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: check_version
+2024-05-16 03:27:40,511 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: run_start
+2024-05-16 03:27:40,545 DEBUG HandlerThread:25247 [system_info.py:__init__():26] System info init
+2024-05-16 03:27:40,545 DEBUG HandlerThread:25247 [system_info.py:__init__():41] System info init done
+2024-05-16 03:27:40,545 INFO HandlerThread:25247 [system_monitor.py:start():194] Starting system monitor
+2024-05-16 03:27:40,545 INFO SystemMonitor:25247 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-16 03:27:40,546 INFO HandlerThread:25247 [system_monitor.py:probe():214] Collecting system info
+2024-05-16 03:27:40,546 INFO SystemMonitor:25247 [interfaces.py:start():188] Started cpu monitoring
+2024-05-16 03:27:40,547 INFO SystemMonitor:25247 [interfaces.py:start():188] Started disk monitoring
+2024-05-16 03:27:40,547 INFO SystemMonitor:25247 [interfaces.py:start():188] Started gpu monitoring
+2024-05-16 03:27:40,548 INFO SystemMonitor:25247 [interfaces.py:start():188] Started memory monitoring
+2024-05-16 03:27:40,549 INFO SystemMonitor:25247 [interfaces.py:start():188] Started network monitoring
+2024-05-16 03:27:40,625 DEBUG HandlerThread:25247 [system_info.py:probe():150] Probing system
+2024-05-16 03:27:40,628 DEBUG HandlerThread:25247 [system_info.py:_probe_git():135] Probing git
+2024-05-16 03:27:40,641 DEBUG HandlerThread:25247 [system_info.py:_probe_git():143] Probing git done
+2024-05-16 03:27:40,641 DEBUG HandlerThread:25247 [system_info.py:probe():198] Probing system done
+2024-05-16 03:27:40,641 DEBUG HandlerThread:25247 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-16T03:27:40.625976', 'startedAt': '2024-05-16T03:27:39.977644', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '039e2a03700730f333471d130dbf1f4d7922b357'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1555.2083229166667, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.001, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 1800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 3000.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.98368072509766}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-16 03:27:40,641 INFO HandlerThread:25247 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-16 03:27:40,642 INFO HandlerThread:25247 [system_monitor.py:probe():227] Publishing system info
+2024-05-16 03:27:40,642 DEBUG HandlerThread:25247 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-16 03:27:41,157 DEBUG HandlerThread:25247 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-16 03:27:41,157 DEBUG HandlerThread:25247 [system_info.py:_save_code():44] Saving code
+2024-05-16 03:27:41,157 WARNING HandlerThread:25247 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-16 03:27:41,157 DEBUG HandlerThread:25247 [system_info.py:_save_patches():82] Saving git patches
+2024-05-16 03:27:41,183 DEBUG HandlerThread:25247 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-16 03:27:41,184 INFO HandlerThread:25247 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-16 03:27:41,189 DEBUG SenderThread:25247 [sender.py:send():378] send: files
+2024-05-16 03:27:41,189 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-16 03:27:41,190 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file diff.patch with policy now
+2024-05-16 03:27:41,395 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-metadata.json
+2024-05-16 03:27:41,396 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/conda-environment.yaml
+2024-05-16 03:27:41,396 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/diff.patch
+2024-05-16 03:27:41,409 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-16 03:27:41,409 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: python_packages
+2024-05-16 03:27:41,410 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:27:41,410 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:27:41,614 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-16 03:27:41,745 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,745 DEBUG SenderThread:25247 [sender.py:send():378] send: config
+2024-05-16 03:27:41,747 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:41,747 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,747 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,747 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:41,748 WARNING SenderThread:25247 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-16 03:27:41,748 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,748 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,748 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: log_artifact
+2024-05-16 03:27:41,838 INFO wandb-upload_0:25247 [upload_job.py:push():130] Uploaded file /tmp/tmpz_h0aag_wandb/6blsqx0w-wandb-metadata.json
+2024-05-16 03:27:42,393 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/requirements.txt
+2024-05-16 03:27:42,394 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/tmp/axolotl_config_h7qiqpxe.yml
+2024-05-16 03:27:42,394 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/tmp
+2024-05-16 03:27:42,990 INFO wandb-upload_0:25247 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmptvdelfam
+2024-05-16 03:27:43,725 INFO SenderThread:25247 [sender.py:send_request_log_artifact():1455] logged artifact config-kmszoyl0 - {'id': 'QXJ0aWZhY3Q6ODM4MjMxMDk3', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTQxMDQ4Mg==', 'latestArtifact': None}}
+2024-05-16 03:27:43,726 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:43,727 DEBUG SenderThread:25247 [sender.py:send():378] send: files
+2024-05-16 03:27:43,727 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file tmp/axolotl_config_h7qiqpxe.yml with policy live
+2024-05-16 03:27:44,118 INFO wandb-upload_0:25247 [upload_job.py:push():130] Uploaded file /tmp/tmpz_h0aag_wandb/pz983g49-tmp/axolotl_config_h7qiqpxe.yml
+2024-05-16 03:27:44,395 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:27:45,622 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:27:46,396 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:27:50,623 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:27:51,779 INFO wandb-upload_1:25247 [upload_job.py:push():130] Uploaded file /tmp/tmpz_h0aag_wandb/s1zo24fv-diff.patch
+2024-05-16 03:27:53,781 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:27:53,786 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:53,787 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:53,787 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:53,788 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:53,788 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:27:53,788 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:27:53,791 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:27:54,401 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:27:55,793 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:27:56,403 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:27:56,410 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:27:56,411 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:01,534 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:01,535 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:01,537 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:01,538 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:01,543 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:02,407 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:04,408 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:06,545 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:09,646 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:09,648 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:09,649 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:09,653 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:10,412 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:11,410 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:28:11,411 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:11,664 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:12,418 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/config.yaml
+2024-05-16 03:28:12,418 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:16,911 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:17,538 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:17,540 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:17,542 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:17,545 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:18,422 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:20,423 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:22,547 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:25,403 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:25,405 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:25,407 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:25,410 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:25,426 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:26,411 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:28:26,412 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:27,644 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:28,428 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:32,644 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:33,324 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:33,326 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:33,327 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:33,334 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:33,432 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:34,432 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:36,434 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:38,336 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:40,550 DEBUG SystemMonitor:25247 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-16 03:28:40,555 DEBUG SenderThread:25247 [sender.py:send():378] send: stats
+2024-05-16 03:28:41,296 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:41,299 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:41,300 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:41,301 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:41,411 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:28:41,412 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:41,437 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:42,438 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:43,629 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:48,630 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:49,254 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:49,257 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:49,259 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:49,262 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:49,443 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:50,444 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:54,264 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:57,171 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:28:57,172 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:57,178 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:57,360 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:57,360 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:57,361 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:57,448 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:58,449 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:59,363 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:04,363 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:05,121 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:29:05,124 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:29:05,125 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:29:05,129 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:29:05,453 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:29:06,454 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:29:10,131 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:10,559 DEBUG SenderThread:25247 [sender.py:send():378] send: stats
+2024-05-16 03:29:12,977 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:29:12,978 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:29:13,022 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:29:13,201 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:29:13,202 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:29:13,203 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:29:13,459 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:29:14,459 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:29:15,204 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:20,205 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:22,578 INFO memory :25247 [interfaces.py:monitor():138] Process proc.memory.rssMB has exited.
+2024-05-16 03:29:22,578 DEBUG SystemMonitor:25247 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-16 03:29:22,580 DEBUG SystemMonitor:25247 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-16 03:29:22,589 DEBUG SenderThread:25247 [sender.py:send():378] send: stats
+2024-05-16 03:29:23,536 INFO MainThread:25247 [internal.py:handle_exit():75] Internal process exited
diff --git a/wandb/debug.log b/wandb/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..67689dec5bd9a6acc4041f1c28375efdad1c362f
--- /dev/null
+++ b/wandb/debug.log
@@ -0,0 +1,31 @@
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Configure stats pid to 24653
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-16 03:27:39,980 WARNING MainThread:24653 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-16 03:27:39,980 WARNING MainThread:24653 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/logs/debug.log
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/logs/debug-internal.log
+2024-05-16 03:27:39,981 INFO MainThread:24653 [wandb_init.py:init():560] calling init triggers
+2024-05-16 03:27:39,981 INFO MainThread:24653 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-16 03:27:39,981 INFO MainThread:24653 [wandb_init.py:init():610] starting backend
+2024-05-16 03:27:39,981 INFO MainThread:24653 [wandb_init.py:init():614] setting up manager
+2024-05-16 03:27:39,983 INFO MainThread:24653 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-16 03:27:39,988 INFO MainThread:24653 [wandb_init.py:init():622] backend started and connected
+2024-05-16 03:27:39,992 INFO MainThread:24653 [wandb_init.py:init():711] updated telemetry
+2024-05-16 03:27:40,002 INFO MainThread:24653 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-16 03:27:40,424 INFO MainThread:24653 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-16 03:27:40,498 INFO MainThread:24653 [wandb_run.py:_on_init():2405] got version response
+2024-05-16 03:27:40,498 INFO MainThread:24653 [wandb_init.py:init():795] starting run threads in backend
+2024-05-16 03:27:41,410 INFO MainThread:24653 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-16 03:27:41,410 INFO MainThread:24653 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-16 03:27:41,410 INFO MainThread:24653 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-16 03:27:41,410 INFO MainThread:24653 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-16 03:27:41,411 INFO MainThread:24653 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-16 03:27:41,413 INFO MainThread:24653 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.2', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 1, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 8628, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May16_03-27-11_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero1.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_bnb_8bit', 'optim_args': 'amsgrad=True', 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.92, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 1, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-16 03:27:41,415 INFO MainThread:24653 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240512_102028-p88snecz/files/conda-environment.yaml b/wandb/run-20240512_102028-p88snecz/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240512_102028-p88snecz/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240512_102028-p88snecz/files/config.yaml b/wandb/run-20240512_102028-p88snecz/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..92214715f409a1074056bc80f5adc62828470d21
--- /dev/null
+++ b/wandb/run-20240512_102028-p88snecz/files/config.yaml
@@ -0,0 +1,750 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715509228
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.0.dev0
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./pyg1m2.3
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: true
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: steps
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 2
+per_device_eval_batch_size:
+ desc: null
+ value: 2
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 2
+eval_accumulation_steps:
+ desc: null
+ value: 2
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 5.0e-05
+weight_decay:
+ desc: null
+ value: 0.03
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.999
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 1.0
+num_train_epochs:
+ desc: null
+ value: 2
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value: {}
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 71
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./pyg1m2.3/runs/May12_10-17-42_a9815c79a99f
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.03333333333333333
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: true
+eval_steps:
+ desc: null
+ value: 0.015
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./pyg1m2.3
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value:
+ - full_shard
+ - auto_wrap
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: true
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: false
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_sharding_strategy: FULL_SHARD
+ fsdp_forward_prefetch: FORWARD_PRE
+ fsdp_backward_prefetch: BACKWARD_PRE
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: null
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_torch
+optim_args:
+ desc: null
+ value: adam_beta1=0.98,adam_beta2=0.97,amsgrad=True
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: null
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.3
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: null
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: false
+sample_packing_efficiency:
+ desc: null
+ value: 0.96
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 2
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240512_102028-p88snecz/files/output.log b/wandb/run-20240512_102028-p88snecz/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..06640dc8381a1aae024a9adad5d4ef43a7a89c47
--- /dev/null
+++ b/wandb/run-20240512_102028-p88snecz/files/output.log
@@ -0,0 +1,11 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/2966 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+[2024-05-12 10:20:29,505] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:37041] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-12 10:20:30,530] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:37041] [RANK:0] packing_efficiency_estimate: 0.96 total_num_tokens per device: 47141994
+[2024-05-12 10:20:31,296] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:37041] [RANK:0] packing_efficiency_estimate: 0.96 total_num_tokens per device: 47141994
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[rank0]:[2024-05-12 10:20:43,887] [11/0] torch._dynamo.variables.higher_order_ops: [WARNING] speculate_subgraph: while introspecting the user-defined autograd.Function, we were unable to trace function `trampoline_autograd_fwd` into a single graph. This means that Dynamo was unable to prove safety for this API and will fall back to eager-mode PyTorch, which could lead to a slowdown.
+[rank0]:[2024-05-12 10:20:43,887] [11/0] torch._dynamo.variables.higher_order_ops: [ERROR] Tensor.data_ptr
+[rank0]:[2024-05-12 10:20:43,902] [12/0] torch._dynamo.variables.higher_order_ops: [WARNING] speculate_subgraph: while introspecting the user-defined autograd.Function, we were unable to trace function `trampoline_autograd_fwd` into a single graph. This means that Dynamo was unable to prove safety for this API and will fall back to eager-mode PyTorch, which could lead to a slowdown.
+[rank0]:[2024-05-12 10:20:43,902] [12/0] torch._dynamo.variables.higher_order_ops: [ERROR] Tensor.data_ptr
\ No newline at end of file
diff --git a/wandb/run-20240512_102028-p88snecz/files/requirements.txt b/wandb/run-20240512_102028-p88snecz/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..54e63626ab0d75f9dec791c2e6499b990c4392b8
--- /dev/null
+++ b/wandb/run-20240512_102028-p88snecz/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash-attn==2.5.5
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240512_102028-p88snecz/files/wandb-metadata.json b/wandb/run-20240512_102028-p88snecz/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..c78406645ccdf303eab0c3146c5334ed01297438
--- /dev/null
+++ b/wandb/run-20240512_102028-p88snecz/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-12T10:20:28.667766",
+ "startedAt": "2024-05-12T10:20:28.026968",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "2147cf6837e2b90a2ea7045262083cbb0da03858"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "a9815c79a99f",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1268.72221875,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.003,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.003,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.002,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 149.2169532775879
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240512_102028-p88snecz/logs/debug-internal.log b/wandb/run-20240512_102028-p88snecz/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..151d9652327dcc33f3276a30a5b64b9a5d825781
--- /dev/null
+++ b/wandb/run-20240512_102028-p88snecz/logs/debug-internal.log
@@ -0,0 +1,99 @@
+2024-05-12 10:20:28,039 INFO StreamThr :38845 [internal.py:wandb_internal():85] W&B internal server running at pid: 38845, started at: 2024-05-12 10:20:28.039296
+2024-05-12 10:20:28,042 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: status
+2024-05-12 10:20:28,056 INFO WriterThread:38845 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/run-p88snecz.wandb
+2024-05-12 10:20:28,057 DEBUG SenderThread:38845 [sender.py:send():378] send: header
+2024-05-12 10:20:28,069 DEBUG SenderThread:38845 [sender.py:send():378] send: run
+2024-05-12 10:20:28,444 INFO SenderThread:38845 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files
+2024-05-12 10:20:28,445 INFO SenderThread:38845 [sender.py:_start_run_threads():1123] run started: p88snecz with start time 1715509228.040864
+2024-05-12 10:20:28,475 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: check_version
+2024-05-12 10:20:28,476 DEBUG SenderThread:38845 [sender.py:send_request():405] send_request: check_version
+2024-05-12 10:20:28,559 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: run_start
+2024-05-12 10:20:28,593 DEBUG HandlerThread:38845 [system_info.py:__init__():26] System info init
+2024-05-12 10:20:28,594 DEBUG HandlerThread:38845 [system_info.py:__init__():41] System info init done
+2024-05-12 10:20:28,594 INFO HandlerThread:38845 [system_monitor.py:start():194] Starting system monitor
+2024-05-12 10:20:28,594 INFO SystemMonitor:38845 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-12 10:20:28,594 INFO HandlerThread:38845 [system_monitor.py:probe():214] Collecting system info
+2024-05-12 10:20:28,594 INFO SystemMonitor:38845 [interfaces.py:start():188] Started cpu monitoring
+2024-05-12 10:20:28,595 INFO SystemMonitor:38845 [interfaces.py:start():188] Started disk monitoring
+2024-05-12 10:20:28,596 INFO SystemMonitor:38845 [interfaces.py:start():188] Started gpu monitoring
+2024-05-12 10:20:28,596 INFO SystemMonitor:38845 [interfaces.py:start():188] Started memory monitoring
+2024-05-12 10:20:28,597 INFO SystemMonitor:38845 [interfaces.py:start():188] Started network monitoring
+2024-05-12 10:20:28,667 DEBUG HandlerThread:38845 [system_info.py:probe():150] Probing system
+2024-05-12 10:20:28,670 DEBUG HandlerThread:38845 [system_info.py:_probe_git():135] Probing git
+2024-05-12 10:20:28,680 DEBUG HandlerThread:38845 [system_info.py:_probe_git():143] Probing git done
+2024-05-12 10:20:28,681 DEBUG HandlerThread:38845 [system_info.py:probe():198] Probing system done
+2024-05-12 10:20:28,681 DEBUG HandlerThread:38845 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-12T10:20:28.667766', 'startedAt': '2024-05-12T10:20:28.026968', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '2147cf6837e2b90a2ea7045262083cbb0da03858'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'a9815c79a99f', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1268.72221875, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1400.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.003, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.003, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.002, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 149.2169532775879}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-12 10:20:28,681 INFO HandlerThread:38845 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-12 10:20:28,681 INFO HandlerThread:38845 [system_monitor.py:probe():227] Publishing system info
+2024-05-12 10:20:28,681 DEBUG HandlerThread:38845 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-12 10:20:29,096 DEBUG HandlerThread:38845 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-12 10:20:29,099 INFO HandlerThread:38845 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-12 10:20:29,110 DEBUG SenderThread:38845 [sender.py:send():378] send: files
+2024-05-12 10:20:29,111 INFO SenderThread:38845 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-12 10:20:29,295 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-12 10:20:29,295 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-12 10:20:29,296 DEBUG SenderThread:38845 [sender.py:send_request():405] send_request: python_packages
+2024-05-12 10:20:29,297 DEBUG SenderThread:38845 [sender.py:send_request():405] send_request: stop_status
+2024-05-12 10:20:29,447 INFO Thread-12 :38845 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/requirements.txt
+2024-05-12 10:20:29,447 INFO Thread-12 :38845 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/conda-environment.yaml
+2024-05-12 10:20:29,447 INFO Thread-12 :38845 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/wandb-metadata.json
+2024-05-12 10:20:29,503 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-12 10:20:29,527 DEBUG SenderThread:38845 [sender.py:send():378] send: telemetry
+2024-05-12 10:20:29,527 DEBUG SenderThread:38845 [sender.py:send():378] send: config
+2024-05-12 10:20:29,529 DEBUG SenderThread:38845 [sender.py:send():378] send: metric
+2024-05-12 10:20:29,529 DEBUG SenderThread:38845 [sender.py:send():378] send: telemetry
+2024-05-12 10:20:29,529 DEBUG SenderThread:38845 [sender.py:send():378] send: telemetry
+2024-05-12 10:20:29,529 DEBUG SenderThread:38845 [sender.py:send():378] send: metric
+2024-05-12 10:20:29,529 WARNING SenderThread:38845 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-12 10:20:29,530 DEBUG SenderThread:38845 [sender.py:send():378] send: telemetry
+2024-05-12 10:20:29,530 DEBUG SenderThread:38845 [sender.py:send_request():405] send_request: log_artifact
+2024-05-12 10:20:30,446 INFO Thread-12 :38845 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/tmp/axolotl_config_deqovz7f.yml
+2024-05-12 10:20:30,446 INFO Thread-12 :38845 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/tmp
+2024-05-12 10:20:30,736 INFO wandb-upload_1:38845 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmpeu2mpquh
+2024-05-12 10:20:31,504 INFO SenderThread:38845 [sender.py:send_request_log_artifact():1455] logged artifact config-p88snecz - {'id': 'QXJ0aWZhY3Q6ODMyOTA3NjYw', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3MzY0NTcyMQ==', 'latestArtifact': None}}
+2024-05-12 10:20:31,506 DEBUG SenderThread:38845 [sender.py:send():378] send: telemetry
+2024-05-12 10:20:31,506 DEBUG SenderThread:38845 [sender.py:send():378] send: files
+2024-05-12 10:20:31,506 INFO SenderThread:38845 [sender.py:_save_file():1389] saving file tmp/axolotl_config_deqovz7f.yml with policy live
+2024-05-12 10:20:31,897 INFO wandb-upload_1:38845 [upload_job.py:push():130] Uploaded file /tmp/tmp_8vmtykewandb/4h9pat2i-tmp/axolotl_config_deqovz7f.yml
+2024-05-12 10:20:32,447 INFO Thread-12 :38845 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/output.log
+2024-05-12 10:20:33,243 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:20:34,448 INFO Thread-12 :38845 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/output.log
+2024-05-12 10:20:38,244 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:20:42,454 INFO Thread-12 :38845 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/output.log
+2024-05-12 10:20:43,889 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:20:44,298 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-12 10:20:44,299 DEBUG SenderThread:38845 [sender.py:send_request():405] send_request: stop_status
+2024-05-12 10:20:46,456 INFO Thread-12 :38845 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/output.log
+2024-05-12 10:20:49,542 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:20:54,543 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:20:58,613 INFO memory :38845 [interfaces.py:monitor():138] Process proc.memory.rssMB has exited.
+2024-05-12 10:20:58,613 DEBUG SystemMonitor:38845 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-12 10:20:58,616 DEBUG SystemMonitor:38845 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-12 10:20:58,616 DEBUG SystemMonitor:38845 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-12 10:20:58,618 DEBUG SenderThread:38845 [sender.py:send():378] send: stats
+2024-05-12 10:20:59,626 DEBUG HandlerThread:38845 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:21:00,463 INFO Thread-12 :38845 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/config.yaml
+2024-05-12 10:21:01,075 WARNING StreamThr :38845 [internal.py:is_dead():413] Internal process exiting, parent pid 37041 disappeared
+2024-05-12 10:21:01,076 ERROR StreamThr :38845 [internal.py:wandb_internal():151] Internal process shutdown.
+2024-05-12 10:21:01,627 INFO HandlerThread:38845 [handler.py:finish():882] shutting down handler
+2024-05-12 10:21:01,627 INFO HandlerThread:38845 [system_monitor.py:finish():203] Stopping system monitor
+2024-05-12 10:21:01,627 INFO HandlerThread:38845 [interfaces.py:finish():200] Joined cpu monitor
+2024-05-12 10:21:01,627 INFO HandlerThread:38845 [interfaces.py:finish():200] Joined disk monitor
+2024-05-12 10:21:01,628 INFO HandlerThread:38845 [interfaces.py:finish():200] Joined gpu monitor
+2024-05-12 10:21:01,628 INFO HandlerThread:38845 [interfaces.py:finish():200] Joined memory monitor
+2024-05-12 10:21:01,628 INFO HandlerThread:38845 [interfaces.py:finish():200] Joined network monitor
+2024-05-12 10:21:01,632 INFO WriterThread:38845 [datastore.py:close():296] close: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/run-p88snecz.wandb
+2024-05-12 10:21:01,906 INFO SenderThread:38845 [sender.py:finish():1545] shutting down sender
+2024-05-12 10:21:05,577 INFO SenderThread:38845 [dir_watcher.py:finish():358] shutting down directory watcher
+2024-05-12 10:21:06,466 INFO SenderThread:38845 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/output.log
+2024-05-12 10:21:06,467 INFO SenderThread:38845 [dir_watcher.py:finish():388] scan: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files
+2024-05-12 10:21:06,467 INFO SenderThread:38845 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/wandb-metadata.json wandb-metadata.json
+2024-05-12 10:21:06,467 INFO SenderThread:38845 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/requirements.txt requirements.txt
+2024-05-12 10:21:06,468 INFO SenderThread:38845 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/config.yaml config.yaml
+2024-05-12 10:21:06,468 INFO SenderThread:38845 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/conda-environment.yaml conda-environment.yaml
+2024-05-12 10:21:06,468 INFO SenderThread:38845 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/output.log output.log
+2024-05-12 10:21:06,469 INFO SenderThread:38845 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/files/tmp/axolotl_config_deqovz7f.yml tmp/axolotl_config_deqovz7f.yml
+2024-05-12 10:21:06,469 INFO SenderThread:38845 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-12 10:21:06,469 INFO SenderThread:38845 [file_pusher.py:join():175] waiting for file pusher
+2024-05-12 10:21:06,971 INFO SenderThread:38845 [file_stream.py:finish():601] file stream finish called
+2024-05-12 10:21:07,115 INFO SenderThread:38845 [file_stream.py:finish():605] file stream finish is done
diff --git a/wandb/run-20240512_102028-p88snecz/logs/debug.log b/wandb/run-20240512_102028-p88snecz/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..097de21c13ce26cb390c723251240e78575d47fb
--- /dev/null
+++ b/wandb/run-20240512_102028-p88snecz/logs/debug.log
@@ -0,0 +1,30 @@
+2024-05-12 10:20:28,032 INFO MainThread:37041 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-12 10:20:28,033 INFO MainThread:37041 [wandb_setup.py:_flush():76] Configure stats pid to 37041
+2024-05-12 10:20:28,033 INFO MainThread:37041 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-12 10:20:28,033 INFO MainThread:37041 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-12 10:20:28,033 INFO MainThread:37041 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-12 10:20:28,033 INFO MainThread:37041 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-12 10:20:28,033 WARNING MainThread:37041 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-12 10:20:28,033 INFO MainThread:37041 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-12 10:20:28,033 INFO MainThread:37041 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-12 10:20:28,033 INFO MainThread:37041 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
+2024-05-12 10:20:28,034 INFO MainThread:37041 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/logs/debug.log
+2024-05-12 10:20:28,034 INFO MainThread:37041 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240512_102028-p88snecz/logs/debug-internal.log
+2024-05-12 10:20:28,034 INFO MainThread:37041 [wandb_init.py:init():560] calling init triggers
+2024-05-12 10:20:28,034 INFO MainThread:37041 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-12 10:20:28,034 INFO MainThread:37041 [wandb_init.py:init():610] starting backend
+2024-05-12 10:20:28,034 INFO MainThread:37041 [wandb_init.py:init():614] setting up manager
+2024-05-12 10:20:28,037 INFO MainThread:37041 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-12 10:20:28,040 INFO MainThread:37041 [wandb_init.py:init():622] backend started and connected
+2024-05-12 10:20:28,056 INFO MainThread:37041 [wandb_init.py:init():711] updated telemetry
+2024-05-12 10:20:28,068 INFO MainThread:37041 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-12 10:20:28,475 INFO MainThread:37041 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-12 10:20:28,547 INFO MainThread:37041 [wandb_run.py:_on_init():2405] got version response
+2024-05-12 10:20:28,548 INFO MainThread:37041 [wandb_init.py:init():795] starting run threads in backend
+2024-05-12 10:20:29,296 INFO MainThread:37041 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-12 10:20:29,296 INFO MainThread:37041 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-12 10:20:29,296 INFO MainThread:37041 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-12 10:20:29,296 INFO MainThread:37041 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-12 10:20:29,297 INFO MainThread:37041 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-12 10:20:29,300 INFO MainThread:37041 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyg1m2.3', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 2, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.03, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 71, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyg1m2.3/runs/May12_10-17-42_a9815c79a99f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.03333333333333333, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': True, 'eval_steps': 0.015, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyg1m2.3', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': ['full_shard', 'auto_wrap'], 'fsdp_min_num_params': 0, 'fsdp_config': {'fsdp_limit_all_gathers': True, 'fsdp_sync_module_states': True, 'fsdp_offload_params': True, 'fsdp_use_orig_params': False, 'fsdp_cpu_ram_efficient_loading': False, 'fsdp_auto_wrap_policy': 'TRANSFORMER_BASED_WRAP', 'fsdp_transformer_layer_cls_to_wrap': 'LlamaDecoderLayer', 'fsdp_state_dict_type': 'FULL_STATE_DICT', 'fsdp_sharding_strategy': 'FULL_SHARD', 'fsdp_forward_prefetch': 'FORWARD_PRE', 'fsdp_backward_prefetch': 'BACKWARD_PRE', 'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': 'adam_beta1=0.98,adam_beta2=0.97,amsgrad=True', 'adafactor': False, 'group_by_length': None, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.3', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': False, 'sample_packing_efficiency': 0.96, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 2, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
diff --git a/wandb/run-20240512_102028-p88snecz/run-p88snecz.wandb b/wandb/run-20240512_102028-p88snecz/run-p88snecz.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..133d885332b9be2f1eb5de5f4e6009a3448e5278
Binary files /dev/null and b/wandb/run-20240512_102028-p88snecz/run-p88snecz.wandb differ
diff --git a/wandb/run-20240512_102340-g9bmlvdf/files/conda-environment.yaml b/wandb/run-20240512_102340-g9bmlvdf/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240512_102340-g9bmlvdf/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240512_102340-g9bmlvdf/files/config.yaml b/wandb/run-20240512_102340-g9bmlvdf/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..5607b05c501d0a2b552b248db3329e754cd6e009
--- /dev/null
+++ b/wandb/run-20240512_102340-g9bmlvdf/files/config.yaml
@@ -0,0 +1,766 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715509420
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+ - 1: train/loss
+ 5: 1
+ 6:
+ - 1
+ - 1: train/grad_norm
+ 5: 1
+ 6:
+ - 1
+ - 1: train/learning_rate
+ 5: 1
+ 6:
+ - 1
+ - 1: train/epoch
+ 5: 1
+ 6:
+ - 1
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.0.dev0
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./pyg1m2.3
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: true
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: steps
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 1
+per_device_eval_batch_size:
+ desc: null
+ value: 1
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 1
+eval_accumulation_steps:
+ desc: null
+ value: 1
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 5.0e-05
+weight_decay:
+ desc: null
+ value: 0.03
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.999
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 1.0
+num_train_epochs:
+ desc: null
+ value: 2
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value: {}
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 569
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./pyg1m2.3/runs/May12_10-23-28_a9815c79a99f
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.03333333333333333
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: true
+eval_steps:
+ desc: null
+ value: 0.015
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./pyg1m2.3
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value:
+ - full_shard
+ - auto_wrap
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: true
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: false
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_sharding_strategy: FULL_SHARD
+ fsdp_forward_prefetch: FORWARD_PRE
+ fsdp_backward_prefetch: BACKWARD_PRE
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: null
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_torch
+optim_args:
+ desc: null
+ value: adam_beta1=0.98,adam_beta2=0.97,amsgrad=True
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: null
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.3
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: null
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: false
+sample_packing_efficiency:
+ desc: null
+ value: 0.92
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 1
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240512_102340-g9bmlvdf/files/output.log b/wandb/run-20240512_102340-g9bmlvdf/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..0f9ad11bebcf9a16cb74851548e27ad3691adc5b
--- /dev/null
+++ b/wandb/run-20240512_102340-g9bmlvdf/files/output.log
@@ -0,0 +1,54 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/12382 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[rank0]:[2024-05-12 10:23:54,974] [22/0] torch._dynamo.variables.higher_order_ops: [WARNING] speculate_subgraph: while introspecting the user-defined autograd.Function, we were unable to trace function `trampoline_autograd_fwd` into a single graph. This means that Dynamo was unable to prove safety for this API and will fall back to eager-mode PyTorch, which could lead to a slowdown.
+[rank0]:[2024-05-12 10:23:54,975] [22/0] torch._dynamo.variables.higher_order_ops: [ERROR] Tensor.data_ptr
+[rank0]:[2024-05-12 10:23:54,989] [23/0] torch._dynamo.variables.higher_order_ops: [WARNING] speculate_subgraph: while introspecting the user-defined autograd.Function, we were unable to trace function `trampoline_autograd_fwd` into a single graph. This means that Dynamo was unable to prove safety for this API and will fall back to eager-mode PyTorch, which could lead to a slowdown.
+[rank0]:[2024-05-12 10:23:54,989] [23/0] torch._dynamo.variables.higher_order_ops: [ERROR] Tensor.data_ptr
+[2024-05-12 10:23:41,802] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:41518] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-12 10:23:42,956] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:41518] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47141994
+[2024-05-12 10:23:44,000] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:41518] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47141994
+{'loss': 2.1033, 'grad_norm': 21.75, 'learning_rate': 8.787346221441124e-08, 'epoch': 0.0}
+ 0%| | 1/12382 [00:34<115:27:20, 33.57s/it]Traceback (most recent call last):
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/runpy.py", line 196, in _run_module_as_main
+ return _run_code(code, main_globals, None,
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/runpy.py", line 86, in _run_code
+ exec(code, run_globals)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 70, in
+ fire.Fire(do_cli)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 143, in Fire
+ component_trace = _Fire(component, args, parsed_flag_args, context, name)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 477, in _Fire
+ component, remaining_args = _CallAndUpdateTrace(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 693, in _CallAndUpdateTrace
+ component = fn(*varargs, **kwargs)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 38, in do_cli
+ return do_train(parsed_cfg, parsed_cli_args)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 66, in do_train
+ return train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ File "/workspace/disk1/axolotl/src/axolotl/train.py", line 170, in train
+ trainer.train(resume_from_checkpoint=resume_from_checkpoint)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 1828, in train
+ return inner_training_loop(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 2256, in _inner_training_loop
+ self._maybe_log_save_evaluate(tr_loss, grad_norm, model, trial, epoch, ignore_keys_for_eval)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 2640, in _maybe_log_save_evaluate
+ metrics = self.evaluate(ignore_keys=ignore_keys_for_eval)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 3445, in evaluate
+ output = eval_loop(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 3624, in evaluation_loop
+ for step, inputs in enumerate(dataloader):
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/accelerate/data_loader.py", line 452, in __iter__
+ current_batch = next(dataloader_iter)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 630, in __next__
+ data = self._next_data()
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/data/dataloader.py", line 674, in _next_data
+ data = self._dataset_fetcher.fetch(index) # may raise StopIteration
+ File "/workspace/disk1/axolotl/src/axolotl/monkeypatch/data/batch_dataset_fetcher.py", line 32, in fetch
+ return self.collate_fn(data)
+ File "/workspace/disk1/axolotl/src/axolotl/utils/collators.py", line 106, in __call__
+ features = self.tokenizer.pad(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 3274, in pad
+ raise ValueError(
+ValueError: You should supply an encoding or a list of encodings to this method that includes input_ids, but you provided ['labels']
\ No newline at end of file
diff --git a/wandb/run-20240512_102340-g9bmlvdf/files/requirements.txt b/wandb/run-20240512_102340-g9bmlvdf/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..54e63626ab0d75f9dec791c2e6499b990c4392b8
--- /dev/null
+++ b/wandb/run-20240512_102340-g9bmlvdf/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash-attn==2.5.5
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240512_102340-g9bmlvdf/files/wandb-metadata.json b/wandb/run-20240512_102340-g9bmlvdf/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..6f28e6061ae2d4bb8a45f246ec933100a92305d1
--- /dev/null
+++ b/wandb/run-20240512_102340-g9bmlvdf/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-12T10:23:40.840803",
+ "startedAt": "2024-05-12T10:23:40.185257",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "2147cf6837e2b90a2ea7045262083cbb0da03858"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "a9815c79a99f",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1492.0334687499999,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1700.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1700.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1700.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3499.997,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3499.997,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2199.998,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2000.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 149.2389030456543
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240512_102340-g9bmlvdf/files/wandb-summary.json b/wandb/run-20240512_102340-g9bmlvdf/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..059eb8ea46e7fad2130adfb9e5510e134c840c02
--- /dev/null
+++ b/wandb/run-20240512_102340-g9bmlvdf/files/wandb-summary.json
@@ -0,0 +1 @@
+{"train/loss": 2.1033, "train/grad_norm": 21.75, "train/learning_rate": 8.787346221441124e-08, "train/epoch": 0.0, "train/global_step": 1, "_timestamp": 1715509455.991621, "_runtime": 35.7974419593811, "_step": 0, "_wandb": {"runtime": 35}}
\ No newline at end of file
diff --git a/wandb/run-20240512_102340-g9bmlvdf/logs/debug-internal.log b/wandb/run-20240512_102340-g9bmlvdf/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..57d5a6563509ec3b958bc741626479f6d1758899
--- /dev/null
+++ b/wandb/run-20240512_102340-g9bmlvdf/logs/debug-internal.log
@@ -0,0 +1,210 @@
+2024-05-12 10:23:40,195 INFO StreamThr :43319 [internal.py:wandb_internal():85] W&B internal server running at pid: 43319, started at: 2024-05-12 10:23:40.194148
+2024-05-12 10:23:40,197 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: status
+2024-05-12 10:23:40,205 INFO WriterThread:43319 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/run-g9bmlvdf.wandb
+2024-05-12 10:23:40,205 DEBUG SenderThread:43319 [sender.py:send():378] send: header
+2024-05-12 10:23:40,215 DEBUG SenderThread:43319 [sender.py:send():378] send: run
+2024-05-12 10:23:40,580 INFO SenderThread:43319 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files
+2024-05-12 10:23:40,581 INFO SenderThread:43319 [sender.py:_start_run_threads():1123] run started: g9bmlvdf with start time 1715509420.194179
+2024-05-12 10:23:40,609 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: check_version
+2024-05-12 10:23:40,609 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: check_version
+2024-05-12 10:23:40,696 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: run_start
+2024-05-12 10:23:40,732 DEBUG HandlerThread:43319 [system_info.py:__init__():26] System info init
+2024-05-12 10:23:40,732 DEBUG HandlerThread:43319 [system_info.py:__init__():41] System info init done
+2024-05-12 10:23:40,732 INFO HandlerThread:43319 [system_monitor.py:start():194] Starting system monitor
+2024-05-12 10:23:40,733 INFO SystemMonitor:43319 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-12 10:23:40,733 INFO HandlerThread:43319 [system_monitor.py:probe():214] Collecting system info
+2024-05-12 10:23:40,733 INFO SystemMonitor:43319 [interfaces.py:start():188] Started cpu monitoring
+2024-05-12 10:23:40,734 INFO SystemMonitor:43319 [interfaces.py:start():188] Started disk monitoring
+2024-05-12 10:23:40,736 INFO SystemMonitor:43319 [interfaces.py:start():188] Started gpu monitoring
+2024-05-12 10:23:40,737 INFO SystemMonitor:43319 [interfaces.py:start():188] Started memory monitoring
+2024-05-12 10:23:40,738 INFO SystemMonitor:43319 [interfaces.py:start():188] Started network monitoring
+2024-05-12 10:23:40,840 DEBUG HandlerThread:43319 [system_info.py:probe():150] Probing system
+2024-05-12 10:23:40,844 DEBUG HandlerThread:43319 [system_info.py:_probe_git():135] Probing git
+2024-05-12 10:23:40,859 DEBUG HandlerThread:43319 [system_info.py:_probe_git():143] Probing git done
+2024-05-12 10:23:40,859 DEBUG HandlerThread:43319 [system_info.py:probe():198] Probing system done
+2024-05-12 10:23:40,859 DEBUG HandlerThread:43319 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-12T10:23:40.840803', 'startedAt': '2024-05-12T10:23:40.185257', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '2147cf6837e2b90a2ea7045262083cbb0da03858'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'a9815c79a99f', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1492.0334687499999, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 1700.0, 'min': 800.0, 'max': 3500.0}, {'current': 1500.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 900.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1700.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 1700.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 1400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3499.997, 'min': 800.0, 'max': 3500.0}, {'current': 3499.997, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 1200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 1100.0, 'min': 800.0, 'max': 3500.0}, {'current': 1100.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2199.998, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2000.0, 'min': 800.0, 'max': 3500.0}, {'current': 1800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1500.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 1200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 1100.0, 'min': 800.0, 'max': 3500.0}, {'current': 1400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 1200.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 1900.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1200.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 149.2389030456543}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-12 10:23:40,859 INFO HandlerThread:43319 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-12 10:23:40,859 INFO HandlerThread:43319 [system_monitor.py:probe():227] Publishing system info
+2024-05-12 10:23:40,859 DEBUG HandlerThread:43319 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-12 10:23:41,325 DEBUG HandlerThread:43319 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-12 10:23:41,325 DEBUG HandlerThread:43319 [system_info.py:_save_code():44] Saving code
+2024-05-12 10:23:41,326 WARNING HandlerThread:43319 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-12 10:23:41,326 DEBUG HandlerThread:43319 [system_info.py:_save_patches():82] Saving git patches
+2024-05-12 10:23:41,361 DEBUG HandlerThread:43319 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-12 10:23:41,364 INFO HandlerThread:43319 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-12 10:23:41,372 DEBUG SenderThread:43319 [sender.py:send():378] send: files
+2024-05-12 10:23:41,372 INFO SenderThread:43319 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-12 10:23:41,583 INFO Thread-12 :43319 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/conda-environment.yaml
+2024-05-12 10:23:41,583 INFO Thread-12 :43319 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/wandb-metadata.json
+2024-05-12 10:23:41,601 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-12 10:23:41,602 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-12 10:23:41,602 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: python_packages
+2024-05-12 10:23:41,604 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: stop_status
+2024-05-12 10:23:41,799 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-12 10:23:42,012 INFO wandb-upload_0:43319 [upload_job.py:push():130] Uploaded file /tmp/tmpqfwr5phhwandb/2davjhut-wandb-metadata.json
+2024-05-12 10:23:42,582 INFO Thread-12 :43319 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/requirements.txt
+2024-05-12 10:23:42,582 INFO Thread-12 :43319 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/tmp/axolotl_config_kx6m3r0w.yml
+2024-05-12 10:23:42,583 INFO Thread-12 :43319 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/tmp
+2024-05-12 10:23:46,864 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-12 10:23:51,873 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-12 10:23:56,874 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-12 10:24:01,909 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-12 10:24:03,094 DEBUG SenderThread:43319 [sender.py:send():378] send: telemetry
+2024-05-12 10:24:03,095 DEBUG SenderThread:43319 [sender.py:send():378] send: config
+2024-05-12 10:24:03,097 DEBUG SenderThread:43319 [sender.py:send():378] send: metric
+2024-05-12 10:24:03,097 DEBUG SenderThread:43319 [sender.py:send():378] send: telemetry
+2024-05-12 10:24:03,097 DEBUG SenderThread:43319 [sender.py:send():378] send: telemetry
+2024-05-12 10:24:03,097 DEBUG SenderThread:43319 [sender.py:send():378] send: metric
+2024-05-12 10:24:03,098 WARNING SenderThread:43319 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-12 10:24:03,098 DEBUG SenderThread:43319 [sender.py:send():378] send: telemetry
+2024-05-12 10:24:03,098 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: log_artifact
+2024-05-12 10:24:03,099 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:24:04,358 INFO wandb-upload_0:43319 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmpln0owjlr
+2024-05-12 10:24:05,076 INFO SenderThread:43319 [sender.py:send_request_log_artifact():1455] logged artifact config-g9bmlvdf - {'id': 'QXJ0aWZhY3Q6ODMyOTEwNzI2', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3MzY0Njg3NQ==', 'latestArtifact': None}}
+2024-05-12 10:24:05,078 DEBUG SenderThread:43319 [sender.py:send():378] send: telemetry
+2024-05-12 10:24:05,078 DEBUG SenderThread:43319 [sender.py:send():378] send: files
+2024-05-12 10:24:05,079 INFO SenderThread:43319 [sender.py:_save_file():1389] saving file tmp/axolotl_config_kx6m3r0w.yml with policy live
+2024-05-12 10:24:05,525 INFO wandb-upload_0:43319 [upload_job.py:push():130] Uploaded file /tmp/tmpqfwr5phhwandb/5mfbffv1-tmp/axolotl_config_kx6m3r0w.yml
+2024-05-12 10:24:05,593 INFO Thread-12 :43319 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/output.log
+2024-05-12 10:24:06,911 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-12 10:24:07,595 INFO Thread-12 :43319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/output.log
+2024-05-12 10:24:09,087 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:24:11,914 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-12 10:24:14,094 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:24:14,600 INFO Thread-12 :43319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/config.yaml
+2024-05-12 10:24:15,992 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-12 10:24:15,997 DEBUG SenderThread:43319 [sender.py:send():378] send: metric
+2024-05-12 10:24:15,997 DEBUG SenderThread:43319 [sender.py:send():378] send: metric
+2024-05-12 10:24:15,998 DEBUG SenderThread:43319 [sender.py:send():378] send: metric
+2024-05-12 10:24:15,998 DEBUG SenderThread:43319 [sender.py:send():378] send: metric
+2024-05-12 10:24:15,998 DEBUG SenderThread:43319 [sender.py:send():378] send: history
+2024-05-12 10:24:15,998 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: summary_record
+2024-05-12 10:24:16,001 INFO SenderThread:43319 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-12 10:24:16,420 DEBUG SenderThread:43319 [sender.py:send():378] send: exit
+2024-05-12 10:24:16,420 INFO SenderThread:43319 [sender.py:send_exit():585] handling exit code: 1
+2024-05-12 10:24:16,420 INFO SenderThread:43319 [sender.py:send_exit():587] handling runtime: 35
+2024-05-12 10:24:16,421 INFO SenderThread:43319 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-12 10:24:16,421 INFO SenderThread:43319 [sender.py:send_exit():593] send defer
+2024-05-12 10:24:16,421 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:16,422 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 0
+2024-05-12 10:24:16,422 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:16,422 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 0
+2024-05-12 10:24:16,422 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 1
+2024-05-12 10:24:16,422 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:16,422 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 1
+2024-05-12 10:24:16,422 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:16,422 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 1
+2024-05-12 10:24:16,422 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 2
+2024-05-12 10:24:16,422 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:16,422 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 2
+2024-05-12 10:24:16,422 INFO HandlerThread:43319 [system_monitor.py:finish():203] Stopping system monitor
+2024-05-12 10:24:16,423 INFO HandlerThread:43319 [interfaces.py:finish():200] Joined cpu monitor
+2024-05-12 10:24:16,423 DEBUG SystemMonitor:43319 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-12 10:24:16,423 INFO HandlerThread:43319 [interfaces.py:finish():200] Joined disk monitor
+2024-05-12 10:24:16,424 DEBUG SystemMonitor:43319 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-12 10:24:16,424 DEBUG SystemMonitor:43319 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-12 10:24:16,601 INFO Thread-12 :43319 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/wandb-summary.json
+2024-05-12 10:24:16,608 INFO HandlerThread:43319 [interfaces.py:finish():200] Joined gpu monitor
+2024-05-12 10:24:16,608 INFO HandlerThread:43319 [interfaces.py:finish():200] Joined memory monitor
+2024-05-12 10:24:16,608 INFO HandlerThread:43319 [interfaces.py:finish():200] Joined network monitor
+2024-05-12 10:24:16,609 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:16,609 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 2
+2024-05-12 10:24:16,609 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 3
+2024-05-12 10:24:16,609 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:16,609 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 3
+2024-05-12 10:24:16,610 DEBUG SenderThread:43319 [sender.py:send():378] send: stats
+2024-05-12 10:24:16,612 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:16,613 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 3
+2024-05-12 10:24:16,613 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 4
+2024-05-12 10:24:16,613 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:16,613 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 4
+2024-05-12 10:24:16,613 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:16,613 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 4
+2024-05-12 10:24:16,614 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 5
+2024-05-12 10:24:16,614 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:16,614 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 5
+2024-05-12 10:24:16,614 DEBUG SenderThread:43319 [sender.py:send():378] send: summary
+2024-05-12 10:24:16,615 INFO SenderThread:43319 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-12 10:24:16,616 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:16,616 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 5
+2024-05-12 10:24:16,616 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 6
+2024-05-12 10:24:16,616 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:16,616 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 6
+2024-05-12 10:24:16,617 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:16,617 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 6
+2024-05-12 10:24:16,622 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:24:16,940 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 7
+2024-05-12 10:24:16,940 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:16,940 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 7
+2024-05-12 10:24:16,941 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:16,941 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 7
+2024-05-12 10:24:17,420 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:17,602 INFO Thread-12 :43319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/wandb-summary.json
+2024-05-12 10:24:17,602 INFO Thread-12 :43319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/config.yaml
+2024-05-12 10:24:17,603 INFO Thread-12 :43319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/output.log
+2024-05-12 10:24:19,116 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 8
+2024-05-12 10:24:19,117 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:19,117 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:19,118 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 8
+2024-05-12 10:24:19,118 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:19,118 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 8
+2024-05-12 10:24:19,118 INFO SenderThread:43319 [job_builder.py:build():432] Attempting to build job artifact
+2024-05-12 10:24:19,122 INFO SenderThread:43319 [job_builder.py:_get_source_type():565] is repo sourced job
+2024-05-12 10:24:19,122 WARNING SenderThread:43319 [job_builder.py:_log_if_verbose():267] No program path found, not creating job artifact. See https://docs.wandb.ai/guides/launch/create-job
+2024-05-12 10:24:19,123 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 9
+2024-05-12 10:24:19,123 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:19,123 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 9
+2024-05-12 10:24:19,123 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:19,123 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 9
+2024-05-12 10:24:19,123 INFO SenderThread:43319 [dir_watcher.py:finish():358] shutting down directory watcher
+2024-05-12 10:24:19,421 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:19,602 INFO Thread-12 :43319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/output.log
+2024-05-12 10:24:19,604 INFO SenderThread:43319 [dir_watcher.py:finish():388] scan: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files
+2024-05-12 10:24:19,604 INFO SenderThread:43319 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/wandb-metadata.json wandb-metadata.json
+2024-05-12 10:24:19,605 INFO SenderThread:43319 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/wandb-summary.json wandb-summary.json
+2024-05-12 10:24:19,605 INFO SenderThread:43319 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/requirements.txt requirements.txt
+2024-05-12 10:24:19,611 INFO SenderThread:43319 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/config.yaml config.yaml
+2024-05-12 10:24:19,617 INFO SenderThread:43319 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/conda-environment.yaml conda-environment.yaml
+2024-05-12 10:24:19,623 INFO SenderThread:43319 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/output.log output.log
+2024-05-12 10:24:19,628 INFO SenderThread:43319 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/tmp/axolotl_config_kx6m3r0w.yml tmp/axolotl_config_kx6m3r0w.yml
+2024-05-12 10:24:19,628 INFO SenderThread:43319 [sender.py:transition_state():613] send defer: 10
+2024-05-12 10:24:19,629 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:19,629 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: defer
+2024-05-12 10:24:19,632 INFO HandlerThread:43319 [handler.py:handle_request_defer():184] handle defer: 10
+2024-05-12 10:24:19,632 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: defer
+2024-05-12 10:24:19,633 INFO SenderThread:43319 [sender.py:send_request_defer():609] handle sender defer: 10
+2024-05-12 10:24:19,633 INFO SenderThread:43319 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-12 10:24:20,036 INFO wandb-upload_0:43319 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/wandb-summary.json
+2024-05-12 10:24:20,098 INFO wandb-upload_1:43319 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/requirements.txt
+2024-05-12 10:24:20,119 INFO wandb-upload_2:43319 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/config.yaml
+2024-05-12 10:24:20,152 INFO wandb-upload_4:43319 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/output.log
+2024-05-12 10:24:20,423 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:20,424 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:21,423 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:21,424 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:22,426 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:22,427 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:23,426 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:23,427 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:24,428 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:24,429 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:25,428 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:25,429 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:25,430 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: status_report
+2024-05-12 10:24:26,429 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:26,430 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:27,430 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:27,431 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:28,431 DEBUG HandlerThread:43319 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-12 10:24:28,432 DEBUG SenderThread:43319 [sender.py:send_request():405] send_request: poll_exit
+2024-05-12 10:24:29,250 WARNING StreamThr :43319 [internal.py:is_dead():413] Internal process exiting, parent pid 41518 disappeared
+2024-05-12 10:24:29,251 ERROR StreamThr :43319 [internal.py:wandb_internal():151] Internal process shutdown.
+2024-05-12 10:24:29,431 INFO HandlerThread:43319 [handler.py:finish():882] shutting down handler
+2024-05-12 10:24:29,432 INFO WriterThread:43319 [datastore.py:close():296] close: /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/run-g9bmlvdf.wandb
+2024-05-12 10:24:29,432 INFO SenderThread:43319 [sender.py:finish():1545] shutting down sender
+2024-05-12 10:24:29,433 INFO SenderThread:43319 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-12 10:24:29,433 INFO SenderThread:43319 [file_pusher.py:join():175] waiting for file pusher
+2024-05-12 10:24:41,197 INFO wandb-upload_3:43319 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/files/conda-environment.yaml
+2024-05-12 10:24:41,397 INFO Thread-11 (_thread_body):43319 [sender.py:transition_state():613] send defer: 11
+2024-05-12 10:24:41,448 INFO SenderThread:43319 [file_stream.py:finish():601] file stream finish called
+2024-05-12 10:24:41,611 INFO SenderThread:43319 [file_stream.py:finish():605] file stream finish is done
diff --git a/wandb/run-20240512_102340-g9bmlvdf/logs/debug.log b/wandb/run-20240512_102340-g9bmlvdf/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..cdfdde46a250cd5106d1f648d8b28daba8bbf9cf
--- /dev/null
+++ b/wandb/run-20240512_102340-g9bmlvdf/logs/debug.log
@@ -0,0 +1,29 @@
+2024-05-12 10:23:40,189 INFO MainThread:41518 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-12 10:23:40,189 INFO MainThread:41518 [wandb_setup.py:_flush():76] Configure stats pid to 41518
+2024-05-12 10:23:40,189 INFO MainThread:41518 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-12 10:23:40,189 INFO MainThread:41518 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-12 10:23:40,189 INFO MainThread:41518 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-12 10:23:40,189 INFO MainThread:41518 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-12 10:23:40,190 WARNING MainThread:41518 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-12 10:23:40,190 INFO MainThread:41518 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-12 10:23:40,190 INFO MainThread:41518 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-12 10:23:40,190 INFO MainThread:41518 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/logs/debug.log
+2024-05-12 10:23:40,190 INFO MainThread:41518 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240512_102340-g9bmlvdf/logs/debug-internal.log
+2024-05-12 10:23:40,190 INFO MainThread:41518 [wandb_init.py:init():560] calling init triggers
+2024-05-12 10:23:40,190 INFO MainThread:41518 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-12 10:23:40,190 INFO MainThread:41518 [wandb_init.py:init():610] starting backend
+2024-05-12 10:23:40,190 INFO MainThread:41518 [wandb_init.py:init():614] setting up manager
+2024-05-12 10:23:40,192 INFO MainThread:41518 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-12 10:23:40,193 INFO MainThread:41518 [wandb_init.py:init():622] backend started and connected
+2024-05-12 10:23:40,204 INFO MainThread:41518 [wandb_init.py:init():711] updated telemetry
+2024-05-12 10:23:40,214 INFO MainThread:41518 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-12 10:23:40,607 INFO MainThread:41518 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-12 10:23:40,684 INFO MainThread:41518 [wandb_run.py:_on_init():2405] got version response
+2024-05-12 10:23:40,685 INFO MainThread:41518 [wandb_init.py:init():795] starting run threads in backend
+2024-05-12 10:23:41,602 INFO MainThread:41518 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-12 10:23:41,602 INFO MainThread:41518 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-12 10:23:41,602 INFO MainThread:41518 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-12 10:23:41,602 INFO MainThread:41518 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-12 10:23:41,604 INFO MainThread:41518 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-12 10:23:41,607 INFO MainThread:41518 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyg1m2.3', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 1, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.03, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 2, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 569, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyg1m2.3/runs/May12_10-23-28_a9815c79a99f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.03333333333333333, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': True, 'eval_steps': 0.015, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyg1m2.3', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': ['full_shard', 'auto_wrap'], 'fsdp_min_num_params': 0, 'fsdp_config': {'fsdp_limit_all_gathers': True, 'fsdp_sync_module_states': True, 'fsdp_offload_params': True, 'fsdp_use_orig_params': False, 'fsdp_cpu_ram_efficient_loading': False, 'fsdp_auto_wrap_policy': 'TRANSFORMER_BASED_WRAP', 'fsdp_transformer_layer_cls_to_wrap': 'LlamaDecoderLayer', 'fsdp_state_dict_type': 'FULL_STATE_DICT', 'fsdp_sharding_strategy': 'FULL_SHARD', 'fsdp_forward_prefetch': 'FORWARD_PRE', 'fsdp_backward_prefetch': 'BACKWARD_PRE', 'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': 'adam_beta1=0.98,adam_beta2=0.97,amsgrad=True', 'adafactor': False, 'group_by_length': None, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.3', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': False, 'sample_packing_efficiency': 0.92, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 1, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
diff --git a/wandb/run-20240512_102340-g9bmlvdf/run-g9bmlvdf.wandb b/wandb/run-20240512_102340-g9bmlvdf/run-g9bmlvdf.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..a10b8660f80b339fa7c0d9da253efec65d4f3002
Binary files /dev/null and b/wandb/run-20240512_102340-g9bmlvdf/run-g9bmlvdf.wandb differ
diff --git a/wandb/run-20240512_103207-vfqtzg1a/files/conda-environment.yaml b/wandb/run-20240512_103207-vfqtzg1a/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240512_103207-vfqtzg1a/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240512_103207-vfqtzg1a/files/config.yaml b/wandb/run-20240512_103207-vfqtzg1a/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..25067a81d2e1fb13bf2a49a8d096ef143bd903e9
--- /dev/null
+++ b/wandb/run-20240512_103207-vfqtzg1a/files/config.yaml
@@ -0,0 +1,3096 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715509927
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 1
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+ - 1: train/loss
+ 5: 1
+ 6:
+ - 1
+ - 1: train/grad_norm
+ 5: 1
+ 6:
+ - 1
+ - 1: train/learning_rate
+ 5: 1
+ 6:
+ - 1
+ - 1: train/epoch
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.embed_tokens\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.embed_tokens\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.embed_tokens\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\._fsdp_wrapped_module\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.embed_tokens\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.embed_tokens\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.embed_tokens\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.0\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.1\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.2\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.3\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.4\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.5\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.6\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.7\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.8\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.9\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.10\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.11\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.12\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.13\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.14\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.15\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.16\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.17\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.18\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.19\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.20\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.21\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.22\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.23\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.24\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.25\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.26\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.27\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.28\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.29\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.30\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/_orig_mod\.model\.layers\.31\._fsdp_wrapped_module\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.0.dev0
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./buzz-3b-small-v0.6
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: false
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: 'no'
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 1
+per_device_eval_batch_size:
+ desc: null
+ value: 1
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 2
+eval_accumulation_steps:
+ desc: null
+ value: 2
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 5.0e-05
+weight_decay:
+ desc: null
+ value: 0.01
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.95
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 0.6
+num_train_epochs:
+ desc: null
+ value: 3
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value:
+ num_cycles: 3
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 287
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./buzz-3b-small-v0.6/runs/May12_10-31-45_a9815c79a99f
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.022222222222222223
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: false
+eval_steps:
+ desc: null
+ value: null
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./buzz-3b-small-v0.6
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value:
+ - full_shard
+ - auto_wrap
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ fsdp_limit_all_gathers: true
+ fsdp_sync_module_states: true
+ fsdp_offload_params: true
+ fsdp_use_orig_params: false
+ fsdp_cpu_ram_efficient_loading: false
+ fsdp_auto_wrap_policy: TRANSFORMER_BASED_WRAP
+ fsdp_transformer_layer_cls_to_wrap: LlamaDecoderLayer
+ fsdp_state_dict_type: FULL_STATE_DICT
+ fsdp_sharding_strategy: FULL_SHARD
+ fsdp_forward_prefetch: FORWARD_PRE
+ fsdp_backward_prefetch: BACKWARD_PRE
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: null
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_torch
+optim_args:
+ desc: null
+ value: null
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: false
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.3
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: 15.0
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: true
+sample_packing_efficiency:
+ desc: null
+ value: 0.92
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 1
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240512_103207-vfqtzg1a/files/output.log b/wandb/run-20240512_103207-vfqtzg1a/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..6763c6141752405962e4d9183943e4529e7a10bd
--- /dev/null
+++ b/wandb/run-20240512_103207-vfqtzg1a/files/output.log
@@ -0,0 +1,13113 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/9378 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+[2024-05-12 10:32:09,607] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:44754] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-12 10:32:10,720] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:44754] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+[2024-05-12 10:32:11,678] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:44754] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[rank0]:[2024-05-12 10:32:22,463] [11/0] torch._dynamo.variables.higher_order_ops: [WARNING] speculate_subgraph: while introspecting the user-defined autograd.Function, we were unable to trace function `trampoline_autograd_fwd` into a single graph. This means that Dynamo was unable to prove safety for this API and will fall back to eager-mode PyTorch, which could lead to a slowdown.
+[rank0]:[2024-05-12 10:32:22,463] [11/0] torch._dynamo.variables.higher_order_ops: [ERROR] Tensor.data_ptr
+[rank0]:[2024-05-12 10:32:22,478] [12/0] torch._dynamo.variables.higher_order_ops: [WARNING] speculate_subgraph: while introspecting the user-defined autograd.Function, we were unable to trace function `trampoline_autograd_fwd` into a single graph. This means that Dynamo was unable to prove safety for this API and will fall back to eager-mode PyTorch, which could lead to a slowdown.
+[rank0]:[2024-05-12 10:32:22,478] [12/0] torch._dynamo.variables.higher_order_ops: [ERROR] Tensor.data_ptr
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 0%| | 1/9378 [00:37<98:50:45, 37.95s/it]
+
+ 0%| | 1/9378 [00:38<98:50:45, 37.95s/it]
+
+ 0%| | 2/9378 [01:02<76:12:06, 29.26s/it]
+
+ 0%| | 3/9378 [01:25<70:01:49, 26.89s/it]
+
+
+ 0%| | 4/9378 [01:49<67:21:25, 25.87s/it]
+
+
+ 0%| | 5/9378 [02:13<65:55:34, 25.32s/it]
+
+
+ 0%| | 6/9378 [02:38<65:05:49, 25.01s/it]
+
+ 0%| | 6/9378 [02:39<65:05:49, 25.01s/it]
+
+
+ 0%| | 8/9378 [03:27<64:31:10, 24.79s/it]
+
+
+ 0%| | 9/9378 [03:51<64:10:15, 24.66s/it]
+
+
+ 0%| | 10/9378 [04:16<63:52:18, 24.55s/it]
+
+
+ 0%| | 11/9378 [04:40<63:42:43, 24.49s/it]
+
+ 0%| | 11/9378 [04:41<63:42:43, 24.49s/it]
+
+ 0%|▏ | 12/9378 [05:05<63:36:59, 24.45s/it]
+
+
+ 0%|▏ | 14/9378 [05:53<63:35:07, 24.45s/it]
+
+
+ 0%|▏ | 15/9378 [06:18<63:48:17, 24.53s/it]
+
+
+ 0%|▏ | 16/9378 [06:43<64:15:36, 24.71s/it]
+
+
+ 0%|▏ | 17/9378 [07:08<64:04:20, 24.64s/it]
+
+
+ 0%|▏ | 18/9378 [07:32<63:54:15, 24.58s/it]
+
+ 0%|▏ | 18/9378 [07:33<63:54:15, 24.58s/it]
+
+
+ 0%|▏ | 20/9378 [08:22<63:49:51, 24.56s/it]
+
+ 0%|▏ | 21/9378 [08:46<63:43:10, 24.52s/it]
+
+
+ 0%|▏ | 22/9378 [09:10<63:39:19, 24.49s/it]
+
+ 0%|▏ | 22/9378 [09:11<63:39:19, 24.49s/it]
+
+ 0%|▏ | 23/9378 [09:36<63:45:33, 24.54s/it]
+
+
+ 0%|▎ | 25/9378 [10:24<63:40:50, 24.51s/it]
+
+
+ 0%|▎ | 26/9378 [10:48<63:30:56, 24.45s/it]
+
+
+ 0%|▎ | 27/9378 [11:13<63:37:39, 24.50s/it]
+
+ 0%|▎ | 27/9378 [11:14<63:37:39, 24.50s/it]
+
+
+ 0%|▎ | 29/9378 [12:03<63:33:46, 24.48s/it]
+
+ 0%|▎ | 30/9378 [12:26<63:29:32, 24.45s/it]
+
+
+ 0%|▎ | 31/9378 [12:50<63:24:46, 24.42s/it]
+
+
+ 0%|▎ | 32/9378 [13:15<63:29:19, 24.46s/it]
+
+ 0%|▎ | 32/9378 [13:16<63:29:19, 24.46s/it]
+
+
+ 0%|▎ | 34/9378 [14:04<63:42:38, 24.55s/it]
+
+
+ 0%|▎ | 35/9378 [14:28<63:35:56, 24.51s/it]
+
+
+ 0%|▍ | 36/9378 [14:53<63:31:36, 24.48s/it]
+
+ 0%|▍ | 36/9378 [14:54<63:31:36, 24.48s/it]
+
+
+ 0%|▍ | 38/9378 [15:42<63:36:29, 24.52s/it]
+
+
+ 0%|▍ | 39/9378 [16:06<63:37:37, 24.53s/it]
+
+
+ 0%|▍ | 40/9378 [16:31<63:29:10, 24.48s/it]
+
+ 0%|▍ | 40/9378 [16:32<63:29:10, 24.48s/it]
+
+
+ 0%|▍ | 42/9378 [17:21<63:28:36, 24.48s/it]
+
+ 0%|▍ | 43/9378 [17:44<63:33:29, 24.51s/it]
+
+
+ 0%|▍ | 44/9378 [18:09<63:29:31, 24.49s/it]
+
+
+ 0%|▍ | 45/9378 [18:33<63:24:43, 24.46s/it]
+
+ 0%|▍ | 45/9378 [18:34<63:24:43, 24.46s/it]
+
+
+ 1%|▍ | 47/9378 [19:23<63:21:08, 24.44s/it]
+
+ 1%|▌ | 48/9378 [19:46<63:18:29, 24.43s/it]
+
+
+ 1%|▌ | 49/9378 [20:11<63:23:52, 24.46s/it]
+
+
+ 1%|▌ | 50/9378 [20:57<79:20:57, 30.62s/it]
+
+ 1%|▌ | 51/9378 [21:21<74:31:19, 28.76s/it]
+
+ 1%|▌ | 52/9378 [21:45<71:14:00, 27.50s/it]
+
+
+ 1%|▌ | 53/9378 [22:09<68:50:46, 26.58s/it]
+
+ 1%|▌ | 53/9378 [22:10<68:50:46, 26.58s/it]
+
+
+ 1%|▌ | 55/9378 [23:00<66:06:53, 25.53s/it]
+
+ 1%|▌ | 56/9378 [23:23<65:23:20, 25.25s/it]
+
+
+ 1%|▌ | 57/9378 [23:47<64:36:33, 24.95s/it]
+
+
+ 1%|▌ | 58/9378 [24:12<64:09:57, 24.79s/it]
+
+ 1%|▌ | 58/9378 [24:13<64:09:57, 24.79s/it]
+
+
+ 1%|▋ | 60/9378 [25:01<63:51:49, 24.67s/it]
+
+
+ 1%|▋ | 61/9378 [25:25<63:40:53, 24.61s/it]
+
+
+ 1%|▋ | 62/9378 [25:50<63:35:59, 24.58s/it]
+
+
+ 1%|▋ | 63/9378 [26:16<63:42:37, 24.62s/it]
+
+ 1%|▋ | 64/9378 [26:39<63:36:02, 24.58s/it]
+
+
+ 1%|▋ | 65/9378 [27:03<63:24:36, 24.51s/it]
+
+
+ 1%|▋ | 66/9378 [27:28<63:23:46, 24.51s/it]
+
+
+ 1%|▋ | 67/9378 [27:52<63:12:54, 24.44s/it]
+
+
+ 1%|▋ | 68/9378 [28:17<63:12:20, 24.44s/it]
+
+
+ 1%|▋ | 69/9378 [28:42<63:26:08, 24.53s/it]
+
+ 1%|▋ | 70/9378 [29:06<63:24:35, 24.52s/it]
+
+
+ 1%|▋ | 71/9378 [29:30<63:24:37, 24.53s/it]
+
+
+ 1%|▊ | 72/9378 [29:55<63:18:30, 24.49s/it]
+
+
+ 1%|▊ | 73/9378 [30:21<63:28:53, 24.56s/it]
+
+ 1%|▊ | 74/9378 [30:44<63:25:36, 24.54s/it]
+
+
+ 1%|▊ | 75/9378 [31:08<63:24:21, 24.54s/it]
+
+ 1%|▊ | 75/9378 [31:10<63:24:21, 24.54s/it]
+
+
+ 1%|▊ | 77/9378 [31:58<63:12:27, 24.46s/it]
+
+ 1%|▊ | 78/9378 [32:23<63:11:22, 24.46s/it]
+
+ 1%|▊ | 79/9378 [32:46<63:13:48, 24.48s/it]
+
+ 1%|▊ | 79/9378 [32:47<63:13:48, 24.48s/it]
+
+
+ 1%|▊ | 81/9378 [33:37<63:35:21, 24.62s/it]
+
+ 1%|▊ | 82/9378 [34:00<63:22:05, 24.54s/it]
+
+
+ 1%|▉ | 83/9378 [34:25<63:12:38, 24.48s/it]
+
+
+ 1%|▉ | 84/9378 [34:49<63:06:00, 24.44s/it]
+
+ 1%|▉ | 84/9378 [34:50<63:06:00, 24.44s/it]
+
+
+ 1%|▉ | 86/9378 [35:38<63:21:53, 24.55s/it]
+
+
+ 1%|▉ | 87/9378 [36:03<63:15:23, 24.51s/it]
+
+ 1%|▉ | 87/9378 [36:04<63:15:23, 24.51s/it]
+
+
+ 1%|▉ | 89/9378 [36:53<63:25:45, 24.58s/it]
+
+ 1%|▉ | 90/9378 [37:17<63:38:25, 24.67s/it]
+
+
+ 1%|▉ | 91/9378 [37:41<63:25:46, 24.59s/it]
+
+
+ 1%|▉ | 92/9378 [38:07<63:27:44, 24.60s/it]
+
+ 1%|▉ | 93/9378 [38:31<63:13:57, 24.52s/it]
+
+ 1%|▉ | 94/9378 [38:55<63:22:26, 24.57s/it]
+
+
+ 1%|█ | 95/9378 [39:19<63:15:22, 24.53s/it]
+
+ 1%|█ | 95/9378 [39:20<63:15:22, 24.53s/it]
+
+
+ 1%|█ | 97/9378 [40:10<63:15:24, 24.54s/it]
+
+ 1%|█ | 98/9378 [40:33<63:10:27, 24.51s/it]
+
+
+ 1%|█ | 99/9378 [40:57<63:08:54, 24.50s/it]
+
+
+ 1%|█ | 100/9378 [41:43<79:29:30, 30.84s/it]
+
+
+ 1%|█ | 101/9378 [42:07<74:33:25, 28.93s/it]
+
+
+ 1%|█ | 102/9378 [42:33<71:20:29, 27.69s/it]
+
+ 1%|█ | 103/9378 [42:57<69:01:35, 26.79s/it]
+
+
+ 1%|█ | 104/9378 [43:21<67:13:56, 26.10s/it]
+
+
+ 1%|█ | 105/9378 [43:46<65:58:40, 25.61s/it]
+
+
+ 1%|█ | 106/9378 [44:11<65:05:13, 25.27s/it]
+
+ 1%|█ | 107/9378 [44:35<64:43:35, 25.13s/it]
+
+
+ 1%|█▏ | 108/9378 [45:00<64:30:57, 25.05s/it]
+
+
+ 1%|█▏ | 109/9378 [45:26<64:06:44, 24.90s/it]
+
+ 1%|█▏ | 110/9378 [45:49<63:46:51, 24.77s/it]
+
+
+ 1%|█▏ | 111/9378 [46:14<63:44:27, 24.76s/it]
+
+ 1%|█▏ | 111/9378 [46:15<63:44:27, 24.76s/it]
+
+
+ 1%|█▏ | 113/9378 [47:04<63:26:29, 24.65s/it]
+
+ 1%|█▏ | 114/9378 [47:27<63:21:37, 24.62s/it]
+
+
+ 1%|█▏ | 115/9378 [47:52<63:14:10, 24.58s/it]
+
+
+ 1%|█▏ | 116/9378 [48:18<63:17:57, 24.60s/it]
+
+ 1%|█▏ | 117/9378 [48:42<63:10:34, 24.56s/it]
+
+ 1%|█▏ | 118/9378 [49:05<63:08:23, 24.55s/it]
+
+
+ 1%|█▏ | 119/9378 [49:30<63:11:46, 24.57s/it]
+
+ 1%|█▏ | 119/9378 [49:31<63:11:46, 24.57s/it]
+
+
+ 1%|█▎ | 121/9378 [50:20<62:57:42, 24.49s/it]
+
+ 1%|█▎ | 122/9378 [50:44<63:12:32, 24.58s/it]
+
+
+ 1%|█▎ | 123/9378 [51:08<63:05:46, 24.54s/it]
+
+
+ 1%|█▎ | 124/9378 [51:34<63:09:30, 24.57s/it]
+
+ 1%|█▎ | 125/9378 [51:58<63:01:59, 24.52s/it]
+
+ 1%|█▎ | 126/9378 [52:22<63:27:11, 24.69s/it]
+
+
+ 1%|█▎ | 127/9378 [52:46<62:53:03, 24.47s/it]
+
+ 1%|█▎ | 127/9378 [52:47<62:53:03, 24.47s/it]
+
+
+ 1%|█▎ | 129/9378 [53:36<62:47:12, 24.44s/it]
+
+ 1%|█▎ | 130/9378 [54:00<62:49:34, 24.46s/it]
+
+
+ 1%|█▎ | 131/9378 [54:24<62:49:22, 24.46s/it]
+
+
+ 1%|█▍ | 132/9378 [54:49<62:50:46, 24.47s/it]
+
+ 1%|█▍ | 132/9378 [54:50<62:50:46, 24.47s/it]
+
+
+ 1%|█▍ | 134/9378 [55:39<62:44:01, 24.43s/it]
+
+ 1%|█▍ | 135/9378 [56:02<62:51:42, 24.48s/it]
+
+
+ 1%|█▍ | 136/9378 [56:26<62:49:00, 24.47s/it]
+
+
+ 1%|█▍ | 137/9378 [56:51<62:53:44, 24.50s/it]
+
+ 1%|█▍ | 137/9378 [56:52<62:53:44, 24.50s/it]
+
+
+ 1%|█▍ | 139/9378 [57:40<62:55:42, 24.52s/it]
+
+
+ 1%|█▍ | 140/9378 [58:05<63:07:47, 24.60s/it]
+
+ 1%|█▍ | 140/9378 [58:06<63:07:47, 24.60s/it]
+
+
+ 2%|█▍ | 142/9378 [58:55<63:01:48, 24.57s/it]
+
+ 2%|█▍ | 143/9378 [59:18<62:55:24, 24.53s/it]
+
+
+ 2%|█▌ | 144/9378 [59:43<62:51:13, 24.50s/it]
+
+
+ 2%|█▍ | 145/9378 [1:00:07<62:45:26, 24.47s/it]
+
+ 2%|█▍ | 145/9378 [1:00:08<62:45:26, 24.47s/it]
+
+
+ 2%|█▌ | 147/9378 [1:00:57<62:46:23, 24.48s/it]
+
+ 2%|█▌ | 148/9378 [1:01:21<62:44:55, 24.47s/it]
+
+
+ 2%|█▌ | 149/9378 [1:01:45<62:43:08, 24.47s/it]
+
+
+ 2%|█▌ | 150/9378 [1:02:29<77:55:29, 30.40s/it]
+
+ 2%|█▌ | 150/9378 [1:02:30<77:55:29, 30.40s/it]
+
+
+ 2%|█▌ | 152/9378 [1:03:19<70:08:24, 27.37s/it]
+
+ 2%|█▌ | 153/9378 [1:03:43<67:53:36, 26.49s/it]
+
+
+ 2%|█▌ | 154/9378 [1:04:07<66:18:39, 25.88s/it]
+
+ 2%|█▌ | 154/9378 [1:04:08<66:18:39, 25.88s/it]
+
+
+ 2%|█▌ | 156/9378 [1:04:57<64:45:31, 25.28s/it]
+
+
+ 2%|█▌ | 157/9378 [1:05:21<64:10:45, 25.06s/it]
+
+
+ 2%|█▌ | 158/9378 [1:05:45<63:38:38, 24.85s/it]
+
+ 2%|█▌ | 158/9378 [1:05:47<63:38:38, 24.85s/it]
+
+
+ 2%|█▋ | 160/9378 [1:06:36<63:17:30, 24.72s/it]
+
+ 2%|█▋ | 161/9378 [1:06:59<63:06:00, 24.65s/it]
+
+
+ 2%|█▋ | 162/9378 [1:07:24<62:57:12, 24.59s/it]
+
+ 2%|█▋ | 162/9378 [1:07:25<62:57:12, 24.59s/it]
+
+
+ 2%|█▋ | 164/9378 [1:08:14<62:48:58, 24.54s/it]
+
+ 2%|█▋ | 165/9378 [1:08:38<62:38:07, 24.47s/it]
+
+ 2%|█▋ | 166/9378 [1:09:01<62:44:12, 24.52s/it]
+
+
+ 2%|█▋ | 167/9378 [1:09:26<62:43:28, 24.52s/it]
+
+ 2%|█▋ | 167/9378 [1:09:27<62:43:28, 24.52s/it]
+
+
+ 2%|█▋ | 169/9378 [1:10:16<62:43:25, 24.52s/it]
+
+ 2%|█▋ | 170/9378 [1:10:40<62:47:23, 24.55s/it]
+
+ 2%|█▋ | 170/9378 [1:10:41<62:47:23, 24.55s/it]
+
+ 2%|█▊ | 171/9378 [1:11:05<62:48:38, 24.56s/it]
+
+
+ 2%|█▊ | 173/9378 [1:11:53<62:51:53, 24.59s/it]
+
+
+ 2%|█▊ | 174/9378 [1:12:18<62:50:14, 24.58s/it]
+
+ 2%|█▊ | 174/9378 [1:12:19<62:50:14, 24.58s/it]
+
+
+ 2%|█▊ | 176/9378 [1:13:08<62:37:26, 24.50s/it]
+
+ 2%|█▊ | 177/9378 [1:13:31<62:42:41, 24.54s/it]
+
+
+ 2%|█▊ | 178/9378 [1:13:56<62:33:59, 24.48s/it]
+
+
+ 2%|█▊ | 179/9378 [1:14:20<62:35:37, 24.50s/it]
+
+ 2%|█▊ | 179/9378 [1:14:21<62:35:37, 24.50s/it]
+
+
+ 2%|█▊ | 181/9378 [1:15:10<62:31:56, 24.48s/it]
+
+ 2%|█▊ | 182/9378 [1:15:34<62:35:19, 24.50s/it]
+
+
+ 2%|█▊ | 183/9378 [1:15:58<62:33:47, 24.49s/it]
+
+
+ 2%|█▉ | 184/9378 [1:16:23<62:28:11, 24.46s/it]
+
+
+ 2%|█▉ | 185/9378 [1:16:47<62:26:19, 24.45s/it]
+
+
+ 2%|█▉ | 186/9378 [1:17:13<62:23:18, 24.43s/it]
+
+ 2%|█▉ | 187/9378 [1:17:37<62:27:31, 24.46s/it]
+
+ 2%|█▉ | 188/9378 [1:18:00<62:26:41, 24.46s/it]
+
+
+ 2%|█▉ | 189/9378 [1:18:25<62:29:55, 24.49s/it]
+
+
+ 2%|█▉ | 190/9378 [1:18:51<62:40:57, 24.56s/it]
+
+ 2%|█▉ | 191/9378 [1:19:15<62:29:39, 24.49s/it]
+
+ 2%|█▉ | 192/9378 [1:19:39<62:30:21, 24.50s/it]
+
+
+ 2%|█▉ | 193/9378 [1:20:03<62:31:11, 24.50s/it]
+
+
+ 2%|█▉ | 194/9378 [1:20:29<62:35:03, 24.53s/it]
+
+ 2%|█▉ | 195/9378 [1:20:53<62:36:32, 24.54s/it]
+
+ 2%|██ | 196/9378 [1:21:17<62:26:53, 24.48s/it]
+
+
+ 2%|██ | 197/9378 [1:21:41<62:31:11, 24.51s/it]
+
+
+ 2%|██ | 198/9378 [1:22:07<62:21:52, 24.46s/it]
+
+ 2%|██ | 199/9378 [1:22:31<62:32:02, 24.53s/it]
+
+ 2%|██ | 200/9378 [1:23:15<77:01:29, 30.21s/it]
+
+ 2%|██ | 201/9378 [1:23:40<72:49:12, 28.57s/it]
+
+ 2%|██ | 202/9378 [1:24:03<69:44:43, 27.36s/it]
+
+
+ 2%|██ | 203/9378 [1:24:28<67:37:03, 26.53s/it]
+
+
+ 2%|██ | 204/9378 [1:24:53<65:57:13, 25.88s/it]
+
+ 2%|██ | 205/9378 [1:25:18<64:52:49, 25.46s/it]
+
+ 2%|██ | 206/9378 [1:25:41<64:11:54, 25.20s/it]
+
+
+ 2%|██ | 207/9378 [1:26:05<63:37:48, 24.98s/it]
+
+ 2%|██ | 207/9378 [1:26:07<63:37:48, 24.98s/it]
+
+
+ 2%|██▏ | 209/9378 [1:26:56<63:02:05, 24.75s/it]
+{'loss': 1.1856, 'grad_norm': 1.2265625, 'learning_rate': 3.6411149825783976e-05, 'epoch': 0.07}
+[2024-05-12 11:59:51,609] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-209/pytorch_model_fsdp.bin
+[2024-05-12 12:00:03,506] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-209/pytorch_model_fsdp.bin
+[2024-05-12 12:00:28,162] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-209/optimizer.bin
+ 2%|██▏ | 209/9378 [1:26:56<63:02:05, 24.75s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 2%|██▏ | 210/9378 [1:29:10<147:30:13, 57.92s/it]
+
+ 2%|██▏ | 210/9378 [1:29:11<147:30:13, 57.92s/it]
+
+
+ 2%|██▏ | 212/9378 [1:30:02<105:55:22, 41.60s/it]
+
+
+ 2%|██▏ | 213/9378 [1:30:28<94:14:24, 37.02s/it]
+
+ 2%|██▏ | 213/9378 [1:30:29<94:14:24, 37.02s/it]
+[2024-05-12 12:02:56,274] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/820c8683fe64b53f608de37a3a84c76c9c1a09f866fa9525cfa29d86a4fcef8c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T120154Z&X-Amz-Expires=86400&X-Amz-Signature=649100cb85f799d5074baf84eb640627ea87b189bb03f95c01a1c77a11b28396&X-Amz-SignedHeaders=host&partNumber=1&uploadId=8cimuA1DggLlFTGl0o8.2Y3pnYT5nTsnaa_gzi73BYNB9vkFEu4exfHrPlRI66q6zUyU0Ex.5cO0C22ZF7N9GVNOVkmW.VnxyqgUNacWA8ihKVKqZKrOH7vSz4fkznyJ&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: 701308a9-e65e-406f-9434-07fd2bfc57c0)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/820c8683fe64b53f608de37a3a84c76c9c1a09f866fa9525cfa29d86a4fcef8c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T120154Z&X-Amz-Expires=86400&X-Amz-Signature=649100cb85f799d5074baf84eb640627ea87b189bb03f95c01a1c77a11b28396&X-Amz-SignedHeaders=host&partNumber=1&uploadId=8cimuA1DggLlFTGl0o8.2Y3pnYT5nTsnaa_gzi73BYNB9vkFEu4exfHrPlRI66q6zUyU0Ex.5cO0C22ZF7N9GVNOVkmW.VnxyqgUNacWA8ihKVKqZKrOH7vSz4fkznyJ&x-id=UploadPart
+ 2%|██▏ | 213/9378 [1:30:29<94:14:24, 37.02s/it]'(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/820c8683fe64b53f608de37a3a84c76c9c1a09f866fa9525cfa29d86a4fcef8c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T120154Z&X-Amz-Expires=86400&X-Amz-Signature=649100cb85f799d5074baf84eb640627ea87b189bb03f95c01a1c77a11b28396&X-Amz-SignedHeaders=host&partNumber=1&uploadId=8cimuA1DggLlFTGl0o8.2Y3pnYT5nTsnaa_gzi73BYNB9vkFEu4exfHrPlRI66q6zUyU0Ex.5cO0C22ZF7N9GVNOVkmW.VnxyqgUNacWA8ihKVKqZKrOH7vSz4fkznyJ&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: 701308a9-e65e-406f-9434-07fd2bfc57c0)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/820c8683fe64b53f608de37a3a84c76c9c1a09f866fa9525cfa29d86a4fcef8c?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T120154Z&X-Amz-Expires=86400&X-Amz-Signature=649100cb85f799d5074baf84eb640627ea87b189bb03f95c01a1c77a11b28396&X-Amz-SignedHeaders=host&partNumber=1&uploadId=8cimuA1DggLlFTGl0o8.2Y3pnYT5nTsnaa_gzi73BYNB9vkFEu4exfHrPlRI66q6zUyU0Ex.5cO0C22ZF7N9GVNOVkmW.VnxyqgUNacWA8ihKVKqZKrOH7vSz4fkznyJ&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+ 2%|██▏ | 214/9378 [1:30:54<85:51:01, 33.73s/it]
+
+ 2%|██▏ | 214/9378 [1:30:55<85:51:01, 33.73s/it]
+
+ 2%|██▏ | 215/9378 [1:31:21<80:02:55, 31.45s/it]
+
+ 2%|██▏ | 216/9378 [1:31:48<76:03:58, 29.89s/it]
+
+ 2%|██▏ | 217/9378 [1:32:14<73:00:49, 28.69s/it]
+
+
+ 2%|██▏ | 219/9378 [1:33:06<69:35:04, 27.35s/it]
+
+ 2%|██▎ | 220/9378 [1:33:32<68:41:16, 27.00s/it]
+
+ 2%|██▎ | 221/9378 [1:33:58<68:04:09, 26.76s/it]
+
+ 2%|██▎ | 222/9378 [1:34:24<67:36:19, 26.58s/it]
+
+ 2%|██▎ | 223/9378 [1:34:49<67:12:29, 26.43s/it]
+
+
+ 2%|██▎ | 224/9378 [1:35:16<67:25:20, 26.52s/it]
+
+
+ 2%|██▎ | 225/9378 [1:35:43<66:26:11, 26.13s/it]
+
+ 2%|██▎ | 226/9378 [1:36:08<66:20:30, 26.10s/it]
+
+ 2%|██▎ | 227/9378 [1:36:34<66:33:44, 26.19s/it]
+
+ 2%|██▎ | 228/9378 [1:37:01<66:31:46, 26.18s/it]
+
+ 2%|██▎ | 229/9378 [1:37:27<65:58:20, 25.96s/it]
+
+ 2%|██▎ | 230/9378 [1:37:52<65:59:41, 25.97s/it]
+
+ 2%|██▎ | 231/9378 [1:38:19<66:38:37, 26.23s/it]
+
+ 2%|██▎ | 232/9378 [1:38:45<65:54:03, 25.94s/it]
+
+ 2%|██▍ | 233/9378 [1:39:10<65:51:10, 25.92s/it]
+
+ 2%|██▍ | 234/9378 [1:39:36<66:13:42, 26.07s/it]
+
+ 3%|██▍ | 235/9378 [1:40:02<66:02:56, 26.01s/it]
+
+ 3%|██▍ | 236/9378 [1:40:28<65:49:02, 25.92s/it]
+{'loss': 1.4115, 'grad_norm': 1.015625, 'learning_rate': 4.111498257839721e-05, 'epoch': 0.08}
+
+ 3%|██▍ | 237/9378 [1:40:54<65:46:01, 25.90s/it]
+
+ 3%|██▍ | 238/9378 [1:41:20<65:59:06, 25.99s/it]
+
+
+ 3%|██▍ | 240/9378 [1:42:13<65:54:55, 25.97s/it]
+
+ 3%|██▍ | 241/9378 [1:42:39<65:57:26, 25.99s/it]
+{'loss': 1.2516, 'grad_norm': 1.171875, 'learning_rate': 4.198606271777004e-05, 'epoch': 0.08}
+
+
+ 3%|██▍ | 243/9378 [1:43:31<66:24:06, 26.17s/it]
+{'loss': 1.4107, 'grad_norm': 0.79296875, 'learning_rate': 4.233449477351917e-05, 'epoch': 0.08}
+
+ 3%|██▍ | 244/9378 [1:43:57<65:48:10, 25.94s/it]
+
+
+ 3%|██▌ | 246/9378 [1:44:49<66:24:35, 26.18s/it]
+{'loss': 1.2491, 'grad_norm': 0.7578125, 'learning_rate': 4.2857142857142856e-05, 'epoch': 0.08}
+
+ 3%|██▌ | 247/9378 [1:45:15<65:49:00, 25.95s/it]
+
+ 3%|██▌ | 248/9378 [1:45:41<65:55:09, 25.99s/it]
+
+
+ 3%|██▌ | 250/9378 [1:46:53<82:19:51, 32.47s/it]
+
+
+ 3%|██▌ | 251/9378 [1:47:19<77:23:14, 30.52s/it]
+
+
+ 3%|██▌ | 252/9378 [1:47:45<73:51:05, 29.13s/it]
+
+
+ 3%|██▌ | 253/9378 [1:48:11<71:28:29, 28.20s/it]
+
+
+ 3%|██▌ | 254/9378 [1:48:37<69:59:58, 27.62s/it]
+
+ 3%|██▌ | 254/9378 [1:48:39<69:59:58, 27.62s/it]
+
+
+ 3%|██▌ | 256/9378 [1:49:30<67:59:11, 26.83s/it]
+
+
+ 3%|██▋ | 257/9378 [1:49:56<67:21:24, 26.59s/it]
+
+ 3%|██▋ | 257/9378 [1:49:57<67:21:24, 26.59s/it]
+
+
+ 3%|██▋ | 259/9378 [1:50:48<66:23:49, 26.21s/it]
+
+ 3%|██▋ | 259/9378 [1:50:49<66:23:49, 26.21s/it]
+
+
+ 3%|██▋ | 261/9378 [1:51:40<66:07:52, 26.11s/it]
+
+
+ 3%|██▋ | 262/9378 [1:52:06<65:54:58, 26.03s/it]
+
+
+ 3%|██▋ | 263/9378 [1:52:33<65:53:35, 26.02s/it]
+
+ 3%|██▋ | 264/9378 [1:52:58<65:50:26, 26.01s/it]
+
+
+ 3%|██▋ | 265/9378 [1:53:24<66:18:16, 26.19s/it]
+
+
+ 3%|██▋ | 266/9378 [1:53:50<65:42:01, 25.96s/it]
+
+
+ 3%|██▋ | 267/9378 [1:54:16<65:54:46, 26.04s/it]
+
+
+ 3%|██▋ | 268/9378 [1:54:42<66:21:48, 26.22s/it]
+
+
+ 3%|██▊ | 269/9378 [1:55:08<66:01:35, 26.09s/it]
+
+
+ 3%|██▊ | 270/9378 [1:55:34<65:35:08, 25.92s/it]
+
+
+ 3%|██▊ | 271/9378 [1:56:00<65:37:43, 25.94s/it]
+
+
+ 3%|██▊ | 272/9378 [1:56:27<65:32:56, 25.91s/it]
+
+ 3%|██▊ | 273/9378 [1:56:53<65:40:27, 25.97s/it]
+
+ 3%|██▊ | 274/9378 [1:57:19<66:18:52, 26.22s/it]
+
+ 3%|██▊ | 275/9378 [1:57:45<65:35:45, 25.94s/it]
+
+ 3%|██▊ | 276/9378 [1:58:11<65:26:55, 25.89s/it]
+
+ 3%|██▊ | 277/9378 [1:58:37<65:31:19, 25.92s/it]
+
+ 3%|██▊ | 278/9378 [1:59:03<65:30:22, 25.91s/it]
+
+ 3%|██▊ | 279/9378 [1:59:29<65:29:10, 25.91s/it]
+
+ 3%|██▊ | 280/9378 [1:59:54<65:27:31, 25.90s/it]
+
+ 3%|██▉ | 281/9378 [2:00:20<65:36:18, 25.96s/it]
+
+ 3%|██▉ | 282/9378 [2:00:47<65:41:17, 26.00s/it]
+
+ 3%|██▉ | 283/9378 [2:01:12<65:35:27, 25.96s/it]
+
+ 3%|██▉ | 284/9378 [2:01:38<66:02:00, 26.14s/it]
+
+ 3%|██▉ | 285/9378 [2:02:04<65:38:01, 25.99s/it]
+
+ 3%|██▉ | 286/9378 [2:02:29<65:13:56, 25.83s/it]
+
+
+ 3%|██▉ | 287/9378 [2:02:55<65:21:18, 25.88s/it]
+
+
+ 3%|██▉ | 288/9378 [2:03:21<65:26:52, 25.92s/it]
+
+ 3%|██▉ | 288/9378 [2:03:22<65:26:52, 25.92s/it]
+
+
+ 3%|██▉ | 290/9378 [2:04:13<65:22:02, 25.89s/it]
+
+
+
+ 3%|██▉ | 291/9378 [2:04:40<65:41:10, 26.02s/it]
+
+
+ 3%|██▉ | 292/9378 [2:05:06<65:43:28, 26.04s/it]
+
+ 3%|██▉ | 293/9378 [2:05:31<65:31:09, 25.96s/it]
+
+
+ 3%|███ | 294/9378 [2:05:57<65:23:47, 25.92s/it]
+
+
+ 3%|███ | 295/9378 [2:06:23<65:35:00, 25.99s/it]
+
+
+ 3%|███ | 296/9378 [2:06:49<65:34:32, 25.99s/it]
+
+
+ 3%|███ | 297/9378 [2:07:15<65:40:46, 26.04s/it]
+
+
+ 3%|███ | 298/9378 [2:07:42<65:55:45, 26.14s/it]
+
+ 3%|███ | 299/9378 [2:08:07<65:29:33, 25.97s/it]
+
+
+ 3%|███ | 300/9378 [2:08:55<81:48:49, 32.44s/it]
+
+
+ 3%|███ | 301/9378 [2:09:22<76:51:18, 30.48s/it]
+
+ 3%|███ | 302/9378 [2:09:48<73:29:08, 29.15s/it]
+
+ 3%|███ | 303/9378 [2:10:14<71:26:06, 28.34s/it]
+
+ 3%|███ | 304/9378 [2:10:40<69:10:41, 27.45s/it]
+
+ 3%|███ | 305/9378 [2:11:06<68:09:36, 27.04s/it]
+
+ 3%|███▏ | 306/9378 [2:11:32<67:39:59, 26.85s/it]
+
+ 3%|███▏ | 307/9378 [2:11:58<66:42:01, 26.47s/it]
+
+ 3%|███▏ | 308/9378 [2:12:24<66:31:37, 26.41s/it]
+
+ 3%|███▏ | 309/9378 [2:12:50<66:13:01, 26.29s/it]
+
+ 3%|███▏ | 310/9378 [2:13:16<65:54:24, 26.17s/it]
+
+ 3%|███▏ | 311/9378 [2:13:42<65:55:25, 26.17s/it]
+
+ 3%|███▏ | 312/9378 [2:14:08<66:14:50, 26.31s/it]
+
+ 3%|███▏ | 313/9378 [2:14:34<65:17:04, 25.93s/it]
+
+ 3%|███▏ | 314/9378 [2:15:00<65:50:14, 26.15s/it]
+
+ 3%|███▏ | 315/9378 [2:15:26<65:21:08, 25.96s/it]
+
+ 3%|███▏ | 316/9378 [2:15:52<65:25:43, 25.99s/it]
+
+ 3%|███▏ | 317/9378 [2:16:18<65:32:21, 26.04s/it]
+
+ 3%|███▎ | 318/9378 [2:16:44<65:26:24, 26.00s/it]
+
+ 3%|███▎ | 319/9378 [2:17:10<65:22:58, 25.98s/it]
+
+ 3%|███▎ | 320/9378 [2:17:36<65:18:21, 25.96s/it]
+
+ 3%|███▎ | 321/9378 [2:18:02<65:22:32, 25.99s/it]
+{'loss': 1.3795, 'grad_norm': 0.73046875, 'learning_rate': 4.99379035306331e-05, 'epoch': 0.1}
+
+ 3%|███▎ | 322/9378 [2:18:28<65:16:28, 25.95s/it]
+
+
+ 3%|███▎ | 324/9378 [2:19:19<65:09:07, 25.91s/it]
+
+ 3%|███▎ | 324/9378 [2:19:20<65:09:07, 25.91s/it]
+
+
+ 3%|███▎ | 326/9378 [2:20:10<64:57:59, 25.84s/it]
+
+
+ 3%|███▎ | 327/9378 [2:20:36<65:01:18, 25.86s/it]
+
+
+ 3%|███▎ | 328/9378 [2:21:02<65:03:38, 25.88s/it]
+
+
+ 4%|███▎ | 329/9378 [2:21:28<65:10:54, 25.93s/it]
+
+
+ 4%|███▍ | 330/9378 [2:21:55<65:22:52, 26.01s/it]
+
+ 4%|███▍ | 331/9378 [2:22:20<65:01:37, 25.88s/it]
+
+
+ 4%|███▍ | 332/9378 [2:22:46<65:08:36, 25.92s/it]
+
+
+ 4%|███▍ | 333/9378 [2:23:12<65:12:31, 25.95s/it]
+
+
+ 4%|███▍ | 334/9378 [2:23:38<65:11:08, 25.95s/it]
+
+
+ 4%|███▍ | 335/9378 [2:24:04<65:04:59, 25.91s/it]
+
+
+ 4%|███▍ | 336/9378 [2:24:31<65:11:51, 25.96s/it]
+
+ 4%|███▍ | 337/9378 [2:24:57<65:09:09, 25.94s/it]
+
+ 4%|███▍ | 338/9378 [2:25:23<65:07:38, 25.94s/it]
+
+ 4%|███▍ | 339/9378 [2:25:49<65:17:07, 26.00s/it]
+
+ 4%|███▍ | 340/9378 [2:26:15<65:13:39, 25.98s/it]
+
+ 4%|███▍ | 341/9378 [2:26:41<65:09:41, 25.96s/it]
+
+ 4%|███▌ | 342/9378 [2:27:07<65:56:22, 26.27s/it]
+
+ 4%|███▌ | 343/9378 [2:27:32<65:22:44, 26.05s/it]
+
+
+ 4%|███▌ | 344/9378 [2:27:59<65:16:25, 26.01s/it]
+
+ 4%|███▌ | 345/9378 [2:28:24<65:21:45, 26.05s/it]
+
+
+ 4%|███▌ | 346/9378 [2:28:51<65:48:52, 26.23s/it]
+
+
+ 4%|███▌ | 347/9378 [2:29:16<65:12:34, 25.99s/it]
+
+
+ 4%|███▌ | 348/9378 [2:29:44<65:44:55, 26.21s/it]
+
+ 4%|███▌ | 349/9378 [2:30:08<65:03:18, 25.94s/it]
+
+
+ 4%|███▌ | 350/9378 [2:30:57<81:51:57, 32.64s/it]
+
+ 4%|███▌ | 351/9378 [2:31:23<76:18:38, 30.43s/it]
+
+ 4%|███▌ | 352/9378 [2:31:49<73:19:22, 29.24s/it]
+
+ 4%|███▌ | 353/9378 [2:32:15<70:34:15, 28.15s/it]
+
+ 4%|███▌ | 354/9378 [2:32:41<68:53:17, 27.48s/it]
+
+ 4%|███▋ | 355/9378 [2:33:07<67:45:59, 27.04s/it]
+
+ 4%|███▋ | 356/9378 [2:33:33<67:35:39, 26.97s/it]
+
+ 4%|███▋ | 357/9378 [2:33:59<66:52:49, 26.69s/it]
+
+ 4%|███▋ | 358/9378 [2:34:25<65:48:50, 26.27s/it]
+
+ 4%|███▋ | 359/9378 [2:34:51<65:43:15, 26.23s/it]
+
+ 4%|███▋ | 360/9378 [2:35:17<65:28:54, 26.14s/it]
+
+ 4%|███▋ | 361/9378 [2:35:44<65:34:58, 26.18s/it]
+
+ 4%|███▋ | 362/9378 [2:36:10<65:36:42, 26.20s/it]
+
+ 4%|███▋ | 363/9378 [2:36:36<66:03:13, 26.38s/it]
+
+ 4%|███▋ | 364/9378 [2:37:02<65:10:54, 26.03s/it]
+
+ 4%|███▋ | 365/9378 [2:37:28<65:42:10, 26.24s/it]
+
+ 4%|███▋ | 366/9378 [2:37:54<64:52:28, 25.92s/it]
+
+ 4%|███▊ | 367/9378 [2:38:20<65:35:03, 26.20s/it]
+
+ 4%|███▊ | 368/9378 [2:38:46<65:01:58, 25.98s/it]
+
+
+ 4%|███▊ | 369/9378 [2:39:11<64:34:06, 25.80s/it]
+
+ 4%|███▊ | 370/9378 [2:39:38<65:01:14, 25.99s/it]
+
+ 4%|███▊ | 371/9378 [2:40:04<65:17:06, 26.09s/it]
+
+ 4%|███▊ | 372/9378 [2:40:30<65:16:24, 26.09s/it]
+
+ 4%|███▊ | 373/9378 [2:40:56<65:17:36, 26.10s/it]
+
+
+ 4%|███▊ | 374/9378 [2:41:22<64:35:29, 25.83s/it]
+
+ 4%|███▊ | 375/9378 [2:41:48<64:55:07, 25.96s/it]
+
+ 4%|███▊ | 376/9378 [2:42:14<64:59:59, 25.99s/it]
+
+ 4%|███▊ | 377/9378 [2:42:40<65:31:33, 26.21s/it]
+
+ 4%|███▊ | 378/9378 [2:43:06<64:49:07, 25.93s/it]
+
+
+ 4%|███▉ | 379/9378 [2:43:32<64:47:18, 25.92s/it]
+
+ 4%|███▉ | 380/9378 [2:43:58<64:59:51, 26.00s/it]
+
+ 4%|███▉ | 381/9378 [2:44:24<64:58:49, 26.00s/it]
+
+
+ 4%|███▉ | 382/9378 [2:44:50<64:51:42, 25.96s/it]
+{'loss': 1.3232, 'grad_norm': 0.65625, 'learning_rate': 4.9516572336008385e-05, 'epoch': 0.12}
+
+
+
+ 4%|███▉ | 384/9378 [2:45:42<64:41:07, 25.89s/it]
+
+ 4%|███▉ | 385/9378 [2:46:07<64:49:23, 25.95s/it]
+
+
+ 4%|███▉ | 386/9378 [2:46:33<64:50:33, 25.96s/it]
+
+
+ 4%|███▉ | 387/9378 [2:46:59<65:02:22, 26.04s/it]
+
+
+ 4%|███▉ | 388/9378 [2:47:25<64:46:44, 25.94s/it]
+
+
+ 4%|███▉ | 389/9378 [2:47:51<64:38:05, 25.89s/it]
+
+
+ 4%|███▉ | 390/9378 [2:48:16<64:38:39, 25.89s/it]
+
+
+ 4%|████ | 391/9378 [2:48:43<64:47:17, 25.95s/it]
+
+ 4%|████ | 392/9378 [2:49:08<64:46:48, 25.95s/it]
+
+
+ 4%|████ | 393/9378 [2:49:34<64:46:43, 25.95s/it]
+
+
+ 4%|████ | 394/9378 [2:50:00<64:49:29, 25.98s/it]
+
+
+ 4%|████ | 395/9378 [2:50:27<64:41:04, 25.92s/it]
+
+ 4%|████ | 396/9378 [2:50:53<64:42:52, 25.94s/it]
+
+ 4%|████ | 397/9378 [2:51:19<64:45:14, 25.96s/it]
+
+ 4%|████ | 398/9378 [2:51:45<64:52:57, 26.01s/it]
+
+ 4%|████ | 399/9378 [2:52:10<64:53:40, 26.02s/it]
+
+
+ 4%|████ | 400/9378 [2:52:57<79:24:03, 31.84s/it]
+
+ 4%|████ | 401/9378 [2:53:23<75:02:08, 30.09s/it]
+
+ 4%|████ | 402/9378 [2:53:49<71:55:25, 28.85s/it]
+
+ 4%|████▏ | 403/9378 [2:54:15<69:49:30, 28.01s/it]
+
+
+ 4%|████▏ | 404/9378 [2:54:41<68:16:55, 27.39s/it]
+
+
+ 4%|████▏ | 405/9378 [2:55:07<67:14:02, 26.97s/it]
+
+ 4%|████▏ | 406/9378 [2:55:33<66:58:43, 26.88s/it]
+
+ 4%|████▏ | 407/9378 [2:55:59<66:36:48, 26.73s/it]
+
+ 4%|████▏ | 408/9378 [2:56:25<66:01:01, 26.50s/it]
+
+ 4%|████▏ | 409/9378 [2:56:51<65:05:21, 26.13s/it]
+
+ 4%|████▏ | 410/9378 [2:57:17<65:30:48, 26.30s/it]
+
+ 4%|████▏ | 411/9378 [2:57:43<65:20:24, 26.23s/it]
+
+ 4%|████▏ | 412/9378 [2:58:09<64:40:42, 25.97s/it]
+
+
+ 4%|████▏ | 413/9378 [2:58:35<64:32:41, 25.92s/it]
+
+
+ 4%|████▏ | 414/9378 [2:59:01<64:51:48, 26.05s/it]
+
+ 4%|████▏ | 415/9378 [2:59:27<65:12:48, 26.19s/it]
+
+ 4%|████▎ | 416/9378 [2:59:52<64:25:06, 25.88s/it]
+
+
+ 4%|████▎ | 417/9378 [3:00:18<64:26:13, 25.89s/it]
+
+
+ 4%|████▎ | 418/9378 [3:00:44<64:27:49, 25.90s/it]
+
+ 4%|████▎ | 418/9378 [3:00:45<64:27:49, 25.90s/it]
+[2024-05-12 13:33:40,660] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-418/pytorch_model_fsdp.bin
+[2024-05-12 13:33:52,987] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-418/pytorch_model_fsdp.bin
+[2024-05-12 13:34:17,390] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-418/optimizer.bin
+ 4%|████▎ | 418/9378 [3:00:45<64:27:49, 25.90s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 4%|████▏ | 419/9378 [3:03:07<152:19:13, 61.21s/it]
+
+
+ 4%|████▎ | 420/9378 [3:03:34<126:41:19, 50.91s/it]
+
+
+ 4%|████▎ | 421/9378 [3:04:00<108:10:18, 43.48s/it]
+
+
+ 4%|████▎ | 422/9378 [3:04:27<94:31:19, 37.99s/it]
+
+ 5%|████▎ | 423/9378 [3:04:53<85:44:16, 34.47s/it]
+
+ 5%|████▎ | 424/9378 [3:05:18<79:51:50, 32.11s/it]
+
+
+ 5%|████▎ | 425/9378 [3:05:45<75:33:32, 30.38s/it]
+
+
+ 5%|████▎ | 426/9378 [3:06:12<73:03:40, 29.38s/it]
+
+
+ 5%|████▎ | 427/9378 [3:06:39<70:07:31, 28.20s/it]
+
+ 5%|████▍ | 428/9378 [3:07:05<68:42:57, 27.64s/it]
+
+
+ 5%|████▍ | 429/9378 [3:07:31<67:17:00, 27.07s/it]
+
+ 5%|████▍ | 430/9378 [3:07:57<66:35:52, 26.79s/it]
+
+ 5%|████▍ | 431/9378 [3:08:23<66:11:39, 26.63s/it]
+
+ 5%|████▍ | 432/9378 [3:08:49<65:58:43, 26.55s/it]
+
+ 5%|████▍ | 433/9378 [3:09:15<66:21:10, 26.70s/it]
+
+
+ 5%|████▍ | 434/9378 [3:09:41<65:21:47, 26.31s/it]
+
+
+ 5%|████▍ | 435/9378 [3:10:07<65:09:29, 26.23s/it]
+
+ 5%|████▍ | 435/9378 [3:10:08<65:09:29, 26.23s/it]
+
+
+ 5%|████▍ | 437/9378 [3:10:59<64:53:50, 26.13s/it]
+
+
+
+ 5%|████▍ | 438/9378 [3:11:27<65:06:12, 26.22s/it]
+
+ 5%|████▍ | 439/9378 [3:11:53<65:12:22, 26.26s/it]
+
+ 5%|████▌ | 440/9378 [3:12:19<65:37:35, 26.43s/it]
+
+
+ 5%|████▌ | 441/9378 [3:12:45<65:36:41, 26.43s/it]
+
+
+ 5%|████▌ | 442/9378 [3:13:11<65:14:48, 26.29s/it]
+
+
+ 5%|████▌ | 443/9378 [3:13:39<66:14:38, 26.69s/it]
+
+
+ 5%|████▌ | 444/9378 [3:14:05<65:13:04, 26.28s/it]
+
+ 5%|████▌ | 445/9378 [3:14:31<65:11:34, 26.27s/it]
+
+ 5%|████▌ | 446/9378 [3:14:58<65:39:37, 26.46s/it]
+
+ 5%|████▌ | 447/9378 [3:15:24<64:36:55, 26.05s/it]
+
+ 5%|████▌ | 448/9378 [3:15:50<65:30:57, 26.41s/it]
+
+ 5%|████▌ | 449/9378 [3:16:16<64:44:26, 26.10s/it]
+
+ 5%|████▌ | 450/9378 [3:17:02<79:05:23, 31.89s/it]
+
+ 5%|████▌ | 451/9378 [3:17:28<74:58:13, 30.23s/it]
+
+ 5%|████▋ | 452/9378 [3:17:54<72:05:16, 29.07s/it]
+
+ 5%|████▋ | 453/9378 [3:18:19<70:00:34, 28.24s/it]
+
+
+ 5%|████▋ | 454/9378 [3:18:45<68:20:01, 27.57s/it]
+
+
+ 5%|████▋ | 455/9378 [3:19:12<67:18:56, 27.16s/it]
+
+
+ 5%|████▋ | 456/9378 [3:19:37<66:25:08, 26.80s/it]
+
+
+ 5%|████▋ | 457/9378 [3:20:04<66:33:17, 26.86s/it]
+
+
+ 5%|████▋ | 458/9378 [3:20:30<65:30:23, 26.44s/it]
+
+
+ 5%|████▋ | 459/9378 [3:20:56<65:09:14, 26.30s/it]
+
+
+ 5%|████▋ | 460/9378 [3:21:22<64:59:39, 26.24s/it]
+
+
+ 5%|████▋ | 461/9378 [3:21:49<65:14:54, 26.34s/it]
+
+ 5%|████▋ | 462/9378 [3:22:15<65:21:09, 26.39s/it]
+
+
+ 5%|████▋ | 463/9378 [3:22:42<64:32:38, 26.06s/it]
+
+
+ 5%|████▋ | 464/9378 [3:23:08<64:38:04, 26.10s/it]
+
+
+ 5%|████▊ | 465/9378 [3:23:34<64:29:57, 26.05s/it]
+
+
+ 5%|████▊ | 466/9378 [3:24:00<64:36:27, 26.10s/it]
+
+ 5%|████▊ | 467/9378 [3:24:26<64:49:13, 26.19s/it]
+
+ 5%|████▊ | 468/9378 [3:24:52<65:19:25, 26.39s/it]
+
+ 5%|████▊ | 469/9378 [3:25:18<65:04:33, 26.30s/it]
+
+ 5%|████▊ | 470/9378 [3:25:45<64:52:12, 26.22s/it]
+
+ 5%|████▊ | 471/9378 [3:26:11<64:30:02, 26.07s/it]
+
+ 5%|████▊ | 472/9378 [3:26:37<64:03:25, 25.89s/it]
+
+ 5%|████▊ | 473/9378 [3:27:03<64:17:59, 25.99s/it]
+
+ 5%|████▊ | 474/9378 [3:27:29<64:56:02, 26.25s/it]
+
+ 5%|████▊ | 475/9378 [3:27:55<64:12:46, 25.96s/it]
+
+ 5%|████▊ | 476/9378 [3:28:21<64:15:27, 25.99s/it]
+
+ 5%|████▉ | 477/9378 [3:28:47<64:17:40, 26.00s/it]
+
+ 5%|████▉ | 478/9378 [3:29:13<64:22:46, 26.04s/it]
+
+ 5%|████▉ | 479/9378 [3:29:39<65:01:01, 26.30s/it]
+
+ 5%|████▉ | 480/9378 [3:30:05<64:20:17, 26.03s/it]
+
+ 5%|████▉ | 481/9378 [3:30:30<64:21:40, 26.04s/it]
+
+
+ 5%|████▉ | 482/9378 [3:30:57<64:37:33, 26.15s/it]
+
+
+ 5%|████▉ | 483/9378 [3:31:24<65:13:22, 26.40s/it]
+
+
+ 5%|████▉ | 484/9378 [3:31:50<64:31:42, 26.12s/it]
+
+
+ 5%|████▉ | 485/9378 [3:32:16<64:29:33, 26.11s/it]
+
+
+ 5%|████▉ | 486/9378 [3:32:43<64:30:03, 26.11s/it]
+
+
+ 5%|████▉ | 487/9378 [3:33:09<64:28:50, 26.11s/it]
+
+ 5%|████▉ | 488/9378 [3:33:35<64:42:05, 26.20s/it]
+
+ 5%|█████ | 489/9378 [3:34:01<64:40:58, 26.20s/it]
+
+ 5%|█████ | 490/9378 [3:34:27<64:32:05, 26.14s/it]
+
+ 5%|█████ | 491/9378 [3:34:53<64:32:38, 26.15s/it]
+
+ 5%|█████ | 492/9378 [3:35:19<64:49:42, 26.26s/it]
+
+
+ 5%|█████ | 493/9378 [3:35:45<64:38:37, 26.19s/it]
+
+
+ 5%|█████ | 494/9378 [3:36:12<65:11:03, 26.41s/it]
+
+ 5%|█████ | 495/9378 [3:36:38<64:58:30, 26.33s/it]
+
+ 5%|█████ | 496/9378 [3:37:04<64:46:51, 26.26s/it]
+
+ 5%|█████ | 497/9378 [3:37:30<64:23:27, 26.10s/it]
+
+ 5%|█████ | 498/9378 [3:37:55<63:58:38, 25.94s/it]
+
+
+ 5%|█████ | 499/9378 [3:38:21<64:03:50, 25.97s/it]
+
+
+ 5%|█████ | 500/9378 [3:39:10<80:01:03, 32.45s/it]
+
+ 5%|█████▏ | 501/9378 [3:39:36<75:16:16, 30.53s/it]
+
+ 5%|█████▏ | 502/9378 [3:40:02<71:52:24, 29.15s/it]
+
+ 5%|█████▏ | 503/9378 [3:40:28<69:36:07, 28.23s/it]
+
+ 5%|█████▏ | 504/9378 [3:40:54<67:59:04, 27.58s/it]
+
+ 5%|█████▏ | 505/9378 [3:41:20<66:49:33, 27.11s/it]
+
+ 5%|█████▏ | 506/9378 [3:41:46<66:05:46, 26.82s/it]
+
+ 5%|█████▏ | 507/9378 [3:42:11<65:33:59, 26.61s/it]
+
+
+ 5%|█████▏ | 508/9378 [3:42:39<65:15:50, 26.49s/it]
+
+ 5%|█████▏ | 509/9378 [3:43:04<65:10:15, 26.45s/it]
+
+
+ 5%|█████▏ | 510/9378 [3:43:30<65:06:05, 26.43s/it]
+
+
+ 5%|█████▏ | 511/9378 [3:43:58<65:34:44, 26.63s/it]
+
+ 5%|█████▏ | 512/9378 [3:44:24<65:16:16, 26.50s/it]
+
+
+ 5%|█████▎ | 513/9378 [3:44:50<64:26:00, 26.17s/it]
+
+ 5%|█████▎ | 514/9378 [3:45:16<64:57:35, 26.38s/it]
+
+
+ 5%|█████▎ | 515/9378 [3:45:42<64:05:48, 26.04s/it]
+
+
+ 6%|█████▎ | 516/9378 [3:46:08<64:15:08, 26.10s/it]
+
+
+ 6%|█████▎ | 517/9378 [3:46:34<64:10:35, 26.07s/it]
+
+ 6%|█████▎ | 518/9378 [3:47:01<64:29:37, 26.21s/it]
+
+ 6%|█████▎ | 519/9378 [3:47:27<64:08:12, 26.06s/it]
+
+ 6%|█████▎ | 520/9378 [3:47:53<64:47:30, 26.33s/it]
+
+ 6%|█████▎ | 521/9378 [3:48:19<64:02:16, 26.03s/it]
+
+ 6%|█████▎ | 522/9378 [3:48:45<64:12:07, 26.10s/it]
+
+ 6%|█████▎ | 523/9378 [3:49:11<64:04:21, 26.05s/it]
+
+ 6%|█████▎ | 524/9378 [3:49:38<64:19:31, 26.15s/it]
+
+ 6%|█████▎ | 525/9378 [3:50:03<64:22:43, 26.18s/it]
+
+
+ 6%|█████▍ | 526/9378 [3:50:29<64:17:48, 26.15s/it]
+
+
+ 6%|█████▍ | 527/9378 [3:50:56<64:53:07, 26.39s/it]
+
+
+ 6%|█████▍ | 528/9378 [3:51:23<64:28:51, 26.23s/it]
+
+
+ 6%|█████▍ | 529/9378 [3:51:49<64:18:21, 26.16s/it]
+
+
+ 6%|█████▍ | 530/9378 [3:52:15<64:13:02, 26.13s/it]
+
+
+ 6%|█████▍ | 531/9378 [3:52:41<64:08:31, 26.10s/it]
+
+
+ 6%|█████▍ | 532/9378 [3:53:07<64:02:52, 26.07s/it]
+
+
+ 6%|█████▍ | 533/9378 [3:53:33<64:06:31, 26.09s/it]
+
+ 6%|█████▍ | 534/9378 [3:53:59<64:12:35, 26.14s/it]
+
+ 6%|█████▍ | 535/9378 [3:54:26<64:25:23, 26.23s/it]
+
+ 6%|█████▍ | 536/9378 [3:54:52<64:18:46, 26.18s/it]
+
+ 6%|█████▍ | 537/9378 [3:55:18<64:21:26, 26.21s/it]
+
+ 6%|█████▌ | 538/9378 [3:55:43<64:22:01, 26.21s/it]
+
+
+ 6%|█████▌ | 539/9378 [3:56:09<64:21:18, 26.21s/it]
+
+
+ 6%|█████▌ | 540/9378 [3:56:37<65:01:38, 26.49s/it]
+
+ 6%|█████▌ | 541/9378 [3:57:03<64:46:21, 26.39s/it]
+
+
+ 6%|█████▌ | 542/9378 [3:57:29<63:59:12, 26.07s/it]
+
+
+ 6%|█████▌ | 543/9378 [3:57:55<63:55:26, 26.05s/it]
+
+
+ 6%|█████▌ | 544/9378 [3:58:21<63:50:44, 26.02s/it]
+
+
+ 6%|█████▌ | 545/9378 [3:58:47<64:01:07, 26.09s/it]
+
+ 6%|█████▌ | 546/9378 [3:59:13<64:34:32, 26.32s/it]
+
+
+ 6%|█████▌ | 547/9378 [3:59:39<63:41:05, 25.96s/it]
+
+
+ 6%|█████▌ | 548/9378 [4:00:05<63:53:45, 26.05s/it]
+
+
+ 6%|█████▌ | 549/9378 [4:00:32<63:55:27, 26.06s/it]
+
+
+ 6%|█████▋ | 550/9378 [4:01:19<79:59:06, 32.62s/it]
+
+ 6%|█████▋ | 551/9378 [4:01:46<75:28:01, 30.78s/it]
+
+ 6%|█████▋ | 552/9378 [4:02:12<72:12:33, 29.45s/it]
+
+ 6%|█████▋ | 553/9378 [4:02:38<69:38:50, 28.41s/it]
+
+ 6%|█████▋ | 554/9378 [4:03:04<67:57:19, 27.72s/it]
+
+ 6%|█████▋ | 555/9378 [4:03:31<66:48:45, 27.26s/it]
+
+ 6%|█████▋ | 556/9378 [4:03:57<66:17:16, 27.05s/it]
+
+ 6%|█████▋ | 557/9378 [4:04:22<65:12:51, 26.62s/it]
+
+
+ 6%|█████▋ | 558/9378 [4:04:48<64:52:46, 26.48s/it]
+
+
+ 6%|█████▋ | 559/9378 [4:05:14<64:26:54, 26.31s/it]
+
+
+ 6%|█████▋ | 560/9378 [4:05:40<64:39:16, 26.40s/it]
+
+
+ 6%|█████▋ | 561/9378 [4:06:06<64:07:46, 26.18s/it]
+
+
+ 6%|█████▊ | 562/9378 [4:06:32<64:05:18, 26.17s/it]
+
+
+ 6%|█████▊ | 563/9378 [4:06:59<64:40:33, 26.41s/it]
+
+
+ 6%|█████▊ | 564/9378 [4:07:26<63:50:39, 26.08s/it]
+
+
+ 6%|█████▊ | 565/9378 [4:07:52<63:50:13, 26.08s/it]
+
+
+ 6%|█████▊ | 566/9378 [4:08:18<63:56:28, 26.12s/it]
+
+
+ 6%|█████▊ | 567/9378 [4:08:44<63:59:50, 26.15s/it]
+
+
+ 6%|█████▊ | 568/9378 [4:09:10<64:04:33, 26.18s/it]
+
+ 6%|█████▊ | 569/9378 [4:09:37<64:11:16, 26.23s/it]
+
+
+ 6%|█████▊ | 570/9378 [4:10:03<63:57:51, 26.14s/it]
+
+ 6%|█████▊ | 571/9378 [4:10:29<64:09:13, 26.22s/it]
+
+ 6%|█████▊ | 572/9378 [4:10:54<64:19:05, 26.29s/it]
+
+
+ 6%|█████▊ | 573/9378 [4:11:22<64:16:31, 26.28s/it]
+
+ 6%|█████▉ | 574/9378 [4:11:48<64:09:28, 26.23s/it]
+
+ 6%|█████▉ | 575/9378 [4:12:14<64:04:58, 26.21s/it]
+
+
+ 6%|█████▉ | 576/9378 [4:12:40<64:04:16, 26.21s/it]
+
+
+ 6%|█████▉ | 577/9378 [4:13:06<64:04:10, 26.21s/it]
+
+
+ 6%|█████▉ | 578/9378 [4:13:33<64:01:14, 26.19s/it]
+
+
+ 6%|█████▉ | 579/9378 [4:13:59<64:07:52, 26.24s/it]
+
+
+ 6%|█████▉ | 580/9378 [4:14:25<63:56:09, 26.16s/it]
+
+
+ 6%|█████▉ | 581/9378 [4:14:51<63:57:57, 26.18s/it]
+
+
+ 6%|█████▉ | 582/9378 [4:15:17<63:51:22, 26.13s/it]
+
+
+ 6%|█████▉ | 583/9378 [4:15:43<63:52:13, 26.14s/it]
+
+ 6%|█████▉ | 584/9378 [4:16:09<64:06:51, 26.25s/it]
+
+ 6%|█████▉ | 585/9378 [4:16:35<64:15:20, 26.31s/it]
+
+ 6%|█████▉ | 586/9378 [4:17:00<63:38:45, 26.06s/it]
+
+ 6%|█████▉ | 586/9378 [4:17:02<63:38:45, 26.06s/it]
+
+
+ 6%|██████ | 588/9378 [4:17:54<64:01:12, 26.22s/it]
+
+ 6%|██████ | 589/9378 [4:18:20<63:46:16, 26.12s/it]
+{'loss': 1.4272, 'grad_norm': 0.578125, 'learning_rate': 4.52568623182876e-05, 'epoch': 0.19}
+
+
+ 6%|██████ | 591/9378 [4:19:12<63:39:36, 26.08s/it]
+
+ 6%|██████ | 592/9378 [4:19:37<63:47:02, 26.14s/it]
+
+
+ 6%|██████ | 593/9378 [4:20:05<64:33:15, 26.45s/it]
+
+ 6%|██████ | 594/9378 [4:20:30<63:43:07, 26.11s/it]
+
+
+ 6%|██████ | 595/9378 [4:20:57<64:16:42, 26.35s/it]
+
+ 6%|██████ | 596/9378 [4:21:22<63:32:33, 26.05s/it]
+
+
+ 6%|██████ | 597/9378 [4:21:48<63:42:14, 26.12s/it]
+
+
+ 6%|██████ | 598/9378 [4:22:14<63:38:38, 26.10s/it]
+
+
+ 6%|██████▏ | 599/9378 [4:22:41<63:49:14, 26.17s/it]
+
+
+ 6%|██████▏ | 600/9378 [4:23:28<79:19:53, 32.54s/it]
+
+
+ 6%|██████▏ | 601/9378 [4:23:54<74:35:10, 30.59s/it]
+
+
+ 6%|██████▏ | 602/9378 [4:24:20<71:23:04, 29.28s/it]
+
+ 6%|██████▏ | 602/9378 [4:24:22<71:23:04, 29.28s/it]
+
+ 6%|██████▏ | 603/9378 [4:24:48<69:32:55, 28.53s/it]
+
+ 6%|██████▏ | 604/9378 [4:25:14<67:48:28, 27.82s/it]
+
+
+ 6%|██████▏ | 606/9378 [4:26:05<65:11:28, 26.75s/it]
+
+
+ 6%|██████▏ | 607/9378 [4:26:31<64:43:58, 26.57s/it]
+
+
+ 6%|██████▏ | 608/9378 [4:26:57<64:24:18, 26.44s/it]
+
+ 6%|██████▏ | 608/9378 [4:26:59<64:24:18, 26.44s/it]
+
+ 6%|██████▏ | 609/9378 [4:27:25<64:11:30, 26.35s/it]
+
+
+ 7%|██████▎ | 611/9378 [4:28:17<64:31:00, 26.49s/it]
+{'loss': 1.3416, 'grad_norm': 0.6328125, 'learning_rate': 4.45676985045181e-05, 'epoch': 0.2}
+
+
+ 7%|██████▎ | 613/9378 [4:29:08<63:37:18, 26.13s/it]
+
+
+ 7%|██████▎ | 614/9378 [4:29:35<63:44:34, 26.18s/it]
+
+
+ 7%|██████▎ | 615/9378 [4:30:01<63:39:00, 26.15s/it]
+
+
+ 7%|██████▎ | 616/9378 [4:30:27<63:30:35, 26.09s/it]
+
+
+ 7%|██████▎ | 617/9378 [4:30:52<63:23:07, 26.05s/it]
+
+
+ 7%|██████▎ | 618/9378 [4:31:19<63:57:37, 26.29s/it]
+
+
+ 7%|██████▎ | 619/9378 [4:31:45<63:27:43, 26.08s/it]
+
+
+ 7%|██████▎ | 620/9378 [4:32:12<63:59:26, 26.30s/it]
+
+ 7%|██████▎ | 621/9378 [4:32:37<63:11:38, 25.98s/it]
+
+
+ 7%|██████▎ | 622/9378 [4:33:03<63:19:01, 26.03s/it]
+
+
+ 7%|██████▍ | 623/9378 [4:33:29<63:17:36, 26.03s/it]
+
+ 7%|██████▍ | 623/9378 [4:33:30<63:17:36, 26.03s/it]
+
+
+ 7%|██████▍ | 625/9378 [4:34:21<63:12:07, 25.99s/it]
+
+
+ 7%|██████▍ | 626/9378 [4:34:48<63:20:36, 26.06s/it]
+
+
+ 7%|██████▍ | 627/9378 [4:35:14<63:24:02, 26.08s/it]
+
+ 7%|██████▍ | 627/9378 [4:35:15<63:24:02, 26.08s/it]
+[2024-05-12 15:08:11,213] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-627/pytorch_model_fsdp.bin
+[2024-05-12 15:08:24,101] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-627/pytorch_model_fsdp.bin
+[2024-05-12 15:08:48,268] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-627/optimizer.bin
+ 7%|██████▍ | 627/9378 [4:35:15<63:24:02, 26.08s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 7%|██████▎ | 628/9378 [4:37:49<157:14:10, 64.69s/it]
+
+ 7%|██████▎ | 629/9378 [4:38:15<128:43:43, 52.97s/it]
+
+ 7%|██████▍ | 630/9378 [4:38:40<108:32:43, 44.67s/it]
+{'loss': 1.2176, 'grad_norm': 0.65234375, 'learning_rate': 4.393969636875512e-05, 'epoch': 0.2}
+
+ 7%|██████▍ | 631/9378 [4:39:06<94:54:39, 39.06s/it]
+
+ 7%|██████▍ | 632/9378 [4:39:32<85:14:42, 35.09s/it]
+
+ 7%|██████▍ | 633/9378 [4:39:58<79:06:55, 32.57s/it]
+
+ 7%|██████▍ | 634/9378 [4:40:24<74:22:49, 30.62s/it]
+
+
+ 7%|██████▌ | 636/9378 [4:41:17<68:32:20, 28.22s/it]
+
+ 7%|██████▌ | 637/9378 [4:41:43<66:59:47, 27.59s/it]
+
+ 7%|██████▌ | 638/9378 [4:42:09<65:51:40, 27.13s/it]
+
+ 7%|██████▌ | 639/9378 [4:42:35<65:03:24, 26.80s/it]
+{'loss': 1.3138, 'grad_norm': 0.57421875, 'learning_rate': 4.363190216900709e-05, 'epoch': 0.2}
+
+ 7%|██████▌ | 640/9378 [4:43:01<64:22:38, 26.52s/it]
+
+ 7%|██████▌ | 641/9378 [4:43:27<63:58:31, 26.36s/it]
+
+ 7%|██████▌ | 642/9378 [4:43:53<63:43:35, 26.26s/it]
+
+ 7%|██████▌ | 643/9378 [4:44:19<63:37:00, 26.22s/it]
+
+
+ 7%|██████▌ | 645/9378 [4:45:11<63:26:31, 26.15s/it]
+
+ 7%|██████▌ | 646/9378 [4:45:37<63:32:17, 26.20s/it]
+
+ 7%|██████▌ | 647/9378 [4:46:03<63:24:35, 26.15s/it]
+
+ 7%|██████▋ | 648/9378 [4:46:29<63:16:38, 26.09s/it]
+
+ 7%|██████▋ | 649/9378 [4:46:56<63:22:25, 26.14s/it]
+{'loss': 1.2745, 'grad_norm': 0.6171875, 'learning_rate': 4.328230400357687e-05, 'epoch': 0.21}
+
+ 7%|██████▋ | 650/9378 [4:47:41<77:13:28, 31.85s/it]
+
+ 7%|██████▋ | 651/9378 [4:48:07<73:04:37, 30.15s/it]
+
+
+ 7%|██████▋ | 653/9378 [4:48:59<68:47:22, 28.38s/it]
+
+ 7%|██████▋ | 654/9378 [4:49:26<66:36:35, 27.49s/it]
+
+ 7%|██████▋ | 655/9378 [4:49:52<65:37:28, 27.08s/it]
+
+ 7%|██████▋ | 656/9378 [4:50:18<64:55:01, 26.79s/it]
+
+ 7%|██████▋ | 657/9378 [4:50:43<64:28:15, 26.61s/it]
+
+
+ 7%|██████▋ | 658/9378 [4:51:09<63:58:35, 26.41s/it]
+
+
+ 7%|██████▋ | 659/9378 [4:51:35<63:49:55, 26.36s/it]
+
+
+ 7%|██████▊ | 660/9378 [4:52:02<63:57:57, 26.41s/it]
+
+
+ 7%|██████▊ | 661/9378 [4:52:28<63:38:22, 26.28s/it]
+
+
+ 7%|██████▊ | 662/9378 [4:52:54<63:35:36, 26.27s/it]
+
+
+ 7%|██████▊ | 663/9378 [4:53:20<63:29:34, 26.23s/it]
+
+
+ 7%|██████▊ | 664/9378 [4:53:46<63:25:47, 26.20s/it]
+
+ 7%|██████▊ | 664/9378 [4:53:47<63:25:47, 26.20s/it]
+
+ 7%|██████▊ | 665/9378 [4:54:14<63:27:15, 26.22s/it]
+
+
+ 7%|██████▊ | 667/9378 [4:55:06<63:48:07, 26.37s/it]
+
+ 7%|██████▊ | 668/9378 [4:55:32<63:00:22, 26.04s/it]
+
+ 7%|██████▊ | 669/9378 [4:55:58<63:03:48, 26.07s/it]
+
+ 7%|██████▊ | 670/9378 [4:56:24<63:01:02, 26.05s/it]
+
+ 7%|██████▊ | 671/9378 [4:56:50<62:55:56, 26.02s/it]
+
+ 7%|██████▉ | 672/9378 [4:57:16<63:02:27, 26.07s/it]
+
+ 7%|██████▉ | 673/9378 [4:57:42<62:52:37, 26.00s/it]
+
+ 7%|██████▉ | 674/9378 [4:58:08<63:02:26, 26.07s/it]
+
+ 7%|██████▉ | 675/9378 [4:58:34<63:14:39, 26.16s/it]
+
+ 7%|██████▉ | 676/9378 [4:59:00<63:09:58, 26.13s/it]
+
+
+ 7%|██████▉ | 677/9378 [4:59:26<63:10:02, 26.14s/it]
+
+
+ 7%|██████▉ | 678/9378 [4:59:52<63:05:45, 26.11s/it]
+
+
+ 7%|██████▉ | 679/9378 [5:00:18<63:05:09, 26.11s/it]
+
+
+ 7%|██████▉ | 680/9378 [5:00:44<62:56:49, 26.05s/it]
+
+
+ 7%|██████▉ | 681/9378 [5:01:10<63:00:03, 26.08s/it]
+
+ 7%|██████▉ | 681/9378 [5:01:11<63:00:03, 26.08s/it]
+[2024-05-12 15:33:39,731] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 3b4ffc8a-c133-44ed-8b1f-c42411cc58c6)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/5a3f8c701de2594af0559bd7f5ef04654ae014be5081a7a4d0b2dc51ff092214?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T151622Z&X-Amz-Expires=86400&X-Amz-Signature=be11912df1103c3339fc7c5d6550bfe608c3a318c5ed2d0f7e31a643fbc6d404&X-Amz-SignedHeaders=host&partNumber=1&uploadId=.4K.uiPPdiBQK9DgETFDJYWYABdRQW_bWXpmqM3TxvjBO67KGodfrIJQB6wZhZxakDOKT0cCHFylqzwMR5FgxzFHn890l3HdSt5z.dDHwj.teeBl6srECxnsjEZxYfU_&x-id=UploadPart
+ 7%|██████▉ | 681/9378 [5:01:11<63:00:03, 26.08s/it]'(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 3b4ffc8a-c133-44ed-8b1f-c42411cc58c6)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/5a3f8c701de2594af0559bd7f5ef04654ae014be5081a7a4d0b2dc51ff092214?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T151622Z&X-Amz-Expires=86400&X-Amz-Signature=be11912df1103c3339fc7c5d6550bfe608c3a318c5ed2d0f7e31a643fbc6d404&X-Amz-SignedHeaders=host&partNumber=1&uploadId=.4K.uiPPdiBQK9DgETFDJYWYABdRQW_bWXpmqM3TxvjBO67KGodfrIJQB6wZhZxakDOKT0cCHFylqzwMR5FgxzFHn890l3HdSt5z.dDHwj.teeBl6srECxnsjEZxYfU_&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+ 7%|██████▉ | 682/9378 [5:01:36<63:08:06, 26.14s/it]
+
+
+ 7%|██████▉ | 683/9378 [5:02:02<63:07:09, 26.13s/it]
+
+
+ 7%|███████ | 684/9378 [5:02:29<63:10:46, 26.16s/it]
+
+
+ 7%|███████ | 685/9378 [5:02:55<63:11:05, 26.17s/it]
+
+
+ 7%|███████ | 686/9378 [5:03:21<63:08:17, 26.15s/it]
+
+
+ 7%|███████ | 687/9378 [5:03:47<62:58:49, 26.09s/it]
+
+ 7%|███████ | 687/9378 [5:03:48<62:58:49, 26.09s/it]
+
+ 7%|███████ | 688/9378 [5:04:14<63:12:04, 26.18s/it]
+
+ 7%|███████ | 689/9378 [5:04:40<63:02:46, 26.12s/it]
+
+ 7%|███████ | 690/9378 [5:05:06<62:56:06, 26.08s/it]
+
+ 7%|███████ | 691/9378 [5:05:32<62:55:17, 26.08s/it]
+
+
+ 7%|███████ | 693/9378 [5:06:23<62:53:18, 26.07s/it]
+
+
+ 7%|███████ | 694/9378 [5:06:50<63:00:09, 26.12s/it]
+
+ 7%|███████ | 694/9378 [5:06:51<63:00:09, 26.12s/it]
+
+ 7%|███████ | 695/9378 [5:07:17<63:26:21, 26.30s/it]
+
+
+ 7%|███████▏ | 697/9378 [5:08:09<62:42:59, 26.01s/it]
+
+ 7%|███████▏ | 698/9378 [5:08:35<62:46:52, 26.04s/it]
+
+ 7%|███████▏ | 699/9378 [5:09:01<62:50:14, 26.06s/it]
+
+ 7%|███████▏ | 700/9378 [5:09:48<78:20:28, 32.50s/it]
+
+
+ 7%|███████▏ | 701/9378 [5:10:14<73:40:05, 30.56s/it]
+
+
+ 7%|███████▏ | 702/9378 [5:10:40<70:24:51, 29.22s/it]
+
+
+ 7%|███████▏ | 703/9378 [5:11:06<68:07:19, 28.27s/it]
+
+
+ 8%|███████▏ | 704/9378 [5:11:33<66:45:07, 27.70s/it]
+
+ 8%|███████▏ | 705/9378 [5:12:00<65:35:34, 27.23s/it]
+
+ 8%|███████▏ | 706/9378 [5:12:26<64:54:16, 26.94s/it]
+
+ 8%|███████▏ | 707/9378 [5:12:52<64:22:08, 26.72s/it]
+
+ 8%|███████▏ | 708/9378 [5:13:18<63:57:01, 26.55s/it]
+
+ 8%|███████▎ | 709/9378 [5:13:43<63:44:22, 26.47s/it]
+
+
+ 8%|███████▎ | 710/9378 [5:14:09<63:31:33, 26.38s/it]
+
+
+ 8%|███████▎ | 711/9378 [5:14:36<63:23:45, 26.33s/it]
+
+
+ 8%|███████▎ | 712/9378 [5:15:02<63:21:47, 26.32s/it]
+
+ 8%|███████▎ | 712/9378 [5:15:03<63:21:47, 26.32s/it]
+
+
+
+ 8%|███████▎ | 714/9378 [5:15:56<63:09:36, 26.24s/it]
+
+ 8%|███████▎ | 715/9378 [5:16:22<63:08:22, 26.24s/it]
+
+ 8%|███████▎ | 716/9378 [5:16:48<63:07:37, 26.24s/it]
+
+ 8%|███████▎ | 717/9378 [5:17:14<63:06:31, 26.23s/it]
+
+ 8%|███████▎ | 718/9378 [5:17:40<63:16:28, 26.30s/it]
+
+
+ 8%|███████▎ | 719/9378 [5:18:06<63:16:53, 26.31s/it]
+
+ 8%|███████▎ | 719/9378 [5:18:07<63:16:53, 26.31s/it]
+
+ 8%|███████▎ | 720/9378 [5:18:33<63:42:14, 26.49s/it]
+
+
+ 8%|███████▍ | 722/9378 [5:19:26<62:58:43, 26.19s/it]
+
+ 8%|███████▍ | 723/9378 [5:19:52<63:03:30, 26.23s/it]
+
+ 8%|███████▍ | 724/9378 [5:20:19<63:06:00, 26.25s/it]
+
+ 8%|███████▍ | 725/9378 [5:20:44<63:12:21, 26.30s/it]
+
+
+ 8%|███████▍ | 726/9378 [5:21:10<63:16:54, 26.33s/it]
+
+
+ 8%|███████▍ | 727/9378 [5:21:37<63:17:03, 26.33s/it]
+
+
+
+ 8%|███████▍ | 728/9378 [5:22:04<63:17:26, 26.34s/it]
+
+ 8%|███████▍ | 729/9378 [5:22:31<63:22:46, 26.38s/it]
+
+ 8%|███████▍ | 730/9378 [5:22:56<63:30:04, 26.43s/it]
+
+
+ 8%|███████▍ | 731/9378 [5:23:22<63:20:44, 26.37s/it]
+
+
+ 8%|███████▍ | 732/9378 [5:23:48<63:16:28, 26.35s/it]
+
+
+
+ 8%|███████▌ | 733/9378 [5:24:16<63:22:44, 26.39s/it]
+
+
+ 8%|███████▌ | 734/9378 [5:24:43<63:23:50, 26.40s/it]
+
+ 8%|███████▌ | 735/9378 [5:25:09<63:25:07, 26.42s/it]
+
+ 8%|███████▌ | 736/9378 [5:25:36<63:23:31, 26.41s/it]
+
+ 8%|███████▌ | 737/9378 [5:26:01<63:30:44, 26.46s/it]
+
+
+
+ 8%|███████▌ | 738/9378 [5:26:29<63:31:35, 26.47s/it]
+
+ 8%|███████▌ | 739/9378 [5:26:55<63:44:05, 26.56s/it]
+
+ 8%|███████▌ | 740/9378 [5:27:20<63:37:57, 26.52s/it]
+
+
+ 8%|███████▌ | 741/9378 [5:27:47<63:35:29, 26.51s/it]
+
+
+
+ 8%|███████▌ | 742/9378 [5:28:15<63:32:41, 26.49s/it]
+
+
+ 8%|███████▌ | 743/9378 [5:28:41<63:29:17, 26.47s/it]
+
+ 8%|███████▌ | 744/9378 [5:29:08<63:30:39, 26.48s/it]
+
+ 8%|███████▋ | 745/9378 [5:29:33<63:33:27, 26.50s/it]
+
+
+ 8%|███████▋ | 746/9378 [5:29:59<63:27:44, 26.47s/it]
+
+
+ 8%|███████▋ | 747/9378 [5:30:28<63:49:16, 26.62s/it]
+
+ 8%|███████▋ | 748/9378 [5:30:53<63:45:44, 26.60s/it]
+
+
+ 8%|███████▋ | 749/9378 [5:31:20<64:10:15, 26.77s/it]
+
+
+ 8%|███████▋ | 750/9378 [5:32:08<78:36:30, 32.80s/it]
+
+ 8%|███████▋ | 751/9378 [5:32:33<74:01:03, 30.89s/it]
+
+
+
+ 8%|███████▋ | 752/9378 [5:33:01<70:54:14, 29.59s/it]
+
+
+ 8%|███████▋ | 753/9378 [5:33:28<68:37:23, 28.64s/it]
+
+ 8%|███████▋ | 754/9378 [5:33:54<67:08:11, 28.03s/it]
+
+ 8%|███████▋ | 755/9378 [5:34:19<65:59:29, 27.55s/it]
+
+
+
+ 8%|███████▋ | 756/9378 [5:34:48<65:24:36, 27.31s/it]
+
+ 8%|███████▋ | 757/9378 [5:35:14<65:07:40, 27.20s/it]
+
+ 8%|███████▊ | 758/9378 [5:35:40<64:42:22, 27.02s/it]
+
+
+
+ 8%|███████▊ | 759/9378 [5:36:08<64:25:54, 26.91s/it]
+
+
+ 8%|███████▊ | 760/9378 [5:36:34<64:05:48, 26.78s/it]
+
+ 8%|███████▊ | 761/9378 [5:37:01<63:50:55, 26.67s/it]
+
+ 8%|███████▊ | 762/9378 [5:37:26<63:51:31, 26.68s/it]
+
+
+
+ 8%|███████▊ | 763/9378 [5:37:54<63:45:21, 26.64s/it]
+
+ 8%|███████▊ | 764/9378 [5:38:20<63:44:35, 26.64s/it]
+
+ 8%|███████▊ | 765/9378 [5:38:46<63:54:27, 26.71s/it]
+
+
+
+ 8%|███████▊ | 766/9378 [5:39:14<63:57:44, 26.74s/it]
+
+ 8%|███████▊ | 767/9378 [5:39:41<63:57:27, 26.74s/it]
+
+ 8%|███████▊ | 768/9378 [5:40:06<64:01:30, 26.77s/it]
+
+
+
+ 8%|███████▊ | 769/9378 [5:40:34<63:51:46, 26.71s/it]
+
+ 8%|███████▉ | 770/9378 [5:41:01<63:49:31, 26.69s/it]
+
+ 8%|███████▉ | 771/9378 [5:41:26<63:54:19, 26.73s/it]
+
+
+
+ 8%|███████▉ | 772/9378 [5:41:54<63:43:27, 26.66s/it]
+
+ 8%|███████▉ | 773/9378 [5:42:21<63:52:07, 26.72s/it]
+
+
+ 8%|███████▉ | 774/9378 [5:42:48<64:00:46, 26.78s/it]
+
+
+ 8%|███████▉ | 775/9378 [5:43:15<63:53:13, 26.73s/it]
+
+ 8%|███████▉ | 776/9378 [5:43:41<64:24:58, 26.96s/it]
+
+
+ 8%|███████▉ | 777/9378 [5:44:08<63:53:15, 26.74s/it]
+
+ 8%|███████▉ | 778/9378 [5:44:35<63:54:29, 26.75s/it]
+
+ 8%|███████▉ | 779/9378 [5:45:00<63:53:15, 26.75s/it]
+
+
+
+ 8%|███████▉ | 780/9378 [5:45:29<63:57:50, 26.78s/it]
+
+ 8%|███████▉ | 781/9378 [5:45:55<63:53:19, 26.75s/it]
+
+ 8%|████████ | 782/9378 [5:46:21<64:12:39, 26.89s/it]
+
+
+ 8%|████████ | 783/9378 [5:46:49<63:51:12, 26.74s/it]
+
+ 8%|████████ | 784/9378 [5:47:15<63:59:07, 26.80s/it]
+
+
+
+ 8%|████████ | 785/9378 [5:47:43<63:57:19, 26.79s/it]
+
+ 8%|████████ | 786/9378 [5:48:10<64:01:09, 26.82s/it]
+
+ 8%|████████ | 787/9378 [5:48:36<64:33:37, 27.05s/it]
+
+ 8%|████████ | 788/9378 [5:49:03<63:55:34, 26.79s/it]
+
+ 8%|████████ | 789/9378 [5:49:29<64:02:42, 26.84s/it]
+
+
+ 8%|████████ | 790/9378 [5:49:57<64:29:05, 27.03s/it]
+
+ 8%|████████ | 791/9378 [5:50:23<64:10:50, 26.91s/it]
+
+
+
+ 8%|████████ | 792/9378 [5:50:51<63:47:39, 26.75s/it]
+
+ 8%|████████ | 793/9378 [5:51:18<63:54:49, 26.80s/it]
+
+ 8%|████████▏ | 794/9378 [5:51:43<63:57:47, 26.83s/it]
+
+
+ 8%|████████▏ | 795/9378 [5:52:12<64:01:27, 26.85s/it]
+
+ 8%|████████▏ | 796/9378 [5:52:37<64:04:31, 26.88s/it]
+
+
+
+ 8%|████████▏ | 797/9378 [5:53:05<64:02:32, 26.87s/it]
+
+ 9%|████████▏ | 798/9378 [5:53:33<64:19:20, 26.99s/it]
+
+
+ 9%|████████▏ | 799/9378 [5:54:00<64:23:08, 27.02s/it]
+
+ 9%|████████▏ | 800/9378 [5:54:46<78:21:15, 32.88s/it]
+
+
+ 9%|████████▏ | 801/9378 [5:55:13<73:56:56, 31.04s/it]
+
+
+ 9%|████████▏ | 802/9378 [5:55:40<70:55:38, 29.77s/it]
+
+ 9%|████████▏ | 803/9378 [5:56:07<68:50:53, 28.90s/it]
+
+
+ 9%|████████▏ | 804/9378 [5:56:34<67:27:53, 28.33s/it]
+
+ 9%|████████▏ | 805/9378 [5:57:01<66:25:17, 27.89s/it]
+
+
+ 9%|████████▎ | 806/9378 [5:57:28<66:04:27, 27.75s/it]
+
+ 9%|████████▎ | 807/9378 [5:57:55<66:00:19, 27.72s/it]
+
+
+ 9%|████████▎ | 808/9378 [5:58:22<64:52:37, 27.25s/it]
+
+ 9%|████████▎ | 809/9378 [5:58:49<64:38:24, 27.16s/it]
+
+
+ 9%|████████▎ | 810/9378 [5:59:16<64:24:08, 27.06s/it]
+
+
+ 9%|████████▎ | 811/9378 [5:59:43<64:17:00, 27.01s/it]
+
+
+ 9%|████████▎ | 812/9378 [6:00:10<64:23:16, 27.06s/it]
+
+ 9%|████████▎ | 813/9378 [6:00:37<64:24:39, 27.07s/it]
+
+
+ 9%|████████▎ | 814/9378 [6:01:04<64:15:56, 27.02s/it]
+
+
+ 9%|████████▎ | 815/9378 [6:01:31<64:13:38, 27.00s/it]
+
+
+ 9%|████████▎ | 816/9378 [6:01:58<64:13:03, 27.00s/it]
+
+ 9%|████████▎ | 817/9378 [6:02:25<64:25:50, 27.09s/it]
+
+
+ 9%|████████▎ | 818/9378 [6:02:52<64:19:41, 27.05s/it]
+
+
+ 9%|████████▍ | 819/9378 [6:03:19<64:18:14, 27.05s/it]
+
+
+ 9%|████████▍ | 820/9378 [6:03:46<64:46:25, 27.25s/it]
+
+ 9%|████████▍ | 821/9378 [6:04:13<64:10:45, 27.00s/it]
+
+
+ 9%|████████▍ | 822/9378 [6:04:40<64:18:14, 27.06s/it]
+
+ 9%|████████▍ | 823/9378 [6:05:07<64:14:10, 27.03s/it]
+
+
+ 9%|████████▍ | 824/9378 [6:05:34<64:12:51, 27.03s/it]
+
+ 9%|████████▍ | 825/9378 [6:06:02<64:19:54, 27.08s/it]
+
+
+ 9%|████████▍ | 826/9378 [6:06:29<64:50:13, 27.29s/it]
+
+ 9%|████████▍ | 827/9378 [6:06:56<64:09:13, 27.01s/it]
+
+
+ 9%|████████▍ | 828/9378 [6:07:23<64:05:05, 26.98s/it]
+
+ 9%|████████▍ | 829/9378 [6:07:50<64:08:51, 27.01s/it]
+
+ 9%|████████▍ | 830/9378 [6:08:15<64:16:23, 27.07s/it]
+
+
+ 9%|████████▌ | 831/9378 [6:08:44<64:18:15, 27.09s/it]
+
+ 9%|████████▌ | 832/9378 [6:09:10<64:30:16, 27.17s/it]
+
+
+ 9%|████████▌ | 833/9378 [6:09:37<64:18:08, 27.09s/it]
+
+
+ 9%|████████▌ | 834/9378 [6:10:04<64:22:30, 27.12s/it]
+
+
+ 9%|████████▌ | 835/9378 [6:10:31<64:25:39, 27.15s/it]
+
+ 9%|████████▌ | 835/9378 [6:10:33<64:25:39, 27.15s/it]
+
+ 9%|████████▌ | 836/9378 [6:11:00<64:22:32, 27.13s/it]
+[2024-05-12 16:43:55,708] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-836/pytorch_model_fsdp.bin
+[2024-05-12 16:44:08,909] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-836/pytorch_model_fsdp.bin
+[2024-05-12 16:44:32,330] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-836/optimizer.bin
+ 9%|████████▌ | 836/9378 [6:11:00<64:22:32, 27.13s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+{'loss': 1.1725, 'grad_norm': 0.515625, 'learning_rate': 3.5431076936640225e-05, 'epoch': 0.27}
+
+ 9%|████████▍ | 838/9378 [6:13:59<127:23:48, 53.70s/it]
+
+
+ 9%|████████▍ | 839/9378 [6:14:28<108:46:50, 45.86s/it]
+
+ 9%|████████▌ | 840/9378 [6:14:54<95:31:15, 40.28s/it]
+
+
+ 9%|████████▌ | 841/9378 [6:15:21<85:58:48, 36.26s/it]
+
+
+ 9%|████████▌ | 842/9378 [6:15:48<79:31:40, 33.54s/it]
+
+
+ 9%|████████▋ | 843/9378 [6:16:15<74:53:51, 31.59s/it]
+
+
+ 9%|████████▋ | 844/9378 [6:16:43<71:45:17, 30.27s/it]
+
+
+ 9%|████████▋ | 845/9378 [6:17:10<69:28:59, 29.31s/it]
+
+
+ 9%|████████▋ | 846/9378 [6:17:37<67:55:29, 28.66s/it]
+
+
+ 9%|████████▋ | 847/9378 [6:18:05<66:44:43, 28.17s/it]
+
+ 9%|████████▋ | 848/9378 [6:18:31<66:12:12, 27.94s/it]
+
+
+ 9%|████████▋ | 849/9378 [6:18:59<65:58:32, 27.85s/it]
+
+
+ 9%|████████▋ | 850/9378 [6:19:47<80:18:16, 33.90s/it]
+
+
+ 9%|████████▋ | 851/9378 [6:20:15<75:28:42, 31.87s/it]
+{'loss': 1.0295, 'grad_norm': 0.52734375, 'learning_rate': 3.476726204846075e-05, 'epoch': 0.27}
+
+
+ 9%|████████▋ | 853/9378 [6:21:09<69:57:09, 29.54s/it]
+
+ 9%|████████▋ | 853/9378 [6:21:10<69:57:09, 29.54s/it]
+[2024-05-12 16:53:40,536] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/081a1d96a85b406183783828ad620f8efe9e661c8c5b493ef8d486d802ebd5bc?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T165236Z&X-Amz-Expires=86400&X-Amz-Signature=6659f42e8e8f8f54afb9aac9de56318dd3a31bfeb3320e7eb38bda557c86634f&X-Amz-SignedHeaders=host&partNumber=1&uploadId=_KNX1vzaNboTIPv0N_gGzDlpHNgW70rFm1tdtCpBfi_tXVZGvrkm4Jl4XNQ7W5JryrUdemN23ARJp.80Rm177U82HndfjA9UNysSbIHhWHBdcZoie_Id9okLrF7Juxg0&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: dfacc758-cab9-458b-9b93-62c5b12b6bd7)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/081a1d96a85b406183783828ad620f8efe9e661c8c5b493ef8d486d802ebd5bc?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T165236Z&X-Amz-Expires=86400&X-Amz-Signature=6659f42e8e8f8f54afb9aac9de56318dd3a31bfeb3320e7eb38bda557c86634f&X-Amz-SignedHeaders=host&partNumber=1&uploadId=_KNX1vzaNboTIPv0N_gGzDlpHNgW70rFm1tdtCpBfi_tXVZGvrkm4Jl4XNQ7W5JryrUdemN23ARJp.80Rm177U82HndfjA9UNysSbIHhWHBdcZoie_Id9okLrF7Juxg0&x-id=UploadPart
+ 9%|████████▋ | 853/9378 [6:21:10<69:57:09, 29.54s/it]'(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/081a1d96a85b406183783828ad620f8efe9e661c8c5b493ef8d486d802ebd5bc?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T165236Z&X-Amz-Expires=86400&X-Amz-Signature=6659f42e8e8f8f54afb9aac9de56318dd3a31bfeb3320e7eb38bda557c86634f&X-Amz-SignedHeaders=host&partNumber=1&uploadId=_KNX1vzaNboTIPv0N_gGzDlpHNgW70rFm1tdtCpBfi_tXVZGvrkm4Jl4XNQ7W5JryrUdemN23ARJp.80Rm177U82HndfjA9UNysSbIHhWHBdcZoie_Id9okLrF7Juxg0&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: dfacc758-cab9-458b-9b93-62c5b12b6bd7)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/081a1d96a85b406183783828ad620f8efe9e661c8c5b493ef8d486d802ebd5bc?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T165236Z&X-Amz-Expires=86400&X-Amz-Signature=6659f42e8e8f8f54afb9aac9de56318dd3a31bfeb3320e7eb38bda557c86634f&X-Amz-SignedHeaders=host&partNumber=1&uploadId=_KNX1vzaNboTIPv0N_gGzDlpHNgW70rFm1tdtCpBfi_tXVZGvrkm4Jl4XNQ7W5JryrUdemN23ARJp.80Rm177U82HndfjA9UNysSbIHhWHBdcZoie_Id9okLrF7Juxg0&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+ 9%|████████▋ | 854/9378 [6:21:38<68:22:45, 28.88s/it]
+
+ 9%|████████▊ | 855/9378 [6:22:03<67:05:23, 28.34s/it]
+
+
+ 9%|████████▊ | 856/9378 [6:22:31<66:50:28, 28.24s/it]
+
+ 9%|████████▊ | 856/9378 [6:22:32<66:50:28, 28.24s/it]
+
+
+ 9%|████████▊ | 858/9378 [6:23:25<65:02:12, 27.48s/it]
+
+ 9%|████████▊ | 858/9378 [6:23:26<65:02:12, 27.48s/it]
+
+
+ 9%|████████▊ | 860/9378 [6:24:19<64:43:44, 27.36s/it]
+
+
+ 9%|████████▊ | 861/9378 [6:24:47<64:37:16, 27.31s/it]
+
+
+ 9%|████████▊ | 862/9378 [6:25:14<64:22:46, 27.22s/it]
+
+
+ 9%|████████▊ | 863/9378 [6:25:41<64:17:21, 27.18s/it]
+
+ 9%|████████▊ | 863/9378 [6:25:42<64:17:21, 27.18s/it]
+
+
+ 9%|████████▊ | 865/9378 [6:26:35<64:25:35, 27.24s/it]
+
+
+ 9%|████████▊ | 866/9378 [6:27:04<64:27:14, 27.26s/it]
+
+ 9%|████████▉ | 867/9378 [6:27:30<64:31:14, 27.29s/it]
+
+
+ 9%|████████▉ | 868/9378 [6:27:57<64:31:33, 27.30s/it]
+
+
+ 9%|████████▉ | 869/9378 [6:28:26<65:08:23, 27.56s/it]
+
+ 9%|████████▉ | 870/9378 [6:28:52<64:16:42, 27.20s/it]
+
+
+ 9%|████████▉ | 871/9378 [6:29:19<64:18:58, 27.22s/it]
+
+ 9%|████████▉ | 871/9378 [6:29:21<64:18:58, 27.22s/it]
+
+
+ 9%|████████▉ | 873/9378 [6:30:14<64:20:02, 27.23s/it]
+
+
+ 9%|████████▉ | 874/9378 [6:30:42<64:22:05, 27.25s/it]
+{'loss': 1.1979, 'grad_norm': 0.5859375, 'learning_rate': 3.365910642551643e-05, 'epoch': 0.28}
+
+
+ 9%|████████▉ | 876/9378 [6:31:35<64:10:10, 27.17s/it]
+
+ 9%|████████▉ | 876/9378 [6:31:37<64:10:10, 27.17s/it]
+
+
+ 9%|████████▉ | 878/9378 [6:32:30<64:27:20, 27.30s/it]
+
+
+ 9%|████████▉ | 879/9378 [6:32:58<64:31:30, 27.33s/it]
+
+ 9%|████████▉ | 879/9378 [6:32:59<64:31:30, 27.33s/it]
+
+
+ 9%|█████████ | 881/9378 [6:33:52<64:31:00, 27.33s/it]
+
+
+ 9%|█████████ | 882/9378 [6:34:19<64:13:49, 27.22s/it]
+
+ 9%|█████████ | 882/9378 [6:34:21<64:13:49, 27.22s/it]
+
+
+ 9%|█████████ | 884/9378 [6:35:15<64:35:29, 27.38s/it]
+
+
+ 9%|█████████ | 885/9378 [6:35:42<64:30:09, 27.34s/it]
+
+
+ 9%|█████████ | 886/9378 [6:36:09<64:42:29, 27.43s/it]
+
+
+ 9%|█████████ | 887/9378 [6:36:37<64:36:45, 27.39s/it]
+
+
+ 9%|█████████ | 888/9378 [6:37:04<64:34:23, 27.38s/it]
+
+ 9%|█████████ | 888/9378 [6:37:05<64:34:23, 27.38s/it]
+
+
+ 9%|█████████ | 890/9378 [6:37:58<64:15:33, 27.25s/it]
+
+
+ 10%|█████████ | 891/9378 [6:38:26<64:35:05, 27.40s/it]
+
+
+ 10%|█████████▏ | 892/9378 [6:38:53<64:11:15, 27.23s/it]
+
+ 10%|█████████▏ | 892/9378 [6:38:55<64:11:15, 27.23s/it]
+[2024-05-12 17:11:06,771] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 3d906e68-d4eb-469e-a868-9dceb3bec3d3)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/cfc4ecf059021fd74c0f5d627d73e660adf0581d458e11be4109e3354fa73298?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T165234Z&X-Amz-Expires=86400&X-Amz-Signature=611acc64a215d22c2a1076d200f66f92ccb0062f88aee232aae65742bcc9b23a&X-Amz-SignedHeaders=host&partNumber=102&uploadId=ERgSsBxXv0jN5dCjw2tMZqFzoFD4XWLt7w55pOvE806YK_ERdZVebIjbLPfoPuNX7c47Qo43Wnzhu5txK_zxJ6E5.pCu0vcxIpayTjYcTXUyday28BY85ugX_hXMsmHX&x-id=UploadPart
+ 10%|█████████▏ | 892/9378 [6:38:55<64:11:15, 27.23s/it]'(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 3d906e68-d4eb-469e-a868-9dceb3bec3d3)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/cfc4ecf059021fd74c0f5d627d73e660adf0581d458e11be4109e3354fa73298?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T165234Z&X-Amz-Expires=86400&X-Amz-Signature=611acc64a215d22c2a1076d200f66f92ccb0062f88aee232aae65742bcc9b23a&X-Amz-SignedHeaders=host&partNumber=102&uploadId=ERgSsBxXv0jN5dCjw2tMZqFzoFD4XWLt7w55pOvE806YK_ERdZVebIjbLPfoPuNX7c47Qo43Wnzhu5txK_zxJ6E5.pCu0vcxIpayTjYcTXUyday28BY85ugX_hXMsmHX&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+ 10%|█████████▏ | 893/9378 [6:39:20<64:11:43, 27.24s/it]
+
+ 10%|█████████▏ | 893/9378 [6:39:22<64:11:43, 27.24s/it]
+
+
+ 10%|█████████▏ | 895/9378 [6:40:15<64:14:02, 27.26s/it]
+
+
+ 10%|█████████▏ | 896/9378 [6:40:42<64:18:01, 27.29s/it]
+
+
+ 10%|█████████▏ | 897/9378 [6:41:11<64:27:21, 27.36s/it]
+
+ 10%|█████████▏ | 898/9378 [6:41:37<64:17:55, 27.30s/it]
+
+
+ 10%|█████████▏ | 899/9378 [6:42:06<64:12:40, 27.26s/it]
+
+ 10%|█████████▏ | 900/9378 [6:42:51<77:53:41, 33.08s/it]
+
+
+ 10%|█████████▏ | 901/9378 [6:43:20<74:16:05, 31.54s/it]
+{'loss': 1.1288, 'grad_norm': 0.51171875, 'learning_rate': 3.233329404077137e-05, 'epoch': 0.29}
+
+
+ 10%|█████████▏ | 903/9378 [6:44:13<69:08:47, 29.37s/it]
+
+
+ 10%|█████████▎ | 904/9378 [6:44:40<67:00:05, 28.46s/it]
+
+
+ 10%|█████████▎ | 905/9378 [6:45:07<66:08:02, 28.10s/it]
+
+
+ 10%|█████████▎ | 906/9378 [6:45:36<66:10:39, 28.12s/it]
+{'loss': 1.3147, 'grad_norm': 0.58203125, 'learning_rate': 3.208512668876022e-05, 'epoch': 0.29}
+
+
+ 10%|█████████▎ | 908/9378 [6:46:30<65:26:04, 27.81s/it]
+
+
+ 10%|█████████▎ | 909/9378 [6:46:58<65:11:57, 27.71s/it]
+
+ 10%|█████████▎ | 910/9378 [6:47:26<65:02:44, 27.65s/it]
+
+ 10%|█████████▎ | 911/9378 [6:47:52<64:16:03, 27.33s/it]
+
+
+ 10%|█████████▎ | 912/9378 [6:48:20<64:14:11, 27.32s/it]
+
+
+ 10%|█████████▎ | 913/9378 [6:48:48<64:18:47, 27.35s/it]
+{'loss': 1.1475, 'grad_norm': 0.498046875, 'learning_rate': 3.173641908749839e-05, 'epoch': 0.29}
+
+
+ 10%|█████████▎ | 915/9378 [6:49:42<63:59:18, 27.22s/it]
+
+ 10%|█████████▍ | 916/9378 [6:50:08<63:55:13, 27.19s/it]
+
+
+ 10%|█████████▍ | 917/9378 [6:50:35<64:01:28, 27.24s/it]
+
+
+ 10%|█████████▍ | 918/9378 [6:51:04<64:05:19, 27.27s/it]
+{'loss': 1.3027, 'grad_norm': 0.5859375, 'learning_rate': 3.148646913061034e-05, 'epoch': 0.29}
+
+
+ 10%|█████████▍ | 920/9378 [6:51:58<64:14:18, 27.34s/it]
+
+
+ 10%|█████████▍ | 921/9378 [6:52:26<64:04:45, 27.28s/it]
+
+ 10%|█████████▍ | 922/9378 [6:52:52<64:05:13, 27.28s/it]
+
+
+ 10%|█████████▍ | 923/9378 [6:53:20<64:15:13, 27.36s/it]
+
+
+ 10%|█████████▍ | 924/9378 [6:53:48<64:17:30, 27.38s/it]
+
+ 10%|█████████▍ | 925/9378 [6:54:15<64:18:58, 27.39s/it]
+
+
+ 10%|█████████▍ | 926/9378 [6:54:42<64:11:20, 27.34s/it]
+
+
+
+ 10%|█████████▍ | 927/9378 [6:55:10<64:04:11, 27.29s/it]
+
+ 10%|█████████▍ | 928/9378 [6:55:36<63:56:58, 27.24s/it]
+
+
+ 10%|█████████▌ | 929/9378 [6:56:05<64:00:17, 27.27s/it]
+
+ 10%|█████████▌ | 930/9378 [6:56:30<63:49:30, 27.20s/it]
+
+
+ 10%|█████████▌ | 931/9378 [6:56:58<63:57:10, 27.26s/it]
+
+
+
+ 10%|█████████▌ | 932/9378 [6:57:26<63:52:09, 27.22s/it]
+
+ 10%|█████████▌ | 933/9378 [6:57:52<63:59:49, 27.28s/it]
+
+
+ 10%|█████████▌ | 934/9378 [6:58:21<64:07:59, 27.34s/it]
+
+ 10%|█████████▌ | 935/9378 [6:58:47<63:50:57, 27.22s/it]
+
+
+ 10%|█████████▌ | 936/9378 [6:59:14<63:57:55, 27.28s/it]
+
+
+ 10%|█████████▌ | 937/9378 [6:59:43<64:29:48, 27.51s/it]
+{'loss': 1.2152, 'grad_norm': 0.578125, 'learning_rate': 3.05305309047446e-05, 'epoch': 0.3}
+
+
+ 10%|█████████▌ | 939/9378 [7:00:37<64:15:17, 27.41s/it]
+
+ 10%|█████████▌ | 939/9378 [7:00:38<64:15:17, 27.41s/it]
+
+ 10%|█████████▌ | 940/9378 [7:01:05<63:58:10, 27.29s/it]
+
+
+ 10%|█████████▋ | 942/9378 [7:01:59<64:10:05, 27.38s/it]
+
+
+ 10%|█████████▋ | 943/9378 [7:02:27<63:43:16, 27.20s/it]
+
+ 10%|█████████▋ | 944/9378 [7:02:53<63:49:39, 27.24s/it]
+
+
+ 10%|█████████▋ | 945/9378 [7:03:20<63:54:53, 27.28s/it]
+
+ 10%|█████████▋ | 945/9378 [7:03:22<63:54:53, 27.28s/it]
+
+ 10%|█████████▋ | 946/9378 [7:03:49<63:59:59, 27.32s/it]
+
+
+ 10%|█████████▋ | 948/9378 [7:04:43<63:55:07, 27.30s/it]
+
+
+
+ 10%|█████████▋ | 949/9378 [7:05:12<63:57:09, 27.31s/it]
+
+ 10%|█████████▋ | 950/9378 [7:06:00<79:31:55, 33.97s/it]
+
+
+ 10%|█████████▋ | 951/9378 [7:06:26<74:28:56, 31.82s/it]
+
+
+ 10%|█████████▋ | 952/9378 [7:06:54<71:18:33, 30.47s/it]
+
+
+ 10%|█████████▊ | 953/9378 [7:07:21<69:05:57, 29.53s/it]
+
+
+ 10%|█████████▊ | 954/9378 [7:07:50<67:52:25, 29.01s/it]
+
+ 10%|█████████▊ | 955/9378 [7:08:16<66:33:20, 28.45s/it]
+
+
+ 10%|█████████▊ | 956/9378 [7:08:43<65:35:23, 28.04s/it]
+
+
+ 10%|█████████▊ | 957/9378 [7:09:12<65:34:04, 28.03s/it]
+
+ 10%|█████████▊ | 958/9378 [7:09:38<64:47:49, 27.70s/it]
+
+
+ 10%|█████████▊ | 959/9378 [7:10:05<64:17:18, 27.49s/it]
+
+
+ 10%|█████████▊ | 960/9378 [7:10:34<64:13:32, 27.47s/it]
+
+ 10%|█████████▊ | 961/9378 [7:11:00<64:08:04, 27.43s/it]
+
+
+ 10%|█████████▊ | 962/9378 [7:11:28<64:37:02, 27.64s/it]
+
+
+ 10%|█████████▊ | 963/9378 [7:11:54<63:48:20, 27.30s/it]
+
+
+ 10%|█████████▊ | 964/9378 [7:12:22<63:47:04, 27.29s/it]
+
+
+ 10%|█████████▉ | 965/9378 [7:12:49<63:57:29, 27.37s/it]
+
+
+ 10%|█████████▉ | 966/9378 [7:13:16<63:46:36, 27.29s/it]
+
+
+ 10%|█████████▉ | 967/9378 [7:13:44<63:49:40, 27.32s/it]
+
+
+ 10%|█████████▉ | 968/9378 [7:14:13<64:21:30, 27.55s/it]
+
+
+ 10%|█████████▉ | 969/9378 [7:14:40<63:48:30, 27.32s/it]
+
+ 10%|█████████▉ | 970/9378 [7:15:06<63:48:45, 27.32s/it]
+
+
+ 10%|█████████▉ | 971/9378 [7:15:35<64:19:00, 27.54s/it]
+
+ 10%|█████████▉ | 972/9378 [7:16:01<63:41:59, 27.28s/it]
+
+
+ 10%|█████████▉ | 973/9378 [7:16:29<64:12:27, 27.50s/it]
+
+ 10%|█████████▉ | 974/9378 [7:16:57<64:08:56, 27.48s/it]
+
+ 10%|█████████▉ | 975/9378 [7:17:23<63:30:25, 27.21s/it]
+
+
+ 10%|█████████▉ | 976/9378 [7:17:51<63:51:17, 27.36s/it]
+
+
+ 10%|██████████ | 977/9378 [7:18:19<63:31:41, 27.22s/it]
+
+ 10%|██████████ | 978/9378 [7:18:45<63:31:49, 27.23s/it]
+
+
+
+ 10%|██████████ | 979/9378 [7:19:13<63:27:11, 27.20s/it]
+
+ 10%|██████████ | 980/9378 [7:19:39<63:34:52, 27.26s/it]
+
+
+ 10%|██████████ | 981/9378 [7:20:07<63:40:13, 27.30s/it]
+
+
+
+ 10%|██████████ | 982/9378 [7:20:35<63:35:29, 27.27s/it]
+
+ 10%|██████████ | 983/9378 [7:21:01<63:37:45, 27.29s/it]
+
+
+ 10%|██████████ | 984/9378 [7:21:30<63:37:42, 27.29s/it]
+
+
+ 11%|██████████ | 985/9378 [7:21:57<63:53:47, 27.41s/it]
+
+ 11%|██████████ | 986/9378 [7:22:23<63:48:22, 27.37s/it]
+
+
+ 11%|██████████ | 987/9378 [7:22:52<63:49:29, 27.38s/it]
+
+
+ 11%|██████████ | 988/9378 [7:23:19<63:35:11, 27.28s/it]
+
+ 11%|██████████ | 989/9378 [7:23:45<63:36:11, 27.29s/it]
+
+
+ 11%|██████████▏ | 990/9378 [7:24:14<63:33:42, 27.28s/it]
+
+
+ 11%|██████████▏ | 991/9378 [7:24:41<63:36:11, 27.30s/it]
+
+ 11%|██████████▏ | 992/9378 [7:25:07<63:47:39, 27.39s/it]
+
+
+ 11%|██████████▏ | 993/9378 [7:25:36<64:17:30, 27.60s/it]
+
+
+ 11%|██████████▏ | 994/9378 [7:26:03<63:36:45, 27.31s/it]
+
+ 11%|██████████▏ | 995/9378 [7:26:29<63:31:01, 27.28s/it]
+
+
+ 11%|██████████▏ | 996/9378 [7:26:58<63:39:47, 27.34s/it]
+
+
+ 11%|██████████▏ | 997/9378 [7:27:25<63:35:13, 27.31s/it]
+
+ 11%|██████████▏ | 998/9378 [7:27:51<63:37:06, 27.33s/it]
+
+
+ 11%|██████████▏ | 999/9378 [7:28:20<64:14:49, 27.60s/it]
+
+ 11%|██████████▏ | 1000/9378 [7:29:07<78:21:07, 33.67s/it]
+
+
+ 11%|██████████▏ | 1001/9378 [7:29:36<74:07:47, 31.86s/it]
+
+
+ 11%|██████████▏ | 1002/9378 [7:30:04<70:53:05, 30.47s/it]
+
+ 11%|██████████▏ | 1003/9378 [7:30:30<68:55:36, 29.63s/it]
+
+
+ 11%|██████████▏ | 1004/9378 [7:30:58<67:05:44, 28.84s/it]
+
+
+ 11%|██████████▏ | 1005/9378 [7:31:26<66:01:45, 28.39s/it]
+
+ 11%|██████████▏ | 1006/9378 [7:31:53<65:14:05, 28.05s/it]
+
+
+ 11%|██████████▏ | 1007/9378 [7:32:20<64:39:20, 27.81s/it]
+
+
+ 11%|██████████▏ | 1008/9378 [7:32:48<64:41:02, 27.82s/it]
+
+ 11%|██████████▏ | 1009/9378 [7:33:14<64:18:08, 27.66s/it]
+
+
+ 11%|██████████▏ | 1010/9378 [7:33:43<64:01:50, 27.55s/it]
+
+
+ 11%|██████████▏ | 1011/9378 [7:34:10<63:55:30, 27.50s/it]
+
+ 11%|██████████▎ | 1012/9378 [7:34:36<63:56:02, 27.51s/it]
+
+
+ 11%|██████████▎ | 1013/9378 [7:35:05<63:41:41, 27.41s/it]
+
+
+ 11%|██████████▎ | 1014/9378 [7:35:32<63:40:03, 27.40s/it]
+
+ 11%|██████████▎ | 1015/9378 [7:35:59<64:01:33, 27.56s/it]
+
+
+ 11%|██████████▎ | 1016/9378 [7:36:27<63:39:12, 27.40s/it]
+
+
+ 11%|██████████▎ | 1017/9378 [7:36:55<63:38:40, 27.40s/it]
+
+ 11%|██████████▎ | 1018/9378 [7:37:20<63:40:21, 27.42s/it]
+
+
+ 11%|██████████▎ | 1019/9378 [7:37:48<63:49:05, 27.48s/it]
+
+
+ 11%|██████████▎ | 1020/9378 [7:38:17<63:45:34, 27.46s/it]
+
+
+ 11%|██████████▎ | 1021/9378 [7:38:44<63:36:37, 27.40s/it]
+
+ 11%|██████████▎ | 1022/9378 [7:39:10<63:36:04, 27.40s/it]
+
+
+
+ 11%|██████████▎ | 1023/9378 [7:39:39<63:33:29, 27.39s/it]
+
+
+ 11%|██████████▎ | 1024/9378 [7:40:07<63:34:13, 27.39s/it]
+
+ 11%|██████████▍ | 1025/9378 [7:40:32<63:36:28, 27.41s/it]
+
+
+ 11%|██████████▍ | 1026/9378 [7:41:02<63:44:22, 27.47s/it]
+
+
+ 11%|██████████▍ | 1027/9378 [7:41:29<63:44:19, 27.48s/it]
+
+
+ 11%|██████████▍ | 1028/9378 [7:41:57<63:46:18, 27.49s/it]
+
+ 11%|██████████▍ | 1029/9378 [7:42:23<63:47:14, 27.50s/it]
+
+
+
+ 11%|██████████▍ | 1030/9378 [7:42:52<63:47:40, 27.51s/it]
+
+
+ 11%|██████████▍ | 1031/9378 [7:43:19<63:43:20, 27.48s/it]
+
+
+ 11%|██████████▍ | 1032/9378 [7:43:47<64:15:52, 27.72s/it]
+
+ 11%|██████████▍ | 1033/9378 [7:44:14<64:02:36, 27.63s/it]
+
+
+ 11%|██████████▍ | 1034/9378 [7:44:42<63:33:27, 27.42s/it]
+
+
+ 11%|██████████▍ | 1035/9378 [7:45:09<63:36:40, 27.45s/it]
+
+
+ 11%|██████████▍ | 1036/9378 [7:45:37<63:38:09, 27.46s/it]
+
+ 11%|██████████▌ | 1037/9378 [7:46:03<64:00:35, 27.63s/it]
+
+
+
+ 11%|██████████▌ | 1038/9378 [7:46:32<63:50:54, 27.56s/it]
+
+
+ 11%|██████████▌ | 1039/9378 [7:47:00<63:38:31, 27.47s/it]
+
+
+ 11%|██████████▌ | 1040/9378 [7:47:27<63:43:29, 27.51s/it]
+
+ 11%|██████████▌ | 1041/9378 [7:47:53<63:54:15, 27.59s/it]
+
+
+ 11%|██████████▌ | 1042/9378 [7:48:23<64:17:31, 27.77s/it]
+
+
+ 11%|██████████▌ | 1043/9378 [7:48:50<63:47:47, 27.55s/it]
+
+
+ 11%|██████████▌ | 1044/9378 [7:49:18<63:43:13, 27.53s/it]
+
+
+ 11%|██████████▌ | 1045/9378 [7:49:45<63:43:57, 27.53s/it]
+{'loss': 1.2138, 'grad_norm': 0.546875, 'learning_rate': 2.4978401769175634e-05, 'epoch': 0.33}
+[2024-05-12 18:22:40,809] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-1045/pytorch_model_fsdp.bin
+[2024-05-12 18:22:53,172] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-1045/pytorch_model_fsdp.bin
+[2024-05-12 18:23:18,100] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-1045/optimizer.bin
+ 11%|██████████▌ | 1045/9378 [7:49:45<63:43:57, 27.53s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 11%|██████████▍ | 1046/9378 [7:52:25<156:01:55, 67.42s/it]
+
+
+ 11%|██████████▍ | 1047/9378 [7:52:53<127:56:19, 55.29s/it]
+
+
+ 11%|██████████▌ | 1048/9378 [7:53:20<108:50:28, 47.04s/it]
+
+
+ 11%|██████████▋ | 1049/9378 [7:53:48<95:15:51, 41.18s/it]
+
+ 11%|██████████▌ | 1050/9378 [7:54:35<100:48:48, 43.58s/it]
+
+
+
+ 11%|██████████▋ | 1051/9378 [7:55:05<89:46:57, 38.82s/it]
+
+
+ 11%|██████████▋ | 1052/9378 [7:55:33<82:00:35, 35.46s/it]
+
+
+ 11%|██████████▋ | 1053/9378 [7:56:00<76:39:21, 33.15s/it]
+
+
+ 11%|██████████▋ | 1054/9378 [7:56:28<73:26:53, 31.77s/it]
+
+
+ 11%|██████████▋ | 1055/9378 [7:56:56<70:05:51, 30.32s/it]
+
+ 11%|██████████▋ | 1056/9378 [7:57:23<68:47:41, 29.76s/it]
+
+ 11%|██████████▋ | 1057/9378 [7:57:51<67:29:37, 29.20s/it]
+
+ 11%|██████████▋ | 1058/9378 [7:58:19<66:32:03, 28.79s/it]
+
+ 11%|██████████▋ | 1059/9378 [7:58:47<65:52:39, 28.51s/it]
+
+ 11%|██████████▋ | 1060/9378 [7:59:15<65:22:26, 28.29s/it]
+
+
+ 11%|██████████▋ | 1061/9378 [7:59:43<64:43:56, 28.02s/it]
+
+
+ 11%|██████████▊ | 1062/9378 [8:00:11<64:23:38, 27.88s/it]
+
+ 11%|██████████▊ | 1063/9378 [8:00:39<65:23:39, 28.31s/it]
+
+ 11%|██████████▊ | 1064/9378 [8:01:05<64:41:14, 28.01s/it]
+
+
+ 11%|██████████▊ | 1065/9378 [8:01:33<64:40:20, 28.01s/it]
+
+
+ 11%|██████████▊ | 1066/9378 [8:02:01<64:28:39, 27.93s/it]
+
+
+ 11%|██████████▊ | 1067/9378 [8:02:29<64:47:44, 28.07s/it]
+
+
+ 11%|██████████▊ | 1068/9378 [8:02:58<63:47:16, 27.63s/it]
+
+
+ 11%|██████████▊ | 1069/9378 [8:03:26<63:51:51, 27.67s/it]
+
+
+ 11%|██████████▊ | 1070/9378 [8:03:54<63:55:11, 27.70s/it]
+
+
+ 11%|██████████▊ | 1071/9378 [8:04:22<64:10:27, 27.81s/it]
+
+
+ 11%|██████████▊ | 1072/9378 [8:04:49<64:06:45, 27.79s/it]
+
+
+ 11%|██████████▊ | 1073/9378 [8:05:17<64:10:57, 27.82s/it]
+
+
+ 11%|██████████▉ | 1074/9378 [8:05:45<64:06:46, 27.79s/it]
+
+ 11%|██████████▉ | 1075/9378 [8:06:12<64:43:31, 28.06s/it]
+
+
+ 11%|██████████▉ | 1076/9378 [8:06:39<63:53:09, 27.70s/it]
+
+
+ 11%|██████████▉ | 1077/9378 [8:07:07<64:17:19, 27.88s/it]
+
+
+
+ 11%|██████████▉ | 1078/9378 [8:07:37<64:14:39, 27.86s/it]
+
+
+ 12%|██████████▉ | 1079/9378 [8:08:05<64:17:07, 27.89s/it]
+
+
+ 12%|██████████▉ | 1080/9378 [8:08:32<64:12:22, 27.86s/it]
+
+
+ 12%|██████████▉ | 1081/9378 [8:09:00<64:05:31, 27.81s/it]
+
+
+ 12%|██████████▉ | 1082/9378 [8:09:28<64:18:09, 27.90s/it]
+
+
+ 12%|██████████▉ | 1083/9378 [8:09:56<64:25:32, 27.96s/it]
+
+
+ 12%|██████████▉ | 1084/9378 [8:10:24<64:22:50, 27.94s/it]
+
+
+ 12%|██████████▉ | 1085/9378 [8:10:52<64:17:02, 27.91s/it]
+
+
+ 12%|███████████ | 1086/9378 [8:11:20<64:42:39, 28.09s/it]
+
+
+ 12%|███████████ | 1087/9378 [8:11:48<63:59:19, 27.78s/it]
+
+
+ 12%|███████████ | 1088/9378 [8:12:15<64:02:22, 27.81s/it]
+
+
+ 12%|███████████ | 1089/9378 [8:12:44<64:42:23, 28.10s/it]
+
+ 12%|███████████ | 1090/9378 [8:13:09<63:57:43, 27.78s/it]
+
+
+
+ 12%|███████████ | 1091/9378 [8:13:39<64:45:29, 28.13s/it]
+
+ 12%|███████████ | 1092/9378 [8:14:07<64:32:28, 28.04s/it]
+
+
+ 12%|███████████ | 1093/9378 [8:14:35<63:48:45, 27.73s/it]
+
+
+ 12%|███████████ | 1094/9378 [8:15:03<63:58:32, 27.80s/it]
+
+ 12%|███████████ | 1095/9378 [8:15:31<64:36:42, 28.08s/it]
+
+
+ 12%|███████████ | 1096/9378 [8:15:59<63:53:16, 27.77s/it]
+
+
+ 12%|███████████ | 1097/9378 [8:16:27<64:06:16, 27.87s/it]
+
+
+ 12%|███████████ | 1098/9378 [8:16:55<64:13:32, 27.92s/it]
+
+
+ 12%|███████████▏ | 1099/9378 [8:17:23<64:12:43, 27.92s/it]
+
+ 12%|███████████▏ | 1100/9378 [8:18:08<77:30:02, 33.70s/it]
+
+ 12%|███████████▏ | 1100/9378 [8:18:10<77:30:02, 33.70s/it]
+
+ 12%|███████████▏ | 1101/9378 [8:18:38<73:21:15, 31.90s/it]
+
+
+ 12%|███████████▏ | 1103/9378 [8:19:31<68:16:25, 29.70s/it]
+
+
+ 12%|███████████▏ | 1104/9378 [8:19:59<66:55:40, 29.12s/it]
+
+
+ 12%|███████████▏ | 1105/9378 [8:20:27<66:19:06, 28.86s/it]
+
+
+ 12%|███████████▏ | 1106/9378 [8:20:55<65:32:28, 28.52s/it]
+
+
+ 12%|███████████▏ | 1107/9378 [8:21:23<65:01:10, 28.30s/it]
+
+
+ 12%|███████████▏ | 1108/9378 [8:21:51<64:44:22, 28.18s/it]
+
+
+ 12%|███████████▏ | 1109/9378 [8:22:19<64:29:39, 28.08s/it]
+
+ 12%|███████████▏ | 1109/9378 [8:22:20<64:29:39, 28.08s/it]
+
+
+ 12%|███████████▎ | 1111/9378 [8:23:15<64:31:30, 28.10s/it]
+
+
+ 12%|███████████▎ | 1112/9378 [8:23:44<65:20:29, 28.46s/it]
+
+
+ 12%|███████████▎ | 1113/9378 [8:24:11<64:31:38, 28.11s/it]
+
+
+ 12%|███████████▎ | 1114/9378 [8:24:39<64:21:35, 28.04s/it]
+
+
+ 12%|███████████▎ | 1115/9378 [8:25:07<64:07:41, 27.94s/it]
+
+ 12%|███████████▎ | 1115/9378 [8:25:09<64:07:41, 27.94s/it]
+
+
+ 12%|███████████▎ | 1117/9378 [8:26:02<63:49:44, 27.82s/it]
+
+
+ 12%|███████████▎ | 1118/9378 [8:26:30<63:53:51, 27.85s/it]
+
+
+ 12%|███████████▎ | 1119/9378 [8:26:58<63:50:10, 27.83s/it]
+
+
+ 12%|███████████▎ | 1120/9378 [8:27:26<63:47:21, 27.81s/it]
+
+
+ 12%|███████████▎ | 1121/9378 [8:27:54<63:58:05, 27.89s/it]
+
+
+ 12%|███████████▎ | 1122/9378 [8:28:22<64:13:19, 28.00s/it]
+
+
+ 12%|███████████▍ | 1123/9378 [8:28:50<64:04:23, 27.94s/it]
+
+
+ 12%|███████████▍ | 1124/9378 [8:29:17<63:27:55, 27.68s/it]
+
+
+ 12%|███████████▍ | 1125/9378 [8:29:45<63:43:02, 27.79s/it]
+
+
+ 12%|███████████▍ | 1126/9378 [8:30:12<63:27:58, 27.69s/it]
+
+
+ 12%|███████████▍ | 1127/9378 [8:30:41<63:47:45, 27.83s/it]
+
+
+ 12%|███████████▍ | 1128/9378 [8:31:09<63:52:18, 27.87s/it]
+
+
+ 12%|███████████▍ | 1129/9378 [8:31:37<64:00:50, 27.94s/it]
+
+
+ 12%|███████████▍ | 1130/9378 [8:32:05<63:59:25, 27.93s/it]
+
+
+ 12%|███████████▍ | 1131/9378 [8:32:34<63:23:33, 27.67s/it]
+
+ 12%|███████████▍ | 1132/9378 [8:33:01<63:53:03, 27.89s/it]
+
+
+ 12%|███████████▍ | 1133/9378 [8:33:29<63:23:11, 27.68s/it]
+
+ 12%|███████████▍ | 1134/9378 [8:33:55<63:24:57, 27.69s/it]
+
+
+ 12%|███████████▍ | 1135/9378 [8:34:23<63:33:28, 27.76s/it]
+
+
+ 12%|███████████▌ | 1136/9378 [8:34:51<63:33:29, 27.76s/it]
+
+
+ 12%|███████████▌ | 1137/9378 [8:35:20<64:29:56, 28.18s/it]
+
+
+ 12%|███████████▌ | 1138/9378 [8:35:50<64:46:28, 28.30s/it]
+
+ 12%|███████████▌ | 1139/9378 [8:36:17<64:59:22, 28.40s/it]
+
+
+ 12%|███████████▌ | 1140/9378 [8:36:45<64:27:01, 28.16s/it]
+
+
+ 12%|███████████▌ | 1141/9378 [8:37:13<64:39:21, 28.26s/it]
+
+
+ 12%|███████████▌ | 1142/9378 [8:37:41<64:21:49, 28.13s/it]
+
+
+ 12%|███████████▌ | 1143/9378 [8:38:10<63:40:02, 27.83s/it]
+
+
+ 12%|███████████▌ | 1144/9378 [8:38:38<63:37:47, 27.82s/it]
+
+
+ 12%|███████████▌ | 1145/9378 [8:39:06<63:46:58, 27.89s/it]
+
+
+ 12%|███████████▌ | 1146/9378 [8:39:34<63:58:19, 27.98s/it]
+
+
+ 12%|███████████▌ | 1147/9378 [8:40:02<63:54:55, 27.95s/it]
+
+
+ 12%|███████████▋ | 1148/9378 [8:40:30<63:47:28, 27.90s/it]
+
+
+ 12%|███████████▋ | 1149/9378 [8:40:58<63:48:28, 27.91s/it]
+
+ 12%|███████████▋ | 1150/9378 [8:41:45<78:21:45, 34.29s/it]
+
+
+ 12%|███████████▋ | 1151/9378 [8:42:13<73:53:09, 32.33s/it]
+
+
+
+ 12%|███████████▋ | 1152/9378 [8:42:42<70:44:54, 30.96s/it]
+
+
+ 12%|███████████▋ | 1153/9378 [8:43:10<68:35:18, 30.02s/it]
+
+ 12%|███████████▋ | 1154/9378 [8:43:36<66:56:44, 29.30s/it]
+
+
+ 12%|███████████▋ | 1155/9378 [8:44:04<66:25:29, 29.08s/it]
+
+
+ 12%|███████████▋ | 1156/9378 [8:44:32<65:42:04, 28.77s/it]
+
+
+ 12%|███████████▋ | 1157/9378 [8:45:00<64:47:44, 28.37s/it]
+
+
+ 12%|███████████▋ | 1158/9378 [8:45:28<64:44:13, 28.35s/it]
+
+
+ 12%|███████████▋ | 1159/9378 [8:45:55<63:50:12, 27.96s/it]
+
+
+ 12%|███████████▊ | 1160/9378 [8:46:24<64:16:32, 28.16s/it]
+
+
+ 12%|███████████▊ | 1161/9378 [8:46:51<63:30:13, 27.82s/it]
+
+
+
+ 12%|███████████▊ | 1162/9378 [8:47:21<63:23:27, 27.78s/it]
+
+
+ 12%|███████████▊ | 1163/9378 [8:47:48<63:25:11, 27.79s/it]
+
+
+ 12%|███████████▊ | 1164/9378 [8:48:16<63:35:11, 27.87s/it]
+
+
+ 12%|███████████▊ | 1165/9378 [8:48:45<63:53:02, 28.00s/it]
+
+
+ 12%|███████████▊ | 1166/9378 [8:49:13<63:48:59, 27.98s/it]
+
+
+ 12%|███████████▊ | 1167/9378 [8:49:40<63:38:24, 27.90s/it]
+
+ 12%|███████████▊ | 1168/9378 [8:50:06<63:25:59, 27.81s/it]
+
+
+ 12%|███████████▊ | 1169/9378 [8:50:34<63:22:02, 27.79s/it]
+
+
+ 12%|███████████▊ | 1170/9378 [8:51:02<63:30:35, 27.86s/it]
+
+
+ 12%|███████████▊ | 1171/9378 [8:51:29<63:23:44, 27.81s/it]
+
+
+ 12%|███████████▊ | 1172/9378 [8:51:59<63:24:33, 27.82s/it]
+
+ 13%|███████████▉ | 1173/9378 [8:52:27<64:01:44, 28.09s/it]
+
+
+ 13%|███████████▉ | 1174/9378 [8:52:55<63:17:59, 27.78s/it]
+
+
+ 13%|███████████▉ | 1175/9378 [8:53:23<63:54:07, 28.04s/it]
+
+
+ 13%|███████████▉ | 1176/9378 [8:53:51<63:35:41, 27.91s/it]
+
+
+ 13%|███████████▉ | 1177/9378 [8:54:19<63:18:25, 27.79s/it]
+
+
+ 13%|███████████▉ | 1178/9378 [8:54:47<63:22:42, 27.82s/it]
+
+ 13%|███████████▉ | 1179/9378 [8:55:13<63:27:17, 27.86s/it]
+
+
+ 13%|███████████▉ | 1180/9378 [8:55:41<63:26:08, 27.86s/it]
+
+
+ 13%|███████████▉ | 1181/9378 [8:56:09<63:34:45, 27.92s/it]
+
+
+ 13%|███████████▉ | 1182/9378 [8:56:36<63:30:03, 27.89s/it]
+
+
+ 13%|███████████▉ | 1183/9378 [8:57:04<63:22:48, 27.84s/it]
+
+
+
+ 13%|███████████▉ | 1184/9378 [8:57:34<63:25:19, 27.86s/it]
+
+
+ 13%|████████████ | 1185/9378 [8:58:02<63:30:27, 27.91s/it]
+
+ 13%|████████████ | 1186/9378 [8:58:30<63:56:27, 28.10s/it]
+
+
+ 13%|████████████ | 1187/9378 [8:58:58<63:35:50, 27.95s/it]
+
+ 13%|████████████ | 1188/9378 [8:59:24<63:37:15, 27.97s/it]
+
+
+ 13%|████████████ | 1189/9378 [8:59:52<63:32:53, 27.94s/it]
+
+
+ 13%|████████████ | 1190/9378 [9:00:20<63:32:33, 27.94s/it]
+
+
+ 13%|████████████ | 1191/9378 [9:00:48<63:30:36, 27.93s/it]
+
+
+ 13%|████████████ | 1192/9378 [9:01:16<63:30:04, 27.93s/it]
+
+
+ 13%|████████████ | 1193/9378 [9:01:43<63:14:43, 27.82s/it]
+
+
+ 13%|████████████ | 1194/9378 [9:02:11<63:07:34, 27.77s/it]
+
+
+ 13%|████████████ | 1195/9378 [9:02:39<63:11:08, 27.80s/it]
+
+
+ 13%|████████████ | 1196/9378 [9:03:08<63:40:36, 28.02s/it]
+
+
+ 13%|████████████▏ | 1197/9378 [9:03:35<63:16:50, 27.85s/it]
+
+
+ 13%|████████████▏ | 1198/9378 [9:04:05<63:17:10, 27.85s/it]
+
+
+ 13%|████████████▏ | 1199/9378 [9:04:32<63:11:49, 27.82s/it]
+
+ 13%|████████████▏ | 1200/9378 [9:05:20<78:01:55, 34.35s/it]
+
+
+ 13%|████████████▏ | 1201/9378 [9:05:48<73:17:31, 32.27s/it]
+
+
+ 13%|████████████▏ | 1202/9378 [9:06:15<70:10:16, 30.90s/it]
+
+
+
+ 13%|████████████▏ | 1203/9378 [9:06:45<67:56:02, 29.92s/it]
+
+
+ 13%|████████████▏ | 1204/9378 [9:07:13<66:32:19, 29.31s/it]
+
+
+ 13%|████████████▏ | 1205/9378 [9:07:41<65:34:46, 28.89s/it]
+
+
+ 13%|████████████▏ | 1206/9378 [9:08:08<64:54:13, 28.59s/it]
+
+
+ 13%|████████████▏ | 1207/9378 [9:08:36<64:23:51, 28.37s/it]
+
+
+ 13%|████████████▏ | 1208/9378 [9:09:04<64:06:34, 28.25s/it]
+
+ 13%|████████████▏ | 1209/9378 [9:09:30<63:47:08, 28.11s/it]
+
+
+ 13%|████████████▎ | 1210/9378 [9:09:59<64:05:13, 28.25s/it]
+
+
+ 13%|████████████▎ | 1211/9378 [9:10:27<63:51:59, 28.15s/it]
+
+
+ 13%|████████████▎ | 1212/9378 [9:10:54<63:07:33, 27.83s/it]
+
+
+ 13%|████████████▎ | 1213/9378 [9:11:22<63:40:41, 28.08s/it]
+
+
+ 13%|████████████▎ | 1214/9378 [9:11:50<63:00:45, 27.79s/it]
+
+
+
+ 13%|████████████▎ | 1215/9378 [9:12:19<63:00:00, 27.78s/it]
+
+
+ 13%|████████████▎ | 1216/9378 [9:12:47<62:57:05, 27.77s/it]
+
+
+ 13%|████████████▎ | 1217/9378 [9:13:15<63:37:57, 28.07s/it]
+
+
+ 13%|████████████▎ | 1218/9378 [9:13:43<62:54:54, 27.76s/it]
+
+
+ 13%|████████████▎ | 1219/9378 [9:14:11<63:26:42, 27.99s/it]
+
+ 13%|████████████▎ | 1220/9378 [9:14:37<62:52:21, 27.74s/it]
+
+
+ 13%|████████████▎ | 1221/9378 [9:15:04<62:57:11, 27.78s/it]
+
+
+ 13%|████████████▍ | 1222/9378 [9:15:33<63:08:32, 27.87s/it]
+
+
+ 13%|████████████▍ | 1223/9378 [9:16:01<63:37:32, 28.09s/it]
+
+
+
+ 13%|████████████▍ | 1224/9378 [9:16:30<62:57:24, 27.80s/it]
+
+
+ 13%|████████████▍ | 1225/9378 [9:16:58<62:58:42, 27.81s/it]
+
+
+ 13%|████████████▍ | 1226/9378 [9:17:26<62:59:27, 27.82s/it]
+
+
+ 13%|████████████▍ | 1227/9378 [9:17:54<63:00:05, 27.83s/it]
+
+
+ 13%|████████████▍ | 1228/9378 [9:18:22<63:08:39, 27.89s/it]
+
+
+ 13%|████████████▍ | 1229/9378 [9:18:49<63:04:33, 27.87s/it]
+
+
+ 13%|████████████▍ | 1230/9378 [9:19:17<63:07:00, 27.89s/it]
+
+
+ 13%|████████████▍ | 1231/9378 [9:19:45<63:13:59, 27.94s/it]
+
+
+ 13%|████████████▍ | 1232/9378 [9:20:13<63:09:57, 27.92s/it]
+
+
+ 13%|████████████▍ | 1233/9378 [9:20:41<63:22:01, 28.01s/it]
+
+ 13%|████████████▌ | 1234/9378 [9:21:07<63:04:00, 27.88s/it]
+
+
+ 13%|████████████▌ | 1235/9378 [9:21:35<63:00:19, 27.85s/it]
+
+
+ 13%|████████████▌ | 1236/9378 [9:22:03<63:05:05, 27.89s/it]
+
+
+ 13%|████████████▌ | 1237/9378 [9:22:33<63:15:57, 27.98s/it]
+
+ 13%|████████████▌ | 1238/9378 [9:22:59<63:06:54, 27.91s/it]
+
+
+ 13%|████████████▌ | 1239/9378 [9:23:29<63:04:57, 27.90s/it]
+
+
+ 13%|████████████▌ | 1240/9378 [9:23:57<63:04:59, 27.91s/it]
+
+
+ 13%|████████████▌ | 1241/9378 [9:24:25<63:10:40, 27.95s/it]
+
+
+ 13%|████████████▌ | 1242/9378 [9:24:52<63:05:33, 27.92s/it]
+
+
+ 13%|████████████▌ | 1243/9378 [9:25:21<63:08:49, 27.94s/it]
+
+
+ 13%|████████████▌ | 1244/9378 [9:25:48<63:07:08, 27.94s/it]
+
+ 13%|████████████▌ | 1245/9378 [9:26:16<63:33:49, 28.14s/it]
+
+ 13%|████████████▌ | 1246/9378 [9:26:44<63:31:02, 28.12s/it]
+
+
+ 13%|████████████▋ | 1247/9378 [9:27:12<63:08:11, 27.95s/it]
+
+
+ 13%|████████████▋ | 1248/9378 [9:27:40<63:16:24, 28.02s/it]
+
+
+ 13%|████████████▋ | 1249/9378 [9:28:08<63:07:43, 27.96s/it]
+
+ 13%|████████████▋ | 1250/9378 [9:28:56<77:12:49, 34.20s/it]
+
+
+
+ 13%|████████████▋ | 1251/9378 [9:29:25<72:48:07, 32.25s/it]
+
+
+ 13%|████████████▋ | 1252/9378 [9:29:53<69:47:26, 30.92s/it]
+
+
+ 13%|████████████▋ | 1253/9378 [9:30:21<67:59:56, 30.13s/it]
+
+
+ 13%|████████████▋ | 1254/9378 [9:30:49<66:18:47, 29.39s/it]
+{'loss': 1.166, 'grad_norm': 0.45703125, 'learning_rate': 1.4482633828798958e-05, 'epoch': 0.4}
+[2024-05-12 20:03:40,914] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-1254/pytorch_model_fsdp.bin
+[2024-05-12 20:03:52,191] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-1254/pytorch_model_fsdp.bin
+[2024-05-12 20:04:14,469] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-1254/optimizer.bin
+ 13%|████████████▋ | 1254/9378 [9:30:49<66:18:47, 29.39s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 13%|████████████▌ | 1255/9378 [9:33:23<151:19:38, 67.07s/it]
+
+ 13%|████████████▌ | 1256/9378 [9:33:51<125:02:10, 55.42s/it]
+
+ 13%|████████████▌ | 1257/9378 [9:34:19<106:14:51, 47.10s/it]
+
+
+ 13%|████████████▋ | 1258/9378 [9:34:47<92:52:39, 41.18s/it]
+
+
+ 13%|████████████▊ | 1259/9378 [9:35:15<84:01:35, 37.26s/it]
+
+
+ 13%|████████████▊ | 1260/9378 [9:35:43<77:46:19, 34.49s/it]
+
+
+ 13%|████████████▊ | 1261/9378 [9:36:11<73:36:27, 32.65s/it]
+
+
+ 13%|████████████▊ | 1262/9378 [9:36:39<70:06:23, 31.10s/it]
+
+
+ 13%|████████████▊ | 1263/9378 [9:37:07<68:03:54, 30.20s/it]
+
+
+ 13%|████████████▊ | 1264/9378 [9:37:35<66:30:41, 29.51s/it]
+
+
+ 13%|████████████▊ | 1265/9378 [9:38:03<65:32:35, 29.08s/it]
+
+
+ 13%|████████████▊ | 1266/9378 [9:38:31<64:58:31, 28.84s/it]
+
+
+ 14%|████████████▊ | 1267/9378 [9:38:59<64:40:10, 28.70s/it]
+
+
+ 14%|████████████▊ | 1268/9378 [9:39:27<63:45:14, 28.30s/it]
+
+
+ 14%|████████████▊ | 1269/9378 [9:39:55<63:27:05, 28.17s/it]
+
+
+ 14%|████████████▊ | 1270/9378 [9:40:23<63:25:50, 28.16s/it]
+
+
+ 14%|████████████▉ | 1271/9378 [9:40:51<63:39:39, 28.27s/it]
+
+
+ 14%|████████████▉ | 1272/9378 [9:41:20<63:14:48, 28.09s/it]
+
+
+ 14%|████████████▉ | 1273/9378 [9:41:48<63:24:36, 28.16s/it]
+
+
+ 14%|████████████▉ | 1274/9378 [9:42:16<63:14:00, 28.09s/it]
+
+
+ 14%|████████████▉ | 1275/9378 [9:42:44<63:09:24, 28.06s/it]
+
+
+ 14%|████████████▉ | 1276/9378 [9:43:12<63:03:01, 28.02s/it]
+
+
+ 14%|████████████▉ | 1277/9378 [9:43:40<63:38:51, 28.28s/it]
+
+
+ 14%|████████████▉ | 1278/9378 [9:44:08<63:30:45, 28.23s/it]
+
+ 14%|████████████▉ | 1279/9378 [9:44:36<63:31:37, 28.24s/it]
+
+
+ 14%|████████████▉ | 1280/9378 [9:45:04<62:47:43, 27.92s/it]
+
+
+ 14%|████████████▉ | 1281/9378 [9:45:32<62:44:04, 27.89s/it]
+
+
+ 14%|████████████▉ | 1282/9378 [9:46:00<62:52:49, 27.96s/it]
+
+
+ 14%|████████████▉ | 1283/9378 [9:46:28<62:57:02, 28.00s/it]
+
+
+ 14%|█████████████ | 1284/9378 [9:46:56<62:53:20, 27.97s/it]
+
+
+ 14%|█████████████ | 1285/9378 [9:47:24<62:57:31, 28.01s/it]
+
+
+ 14%|█████████████ | 1286/9378 [9:47:52<62:52:57, 27.98s/it]
+
+
+ 14%|█████████████ | 1287/9378 [9:48:20<62:50:32, 27.96s/it]
+
+
+ 14%|█████████████ | 1288/9378 [9:48:48<62:56:29, 28.01s/it]
+
+
+ 14%|█████████████ | 1289/9378 [9:49:16<62:58:40, 28.03s/it]
+
+
+ 14%|█████████████ | 1290/9378 [9:49:44<62:57:00, 28.02s/it]
+
+
+ 14%|█████████████ | 1291/9378 [9:50:12<63:10:49, 28.13s/it]
+
+
+ 14%|█████████████ | 1292/9378 [9:50:40<63:25:40, 28.24s/it]
+
+
+ 14%|█████████████ | 1293/9378 [9:51:08<62:38:14, 27.89s/it]
+
+
+ 14%|█████████████ | 1294/9378 [9:51:36<62:41:10, 27.92s/it]
+
+
+ 14%|█████████████ | 1295/9378 [9:52:04<63:15:41, 28.18s/it]
+
+
+ 14%|█████████████▏ | 1296/9378 [9:52:32<62:43:36, 27.94s/it]
+
+
+ 14%|█████████████▏ | 1297/9378 [9:53:00<62:41:21, 27.93s/it]
+
+
+ 14%|█████████████▏ | 1298/9378 [9:53:28<62:51:51, 28.01s/it]
+
+
+ 14%|█████████████▏ | 1299/9378 [9:53:56<62:58:28, 28.06s/it]
+
+
+ 14%|█████████████▏ | 1300/9378 [9:54:46<77:34:44, 34.57s/it]
+
+ 14%|█████████████▏ | 1301/9378 [9:55:14<73:20:46, 32.69s/it]
+
+
+ 14%|█████████████▏ | 1302/9378 [9:55:42<70:31:08, 31.43s/it]
+
+
+ 14%|█████████████▏ | 1303/9378 [9:56:10<67:36:30, 30.14s/it]
+
+ 14%|█████████████▏ | 1304/9378 [9:56:38<66:39:07, 29.72s/it]
+
+ 14%|█████████████▏ | 1305/9378 [9:57:06<65:30:42, 29.21s/it]
+
+ 14%|█████████████▏ | 1306/9378 [9:57:34<64:17:10, 28.67s/it]
+
+ 14%|█████████████▏ | 1307/9378 [9:58:01<64:02:03, 28.56s/it]
+
+
+ 14%|█████████████▎ | 1308/9378 [9:58:29<63:41:16, 28.41s/it]
+
+
+ 14%|█████████████▎ | 1309/9378 [9:58:58<63:53:30, 28.51s/it]
+
+
+ 14%|█████████████▎ | 1310/9378 [9:59:25<63:18:01, 28.25s/it]
+
+
+ 14%|█████████████▎ | 1311/9378 [9:59:54<63:44:50, 28.45s/it]
+
+
+ 14%|█████████████▏ | 1312/9378 [10:00:21<62:54:44, 28.08s/it]
+
+
+ 14%|█████████████▏ | 1313/9378 [10:00:49<62:53:55, 28.08s/it]
+
+
+ 14%|█████████████▏ | 1314/9378 [10:01:17<62:45:18, 28.02s/it]
+
+
+ 14%|█████████████▏ | 1315/9378 [10:01:45<62:43:08, 28.00s/it]
+
+
+ 14%|█████████████▏ | 1316/9378 [10:02:13<62:49:37, 28.05s/it]
+
+
+ 14%|█████████████▏ | 1317/9378 [10:02:41<62:36:36, 27.96s/it]
+
+
+ 14%|█████████████▏ | 1318/9378 [10:03:09<62:43:16, 28.01s/it]
+
+
+ 14%|█████████████▏ | 1319/9378 [10:03:38<63:26:22, 28.34s/it]
+
+
+ 14%|█████████████▏ | 1320/9378 [10:04:06<62:42:49, 28.02s/it]
+
+
+ 14%|█████████████▏ | 1321/9378 [10:04:34<62:55:54, 28.12s/it]
+
+
+ 14%|█████████████▎ | 1322/9378 [10:05:02<62:35:27, 27.97s/it]
+
+
+ 14%|█████████████▎ | 1323/9378 [10:05:30<62:40:02, 28.01s/it]
+
+
+ 14%|█████████████▎ | 1324/9378 [10:05:58<62:47:07, 28.06s/it]
+
+
+ 14%|█████████████▎ | 1325/9378 [10:06:26<62:45:43, 28.06s/it]
+
+
+ 14%|█████████████▎ | 1326/9378 [10:06:55<63:07:07, 28.22s/it]
+
+
+ 14%|█████████████▎ | 1327/9378 [10:07:22<62:28:18, 27.93s/it]
+
+
+ 14%|█████████████▎ | 1328/9378 [10:07:50<62:36:58, 28.00s/it]
+
+
+ 14%|█████████████▎ | 1329/9378 [10:08:18<62:32:16, 27.97s/it]
+
+
+ 14%|█████████████▎ | 1330/9378 [10:08:46<62:41:08, 28.04s/it]
+
+
+ 14%|█████████████▎ | 1331/9378 [10:09:14<62:36:02, 28.01s/it]
+
+
+ 14%|█████████████▎ | 1332/9378 [10:09:42<62:38:30, 28.03s/it]
+
+
+ 14%|█████████████▎ | 1333/9378 [10:10:12<62:39:50, 28.04s/it]
+
+
+ 14%|█████████████▎ | 1334/9378 [10:10:40<62:32:17, 27.99s/it]
+
+ 14%|█████████████▍ | 1335/9378 [10:11:08<62:37:50, 28.03s/it]
+
+ 14%|█████████████▍ | 1336/9378 [10:11:36<62:52:52, 28.15s/it]
+
+
+ 14%|█████████████▍ | 1337/9378 [10:12:04<62:26:52, 27.96s/it]
+
+ 14%|█████████████▍ | 1338/9378 [10:12:31<63:09:59, 28.28s/it]
+
+
+ 14%|█████████████▍ | 1339/9378 [10:12:59<62:58:38, 28.20s/it]
+
+
+ 14%|█████████████▍ | 1340/9378 [10:13:27<62:25:25, 27.96s/it]
+
+
+ 14%|█████████████▍ | 1341/9378 [10:13:55<62:49:06, 28.14s/it]
+
+
+ 14%|█████████████▍ | 1342/9378 [10:14:23<62:47:27, 28.13s/it]
+
+
+ 14%|█████████████▍ | 1343/9378 [10:14:51<62:42:12, 28.09s/it]
+
+
+ 14%|█████████████▍ | 1344/9378 [10:15:20<63:15:30, 28.35s/it]
+
+
+ 14%|█████████████▍ | 1345/9378 [10:15:48<62:34:51, 28.05s/it]
+
+
+ 14%|█████████████▍ | 1346/9378 [10:16:16<62:44:08, 28.12s/it]
+
+
+ 14%|█████████████▌ | 1347/9378 [10:16:44<62:41:53, 28.11s/it]
+
+
+ 14%|█████████████▌ | 1348/9378 [10:17:12<62:41:51, 28.11s/it]
+
+
+ 14%|█████████████▌ | 1349/9378 [10:17:40<62:40:57, 28.11s/it]
+
+
+ 14%|█████████████▌ | 1350/9378 [10:18:30<77:02:47, 34.55s/it]
+
+
+ 14%|█████████████▌ | 1351/9378 [10:18:58<72:58:07, 32.73s/it]
+
+
+ 14%|█████████████▌ | 1352/9378 [10:19:26<69:25:17, 31.14s/it]
+
+
+ 14%|█████████████▌ | 1353/9378 [10:19:54<67:18:57, 30.20s/it]
+
+
+ 14%|█████████████▌ | 1354/9378 [10:20:22<65:58:52, 29.60s/it]
+
+
+ 14%|█████████████▌ | 1355/9378 [10:20:50<64:55:53, 29.14s/it]
+
+
+ 14%|█████████████▌ | 1356/9378 [10:21:18<64:21:54, 28.88s/it]
+
+
+ 14%|█████████████▌ | 1357/9378 [10:21:46<63:49:03, 28.64s/it]
+
+
+ 14%|█████████████▌ | 1358/9378 [10:22:14<63:27:58, 28.49s/it]
+
+
+ 14%|█████████████▌ | 1359/9378 [10:22:43<63:18:45, 28.42s/it]
+
+
+ 15%|█████████████▋ | 1360/9378 [10:23:11<63:02:13, 28.30s/it]
+
+
+ 15%|█████████████▋ | 1361/9378 [10:23:39<63:03:38, 28.32s/it]
+
+
+ 15%|█████████████▋ | 1362/9378 [10:24:07<62:39:26, 28.14s/it]
+
+
+ 15%|█████████████▋ | 1363/9378 [10:24:35<62:37:45, 28.13s/it]
+
+
+ 15%|█████████████▋ | 1364/9378 [10:25:03<62:36:58, 28.13s/it]
+
+
+ 15%|█████████████▋ | 1365/9378 [10:25:31<62:35:54, 28.12s/it]
+
+
+ 15%|█████████████▋ | 1366/9378 [10:25:59<62:36:04, 28.13s/it]
+
+
+ 15%|█████████████▋ | 1367/9378 [10:26:27<62:30:14, 28.09s/it]
+
+
+ 15%|█████████████▋ | 1368/9378 [10:26:55<62:28:52, 28.08s/it]
+
+
+ 15%|█████████████▋ | 1369/9378 [10:27:23<62:27:06, 28.07s/it]
+
+
+ 15%|█████████████▋ | 1370/9378 [10:27:51<62:26:41, 28.07s/it]
+
+
+ 15%|█████████████▋ | 1371/9378 [10:28:19<62:22:40, 28.05s/it]
+
+ 15%|█████████████▋ | 1371/9378 [10:28:21<62:22:40, 28.05s/it]
+
+ 15%|█████████████▊ | 1372/9378 [10:28:49<63:02:10, 28.35s/it]
+
+
+ 15%|█████████████▊ | 1374/9378 [10:29:44<62:10:03, 27.96s/it]
+
+ 15%|█████████████▊ | 1374/9378 [10:29:45<62:10:03, 27.96s/it]
+
+
+ 15%|█████████████▊ | 1376/9378 [10:30:40<62:07:17, 27.95s/it]
+
+
+ 15%|█████████████▊ | 1377/9378 [10:31:10<62:11:24, 27.98s/it]
+{'loss': 1.3507, 'grad_norm': 0.4921875, 'learning_rate': 9.1011741556488e-06, 'epoch': 0.44}
+
+
+ 15%|█████████████▊ | 1379/9378 [10:32:04<62:12:05, 27.99s/it]
+
+
+ 15%|█████████████▊ | 1380/9378 [10:32:34<62:17:23, 28.04s/it]
+
+ 15%|█████████████▊ | 1381/9378 [10:33:02<62:16:56, 28.04s/it]
+
+ 15%|█████████████▊ | 1382/9378 [10:33:30<62:19:40, 28.06s/it]
+
+ 15%|█████████████▊ | 1383/9378 [10:33:57<62:58:59, 28.36s/it]
+
+
+ 15%|█████████████▊ | 1384/9378 [10:34:25<62:44:43, 28.26s/it]
+
+
+ 15%|█████████████▉ | 1385/9378 [10:34:53<62:35:09, 28.19s/it]
+
+
+ 15%|█████████████▉ | 1386/9378 [10:35:22<62:39:21, 28.22s/it]
+
+ 15%|█████████████▉ | 1386/9378 [10:35:23<62:39:21, 28.22s/it]
+
+
+ 15%|█████████████▉ | 1388/9378 [10:36:18<62:35:25, 28.20s/it]
+
+
+ 15%|█████████████▉ | 1389/9378 [10:36:48<62:36:01, 28.21s/it]
+
+ 15%|█████████████▉ | 1390/9378 [10:37:16<63:18:05, 28.53s/it]
+
+
+ 15%|█████████████▉ | 1391/9378 [10:37:43<62:39:46, 28.24s/it]
+
+
+ 15%|█████████████▉ | 1392/9378 [10:38:11<62:35:48, 28.22s/it]
+
+
+ 15%|█████████████▉ | 1393/9378 [10:38:40<62:29:46, 28.18s/it]
+
+
+ 15%|█████████████▉ | 1394/9378 [10:39:08<62:33:30, 28.21s/it]
+
+
+ 15%|█████████████▉ | 1395/9378 [10:39:36<62:26:55, 28.16s/it]
+
+
+ 15%|█████████████▉ | 1396/9378 [10:40:04<62:35:12, 28.23s/it]
+
+
+ 15%|██████████████ | 1397/9378 [10:40:32<62:33:27, 28.22s/it]
+
+
+
+ 15%|██████████████ | 1398/9378 [10:41:02<62:33:01, 28.22s/it]
+
+ 15%|██████████████ | 1399/9378 [10:41:31<63:11:48, 28.51s/it]
+
+
+ 15%|██████████████ | 1400/9378 [10:42:20<76:41:08, 34.60s/it]
+
+ 15%|██████████████ | 1401/9378 [10:42:47<72:18:01, 32.63s/it]
+
+
+
+ 15%|██████████████ | 1402/9378 [10:43:17<69:25:40, 31.34s/it]
+
+
+ 15%|██████████████ | 1403/9378 [10:43:45<67:21:32, 30.41s/it]
+
+ 15%|██████████████ | 1404/9378 [10:44:13<65:59:46, 29.80s/it]
+
+ 15%|██████████████ | 1405/9378 [10:44:40<64:51:54, 29.29s/it]
+
+
+ 15%|██████████████ | 1406/9378 [10:45:08<64:09:35, 28.97s/it]
+
+
+ 15%|██████████████ | 1407/9378 [10:45:37<64:05:37, 28.95s/it]
+
+
+ 15%|██████████████ | 1408/9378 [10:46:05<63:28:07, 28.67s/it]
+
+
+ 15%|██████████████ | 1409/9378 [10:46:32<62:31:27, 28.25s/it]
+
+
+ 15%|██████████████▏ | 1410/9378 [10:47:01<62:57:53, 28.45s/it]
+
+
+ 15%|██████████████▏ | 1411/9378 [10:47:29<62:19:04, 28.16s/it]
+
+
+ 15%|██████████████▏ | 1412/9378 [10:47:57<62:20:40, 28.17s/it]
+
+
+ 15%|██████████████▏ | 1413/9378 [10:48:25<62:15:38, 28.14s/it]
+
+
+ 15%|██████████████▏ | 1414/9378 [10:48:53<62:19:59, 28.18s/it]
+
+
+ 15%|██████████████▏ | 1415/9378 [10:49:21<62:15:30, 28.15s/it]
+
+
+
+ 15%|██████████████▏ | 1416/9378 [10:49:51<62:27:08, 28.24s/it]
+
+
+ 15%|██████████████▏ | 1417/9378 [10:50:20<62:29:46, 28.26s/it]
+
+ 15%|██████████████▏ | 1418/9378 [10:50:48<62:30:48, 28.27s/it]
+
+ 15%|██████████████▏ | 1419/9378 [10:51:14<62:26:13, 28.24s/it]
+
+
+ 15%|██████████████▏ | 1420/9378 [10:51:43<62:28:58, 28.27s/it]
+
+
+ 15%|██████████████▏ | 1421/9378 [10:52:11<62:26:29, 28.25s/it]
+
+
+ 15%|██████████████▎ | 1422/9378 [10:52:39<62:30:49, 28.29s/it]
+
+
+ 15%|██████████████▎ | 1423/9378 [10:53:08<62:26:10, 28.26s/it]
+
+
+
+ 15%|██████████████▎ | 1424/9378 [10:53:38<62:33:46, 28.32s/it]
+
+
+ 15%|██████████████▎ | 1425/9378 [10:54:06<62:25:51, 28.26s/it]
+
+ 15%|██████████████▎ | 1426/9378 [10:54:34<62:56:17, 28.49s/it]
+
+ 15%|██████████████▎ | 1427/9378 [10:55:02<62:28:25, 28.29s/it]
+
+ 15%|██████████████▎ | 1428/9378 [10:55:30<62:39:40, 28.37s/it]
+
+
+ 15%|██████████████▎ | 1429/9378 [10:55:57<62:11:37, 28.17s/it]
+
+
+ 15%|██████████████▎ | 1430/9378 [10:56:25<62:12:48, 28.18s/it]
+
+
+ 15%|██████████████▎ | 1431/9378 [10:56:54<62:43:28, 28.41s/it]
+
+
+ 15%|██████████████▎ | 1432/9378 [10:57:22<62:07:08, 28.14s/it]
+
+
+ 15%|██████████████▎ | 1433/9378 [10:57:52<62:38:36, 28.38s/it]
+
+ 15%|██████████████▎ | 1434/9378 [10:58:20<62:32:53, 28.35s/it]
+
+
+ 15%|██████████████▍ | 1435/9378 [10:58:48<61:55:15, 28.06s/it]
+
+
+ 15%|██████████████▍ | 1436/9378 [10:59:16<62:04:14, 28.14s/it]
+
+ 15%|██████████████▍ | 1437/9378 [10:59:45<62:07:08, 28.16s/it]
+
+ 15%|██████████████▍ | 1438/9378 [11:00:13<62:01:09, 28.12s/it]
+
+ 15%|██████████████▍ | 1439/9378 [11:00:40<62:36:22, 28.39s/it]
+
+
+ 15%|██████████████▍ | 1440/9378 [11:01:08<61:58:47, 28.11s/it]
+
+
+ 15%|██████████████▍ | 1441/9378 [11:01:36<62:15:00, 28.23s/it]
+
+
+ 15%|██████████████▍ | 1442/9378 [11:02:04<62:09:35, 28.20s/it]
+
+
+
+ 15%|██████████████▍ | 1443/9378 [11:02:34<62:20:07, 28.28s/it]
+
+ 15%|██████████████▍ | 1444/9378 [11:03:02<62:43:06, 28.46s/it]
+
+
+ 15%|██████████████▍ | 1445/9378 [11:03:31<62:09:57, 28.21s/it]
+
+ 15%|██████████████▍ | 1446/9378 [11:03:57<62:14:28, 28.25s/it]
+
+
+ 15%|██████████████▌ | 1447/9378 [11:04:26<62:36:56, 28.42s/it]
+
+
+ 15%|██████████████▌ | 1448/9378 [11:04:54<62:13:14, 28.25s/it]
+
+
+ 15%|██████████████▌ | 1449/9378 [11:05:23<62:43:42, 28.48s/it]
+
+
+ 15%|██████████████▌ | 1450/9378 [11:06:13<76:31:45, 34.75s/it]
+
+
+ 15%|██████████████▌ | 1451/9378 [11:06:41<72:11:27, 32.79s/it]
+
+
+
+ 15%|██████████████▌ | 1452/9378 [11:07:10<69:07:22, 31.40s/it]
+
+
+ 15%|██████████████▌ | 1453/9378 [11:07:39<66:56:25, 30.41s/it]
+
+
+ 16%|██████████████▌ | 1454/9378 [11:08:07<65:34:18, 29.79s/it]
+
+
+ 16%|██████████████▌ | 1455/9378 [11:08:35<64:37:25, 29.36s/it]
+
+
+ 16%|██████████████▌ | 1456/9378 [11:09:04<63:48:44, 29.00s/it]
+
+
+ 16%|██████████████▌ | 1457/9378 [11:09:32<63:21:53, 28.80s/it]
+
+ 16%|██████████████▌ | 1458/9378 [11:10:00<63:04:17, 28.67s/it]
+
+ 16%|██████████████▌ | 1459/9378 [11:10:27<62:44:12, 28.52s/it]
+
+
+ 16%|██████████████▋ | 1460/9378 [11:10:55<62:34:58, 28.45s/it]
+
+
+
+ 16%|██████████████▋ | 1461/9378 [11:11:25<62:25:11, 28.38s/it]
+
+
+ 16%|██████████████▋ | 1462/9378 [11:11:53<62:21:10, 28.36s/it]
+
+ 16%|██████████████▋ | 1463/9378 [11:12:22<62:50:08, 28.58s/it]
+{'loss': 1.1361, 'grad_norm': 0.55078125, 'learning_rate': 5.931202697934357e-06, 'epoch': 0.47}
+[2024-05-12 21:45:14,588] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-1463/pytorch_model_fsdp.bin
+[2024-05-12 21:45:25,815] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-1463/pytorch_model_fsdp.bin
+[2024-05-12 21:45:47,197] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-1463/optimizer.bin
+ 16%|██████████████▋ | 1463/9378 [11:12:22<62:50:08, 28.58s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 16%|██████████████▌ | 1464/9378 [11:14:54<145:14:12, 66.07s/it]
+
+
+ 16%|██████████████▌ | 1465/9378 [11:15:22<120:02:35, 54.61s/it]
+
+
+
+ 16%|██████████████▌ | 1466/9378 [11:15:52<102:34:10, 46.67s/it]
+
+
+ 16%|██████████████▋ | 1467/9378 [11:16:20<90:39:42, 41.26s/it]
+
+ 16%|██████████████▋ | 1468/9378 [11:16:48<82:04:04, 37.35s/it]
+
+ 16%|██████████████▋ | 1469/9378 [11:17:17<75:51:27, 34.53s/it]
+
+ 16%|██████████████▋ | 1470/9378 [11:17:43<71:43:30, 32.65s/it]
+
+
+ 16%|██████████████▋ | 1471/9378 [11:18:12<68:49:58, 31.34s/it]
+
+
+ 16%|██████████████▊ | 1472/9378 [11:18:40<66:50:13, 30.43s/it]
+
+
+ 16%|██████████████▊ | 1473/9378 [11:19:08<65:22:16, 29.77s/it]
+
+
+
+ 16%|██████████████▊ | 1474/9378 [11:19:38<64:24:29, 29.34s/it]
+
+
+ 16%|██████████████▊ | 1475/9378 [11:20:06<63:41:07, 29.01s/it]
+
+ 16%|██████████████▊ | 1476/9378 [11:20:35<63:29:48, 28.93s/it]
+
+ 16%|██████████████▊ | 1477/9378 [11:21:02<63:09:29, 28.78s/it]
+
+
+ 16%|██████████████▊ | 1478/9378 [11:21:30<62:55:08, 28.67s/it]
+
+
+
+ 16%|██████████████▊ | 1479/9378 [11:22:00<62:41:44, 28.57s/it]
+
+
+ 16%|██████████████▊ | 1480/9378 [11:22:29<62:42:32, 28.58s/it]
+
+ 16%|██████████████▊ | 1481/9378 [11:22:56<62:39:46, 28.57s/it]
+
+
+ 16%|██████████████▊ | 1482/9378 [11:23:24<62:45:32, 28.61s/it]
+
+
+ 16%|██████████████▊ | 1483/9378 [11:23:53<62:54:21, 28.68s/it]
+
+
+
+ 16%|██████████████▊ | 1484/9378 [11:24:23<62:22:23, 28.44s/it]
+
+ 16%|██████████████▉ | 1485/9378 [11:24:51<62:30:17, 28.51s/it]
+
+ 16%|██████████████▉ | 1486/9378 [11:25:18<62:25:22, 28.47s/it]
+
+
+ 16%|██████████████▉ | 1487/9378 [11:25:47<62:24:29, 28.47s/it]
+
+
+
+ 16%|██████████████▉ | 1488/9378 [11:26:17<62:27:35, 28.50s/it]
+
+
+ 16%|██████████████▉ | 1489/9378 [11:26:45<62:17:54, 28.43s/it]
+
+ 16%|██████████████▉ | 1490/9378 [11:27:12<62:21:04, 28.46s/it]
+
+
+ 16%|██████████████▉ | 1491/9378 [11:27:41<62:23:20, 28.48s/it]
+
+
+ 16%|██████████████▉ | 1492/9378 [11:28:11<62:55:26, 28.73s/it]
+
+ 16%|██████████████▉ | 1493/9378 [11:28:39<62:40:22, 28.61s/it]
+
+
+ 16%|██████████████▉ | 1494/9378 [11:29:08<61:55:21, 28.28s/it]
+
+ 16%|██████████████▉ | 1495/9378 [11:29:36<62:00:22, 28.32s/it]
+
+ 16%|██████████████▉ | 1496/9378 [11:30:02<61:54:47, 28.28s/it]
+
+
+ 16%|███████████████ | 1497/9378 [11:30:31<62:27:03, 28.53s/it]
+
+
+ 16%|███████████████ | 1498/9378 [11:30:59<61:48:36, 28.24s/it]
+
+
+
+ 16%|███████████████ | 1499/9378 [11:31:29<61:52:52, 28.27s/it]
+
+
+ 16%|███████████████ | 1500/9378 [11:32:19<76:18:04, 34.87s/it]
+
+
+ 16%|███████████████ | 1501/9378 [11:32:47<71:49:45, 32.83s/it]
+
+
+ 16%|███████████████ | 1502/9378 [11:33:16<68:51:04, 31.47s/it]
+
+
+ 16%|███████████████ | 1503/9378 [11:33:44<66:44:22, 30.51s/it]
+
+ 16%|███████████████ | 1504/9378 [11:34:12<65:19:57, 29.87s/it]
+
+ 16%|███████████████ | 1505/9378 [11:34:39<64:27:06, 29.47s/it]
+
+
+ 16%|███████████████ | 1506/9378 [11:35:07<63:32:55, 29.06s/it]
+
+
+ 16%|███████████████ | 1507/9378 [11:35:35<62:58:12, 28.80s/it]
+
+
+
+ 16%|███████████████ | 1508/9378 [11:36:06<62:40:45, 28.67s/it]
+
+
+ 16%|███████████████▏ | 1509/9378 [11:36:34<62:27:07, 28.57s/it]
+
+
+ 16%|███████████████▏ | 1510/9378 [11:37:02<62:18:27, 28.51s/it]
+
+ 16%|███████████████▏ | 1511/9378 [11:37:30<62:45:43, 28.72s/it]
+
+
+ 16%|███████████████▏ | 1512/9378 [11:37:58<62:29:43, 28.60s/it]
+
+
+ 16%|███████████████▏ | 1513/9378 [11:38:26<61:51:06, 28.31s/it]
+
+
+
+ 16%|███████████████▏ | 1514/9378 [11:38:56<61:50:43, 28.31s/it]
+
+
+ 16%|███████████████▏ | 1515/9378 [11:39:24<61:53:00, 28.33s/it]
+
+ 16%|███████████████▏ | 1516/9378 [11:39:52<62:05:49, 28.43s/it]
+
+
+ 16%|███████████████▏ | 1517/9378 [11:40:21<61:46:02, 28.29s/it]
+
+ 16%|███████████████▏ | 1518/9378 [11:40:49<62:12:04, 28.49s/it]
+
+ 16%|███████████████▏ | 1519/9378 [11:41:17<62:16:38, 28.53s/it]
+
+
+
+ 16%|███████████████▏ | 1520/9378 [11:41:46<61:48:20, 28.32s/it]
+
+
+ 16%|███████████████▏ | 1521/9378 [11:42:14<62:02:47, 28.43s/it]
+
+
+ 16%|███████████████▎ | 1522/9378 [11:42:42<61:26:14, 28.15s/it]
+
+
+ 16%|███████████████▎ | 1523/9378 [11:43:11<61:33:59, 28.22s/it]
+
+
+ 16%|███████████████▎ | 1524/9378 [11:43:39<61:47:33, 28.32s/it]
+
+
+ 16%|███████████████▎ | 1525/9378 [11:44:08<61:43:00, 28.29s/it]
+
+ 16%|███████████████▎ | 1526/9378 [11:44:34<61:45:39, 28.32s/it]
+
+
+
+ 16%|███████████████▎ | 1527/9378 [11:45:04<61:48:54, 28.34s/it]
+
+
+ 16%|███████████████▎ | 1528/9378 [11:45:33<61:47:58, 28.34s/it]
+
+
+ 16%|███████████████▎ | 1529/9378 [11:46:01<61:55:55, 28.41s/it]
+
+
+ 16%|███████████████▎ | 1530/9378 [11:46:29<61:50:17, 28.37s/it]
+
+
+ 16%|███████████████▎ | 1531/9378 [11:46:58<61:49:58, 28.37s/it]
+
+ 16%|███████████████▎ | 1532/9378 [11:47:26<61:46:11, 28.34s/it]
+
+
+ 16%|███████████████▎ | 1533/9378 [11:47:54<61:47:10, 28.35s/it]
+
+
+ 16%|███████████████▍ | 1534/9378 [11:48:23<61:49:25, 28.37s/it]
+
+
+ 16%|███████████████▍ | 1535/9378 [11:48:51<61:50:04, 28.38s/it]
+
+
+ 16%|███████████████▍ | 1536/9378 [11:49:20<61:48:57, 28.38s/it]
+
+
+ 16%|███████████████▍ | 1537/9378 [11:49:48<61:51:31, 28.40s/it]
+
+
+ 16%|███████████████▍ | 1538/9378 [11:50:16<61:48:34, 28.38s/it]
+
+
+ 16%|███████████████▍ | 1539/9378 [11:50:45<61:40:35, 28.32s/it]
+
+
+ 16%|███████████████▍ | 1540/9378 [11:51:13<61:43:44, 28.35s/it]
+
+
+ 16%|███████████████▍ | 1541/9378 [11:51:41<61:47:29, 28.38s/it]
+
+
+ 16%|███████████████▍ | 1542/9378 [11:52:10<61:33:56, 28.28s/it]
+
+
+ 16%|███████████████▍ | 1543/9378 [11:52:38<61:39:29, 28.33s/it]
+
+
+ 16%|███████████████▍ | 1544/9378 [11:53:06<61:35:34, 28.30s/it]
+
+
+ 16%|███████████████▍ | 1545/9378 [11:53:35<61:41:54, 28.36s/it]
+
+
+ 16%|███████████████▍ | 1546/9378 [11:54:03<61:46:52, 28.40s/it]
+
+
+ 16%|███████████████▌ | 1547/9378 [11:54:32<61:42:07, 28.37s/it]
+
+ 17%|███████████████▌ | 1548/9378 [11:55:00<62:13:49, 28.61s/it]
+
+ 17%|███████████████▌ | 1549/9378 [11:55:28<61:50:35, 28.44s/it]
+
+ 17%|███████████████▌ | 1550/9378 [11:56:18<75:55:16, 34.92s/it]
+
+
+ 17%|███████████████▌ | 1551/9378 [11:56:47<71:23:10, 32.83s/it]
+
+
+ 17%|███████████████▌ | 1552/9378 [11:57:15<68:25:33, 31.48s/it]
+
+
+ 17%|███████████████▌ | 1553/9378 [11:57:43<66:24:37, 30.55s/it]
+
+
+ 17%|███████████████▌ | 1554/9378 [11:58:12<64:56:09, 29.88s/it]
+
+
+ 17%|███████████████▌ | 1555/9378 [11:58:40<64:02:15, 29.47s/it]
+
+
+ 17%|███████████████▌ | 1556/9378 [11:59:08<63:12:57, 29.09s/it]
+
+ 17%|███████████████▌ | 1557/9378 [11:59:37<62:53:35, 28.95s/it]
+
+
+ 17%|███████████████▌ | 1558/9378 [12:00:05<62:26:16, 28.74s/it]
+
+
+ 17%|███████████████▋ | 1559/9378 [12:00:34<62:06:56, 28.60s/it]
+
+
+ 17%|███████████████▋ | 1560/9378 [12:01:02<62:33:26, 28.81s/it]
+
+ 17%|███████████████▋ | 1561/9378 [12:01:31<62:14:29, 28.66s/it]
+
+
+ 17%|███████████████▋ | 1562/9378 [12:01:59<61:33:43, 28.36s/it]
+
+ 17%|███████████████▋ | 1563/9378 [12:02:26<61:43:11, 28.43s/it]
+
+
+
+ 17%|███████████████▋ | 1564/9378 [12:02:56<62:11:43, 28.65s/it]
+
+
+ 17%|███████████████▋ | 1565/9378 [12:03:24<61:33:29, 28.36s/it]
+
+
+ 17%|███████████████▋ | 1566/9378 [12:03:53<61:33:22, 28.37s/it]
+
+
+ 17%|███████████████▋ | 1567/9378 [12:04:21<61:32:58, 28.37s/it]
+
+ 17%|███████████████▋ | 1568/9378 [12:04:50<61:40:41, 28.43s/it]
+
+ 17%|███████████████▋ | 1569/9378 [12:05:16<61:31:44, 28.37s/it]
+
+
+ 17%|███████████████▋ | 1570/9378 [12:05:45<61:55:26, 28.55s/it]
+
+
+
+ 17%|███████████████▋ | 1571/9378 [12:06:14<61:11:25, 28.22s/it]
+
+
+ 17%|███████████████▊ | 1572/9378 [12:06:43<61:26:23, 28.34s/it]
+
+
+ 17%|███████████████▊ | 1573/9378 [12:07:11<61:31:39, 28.38s/it]
+
+ 17%|███████████████▊ | 1574/9378 [12:07:39<62:10:45, 28.68s/it]
+
+
+ 17%|███████████████▊ | 1575/9378 [12:08:07<61:55:49, 28.57s/it]
+
+
+
+ 17%|███████████████▊ | 1576/9378 [12:08:37<61:22:00, 28.32s/it]
+
+
+ 17%|███████████████▊ | 1577/9378 [12:09:05<61:21:57, 28.32s/it]
+
+
+ 17%|███████████████▊ | 1578/9378 [12:09:34<61:24:32, 28.34s/it]
+
+ 17%|███████████████▊ | 1579/9378 [12:10:02<61:46:20, 28.51s/it]
+
+ 17%|███████████████▊ | 1580/9378 [12:10:30<61:06:12, 28.21s/it]
+
+ 17%|███████████████▊ | 1581/9378 [12:10:57<61:39:26, 28.47s/it]
+
+
+ 17%|███████████████▊ | 1582/9378 [12:11:25<61:00:40, 28.17s/it]
+
+
+
+ 17%|███████████████▊ | 1583/9378 [12:11:55<61:07:26, 28.23s/it]
+
+
+ 17%|███████████████▉ | 1584/9378 [12:12:23<61:10:01, 28.25s/it]
+
+
+ 17%|███████████████▉ | 1585/9378 [12:12:52<61:14:11, 28.29s/it]
+
+
+ 17%|███████████████▉ | 1586/9378 [12:13:20<61:22:53, 28.36s/it]
+
+ 17%|███████████████▉ | 1587/9378 [12:13:48<61:34:13, 28.45s/it]
+
+ 17%|███████████████▉ | 1588/9378 [12:14:15<61:07:22, 28.25s/it]
+
+
+ 17%|███████████████▉ | 1589/9378 [12:14:43<61:12:44, 28.29s/it]
+
+
+
+ 17%|███████████████▉ | 1590/9378 [12:15:14<61:24:14, 28.38s/it]
+
+ 17%|███████████████▉ | 1591/9378 [12:15:42<61:50:35, 28.59s/it]
+
+ 17%|███████████████▉ | 1592/9378 [12:16:10<61:39:50, 28.51s/it]
+
+ 17%|███████████████▉ | 1593/9378 [12:16:39<61:05:12, 28.25s/it]
+
+ 17%|███████████████▉ | 1594/9378 [12:17:05<61:07:22, 28.27s/it]
+
+
+
+ 17%|███████████████▉ | 1595/9378 [12:17:35<61:05:09, 28.26s/it]
+
+
+ 17%|███████████████▉ | 1596/9378 [12:18:03<61:03:47, 28.25s/it]
+
+
+ 17%|████████████████ | 1597/9378 [12:18:32<61:10:44, 28.31s/it]
+
+
+ 17%|████████████████ | 1598/9378 [12:19:00<61:12:59, 28.33s/it]
+
+
+ 17%|████████████████ | 1599/9378 [12:19:29<61:22:55, 28.41s/it]
+
+
+ 17%|████████████████ | 1600/9378 [12:20:19<75:26:04, 34.91s/it]
+
+
+ 17%|████████████████ | 1601/9378 [12:20:47<71:12:53, 32.97s/it]
+
+
+ 17%|████████████████ | 1602/9378 [12:21:16<68:17:08, 31.61s/it]
+
+
+ 17%|████████████████ | 1603/9378 [12:21:44<66:20:18, 30.72s/it]
+
+
+ 17%|████████████████ | 1604/9378 [12:22:12<64:34:54, 29.91s/it]
+
+ 17%|████████████████ | 1605/9378 [12:22:39<63:27:58, 29.39s/it]
+
+
+
+ 17%|████████████████ | 1606/9378 [12:23:09<62:42:12, 29.04s/it]
+
+ 17%|████████████████ | 1607/9378 [12:23:35<62:11:49, 28.81s/it]
+
+
+ 17%|████████████████ | 1608/9378 [12:24:04<61:52:21, 28.67s/it]
+
+
+ 17%|████████████████▏ | 1609/9378 [12:24:34<61:44:06, 28.61s/it]
+
+ 17%|████████████████▏ | 1610/9378 [12:25:00<61:33:10, 28.53s/it]
+
+
+ 17%|████████████████▏ | 1611/9378 [12:25:29<61:25:31, 28.47s/it]
+
+
+ 17%|████████████████▏ | 1612/9378 [12:25:57<61:16:07, 28.40s/it]
+
+
+ 17%|████████████████▏ | 1613/9378 [12:26:25<61:15:14, 28.40s/it]
+
+
+ 17%|████████████████▏ | 1614/9378 [12:26:53<61:01:21, 28.29s/it]
+
+
+ 17%|████████████████▏ | 1615/9378 [12:27:21<60:40:43, 28.14s/it]
+
+
+ 17%|████████████████▏ | 1616/9378 [12:27:49<60:26:20, 28.03s/it]
+
+
+ 17%|████████████████▏ | 1617/9378 [12:28:18<60:48:49, 28.21s/it]
+
+
+ 17%|████████████████▏ | 1618/9378 [12:28:46<59:58:13, 27.82s/it]
+
+ 17%|████████████████▏ | 1619/9378 [12:29:14<59:59:43, 27.84s/it]
+
+ 17%|████████████████▏ | 1620/9378 [12:29:40<59:57:28, 27.82s/it]
+
+
+ 17%|████████████████▏ | 1621/9378 [12:30:08<60:00:44, 27.85s/it]
+
+
+ 17%|████████████████▎ | 1622/9378 [12:30:36<60:06:02, 27.90s/it]
+
+
+ 17%|████████████████▎ | 1623/9378 [12:31:04<60:15:31, 27.97s/it]
+
+
+ 17%|████████████████▎ | 1624/9378 [12:31:32<60:16:59, 27.99s/it]
+
+ 17%|████████████████▎ | 1624/9378 [12:31:34<60:16:59, 27.99s/it]
+
+
+ 17%|████████████████▎ | 1626/9378 [12:32:28<59:49:05, 27.78s/it]
+
+
+ 17%|████████████████▎ | 1627/9378 [12:32:56<59:55:05, 27.83s/it]
+
+
+ 17%|████████████████▎ | 1628/9378 [12:33:24<59:55:53, 27.84s/it]
+
+ 17%|████████████████▎ | 1628/9378 [12:33:25<59:55:53, 27.84s/it]
+
+
+ 17%|████████████████▎ | 1629/9378 [12:33:53<60:14:37, 27.99s/it]
+
+
+ 17%|████████████████▎ | 1630/9378 [12:34:21<59:43:02, 27.75s/it]
+
+
+ 17%|████████████████▎ | 1631/9378 [12:34:49<60:24:29, 28.07s/it]
+
+
+ 17%|████████████████▎ | 1632/9378 [12:35:17<59:55:28, 27.85s/it]
+
+
+ 17%|████████████████▍ | 1634/9378 [12:36:11<59:53:18, 27.84s/it]
+
+
+ 17%|████████████████▍ | 1635/9378 [12:36:39<59:55:50, 27.86s/it]
+
+
+ 17%|████████████████▍ | 1636/9378 [12:37:07<60:07:09, 27.96s/it]
+
+
+ 17%|████████████████▍ | 1637/9378 [12:37:35<60:01:25, 27.91s/it]
+
+
+ 17%|████████████████▍ | 1638/9378 [12:38:03<60:10:19, 27.99s/it]
+
+
+ 17%|████████████████▍ | 1639/9378 [12:38:31<59:51:51, 27.85s/it]
+
+
+ 17%|████████████████▍ | 1640/9378 [12:38:59<59:58:39, 27.90s/it]
+
+
+ 17%|████████████████▍ | 1641/9378 [12:39:26<59:56:10, 27.89s/it]
+
+
+ 18%|████████████████▍ | 1642/9378 [12:39:54<59:54:22, 27.88s/it]
+
+
+ 18%|████████████████▍ | 1643/9378 [12:40:22<59:49:41, 27.85s/it]
+
+
+ 18%|████████████████▍ | 1644/9378 [12:40:50<59:52:19, 27.87s/it]
+
+ 18%|████████████████▍ | 1644/9378 [12:40:51<59:52:19, 27.87s/it]
+
+ 18%|████████████████▍ | 1645/9378 [12:41:19<60:13:15, 28.04s/it]
+
+
+ 18%|████████████████▌ | 1647/9378 [12:42:13<59:35:01, 27.75s/it]
+
+
+ 18%|████████████████▌ | 1648/9378 [12:42:41<59:46:18, 27.84s/it]
+
+
+ 18%|████████████████▌ | 1649/9378 [12:43:09<59:48:54, 27.86s/it]
+
+
+ 18%|████████████████▌ | 1650/9378 [12:43:59<73:42:32, 34.34s/it]
+
+
+ 18%|████████████████▌ | 1651/9378 [12:44:27<69:56:44, 32.59s/it]
+
+
+ 18%|████████████████▌ | 1652/9378 [12:44:55<66:27:40, 30.97s/it]
+
+
+ 18%|████████████████▌ | 1653/9378 [12:45:23<64:54:05, 30.25s/it]
+
+
+ 18%|████████████████▌ | 1654/9378 [12:45:50<62:51:21, 29.30s/it]
+
+
+ 18%|████████████████▌ | 1655/9378 [12:46:20<62:00:09, 28.90s/it]
+{'loss': 1.2425, 'grad_norm': 0.5078125, 'learning_rate': 1.154876266770652e-06, 'epoch': 0.53}
+
+
+ 18%|████████████████▌ | 1657/9378 [12:47:14<60:53:39, 28.39s/it]
+
+
+ 18%|████████████████▌ | 1658/9378 [12:47:42<60:37:07, 28.27s/it]
+
+
+ 18%|████████████████▋ | 1659/9378 [12:48:10<60:18:11, 28.12s/it]
+
+
+ 18%|████████████████▋ | 1660/9378 [12:48:38<60:08:09, 28.05s/it]
+
+
+ 18%|████████████████▋ | 1661/9378 [12:49:06<60:06:57, 28.04s/it]
+
+
+ 18%|████████████████▋ | 1662/9378 [12:49:34<60:24:05, 28.18s/it]
+
+
+ 18%|████████████████▋ | 1663/9378 [12:50:01<59:41:14, 27.85s/it]
+
+
+ 18%|████████████████▋ | 1664/9378 [12:50:29<59:43:13, 27.87s/it]
+
+
+ 18%|████████████████▋ | 1665/9378 [12:50:59<60:16:12, 28.13s/it]
+
+ 18%|████████████████▋ | 1666/9378 [12:51:27<59:39:07, 27.85s/it]
+
+ 18%|████████████████▋ | 1667/9378 [12:51:55<59:46:58, 27.91s/it]
+
+ 18%|████████████████▋ | 1668/9378 [12:52:23<59:45:23, 27.90s/it]
+
+ 18%|████████████████▋ | 1669/9378 [12:52:51<60:15:03, 28.14s/it]
+
+
+ 18%|████████████████▋ | 1670/9378 [12:53:18<59:38:27, 27.86s/it]
+
+
+ 18%|████████████████▋ | 1671/9378 [12:53:46<59:41:02, 27.88s/it]
+{'loss': 1.0679, 'grad_norm': 0.474609375, 'learning_rate': 9.18876987945258e-07, 'epoch': 0.53}
+
+ 18%|████████████████▊ | 1672/9378 [12:54:14<60:12:03, 28.12s/it]
+[2024-05-12 23:27:06,882] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-1672/pytorch_model_fsdp.bin
+[2024-05-12 23:27:18,174] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-1672/pytorch_model_fsdp.bin
+[2024-05-12 23:27:43,416] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-1672/optimizer.bin
+ 18%|████████████████▊ | 1672/9378 [12:54:14<60:12:03, 28.12s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 18%|████████████████▌ | 1673/9378 [12:56:51<143:06:43, 66.87s/it]
+
+
+ 18%|████████████████▌ | 1674/9378 [12:57:19<118:06:32, 55.19s/it]
+
+
+ 18%|████████████████▌ | 1675/9378 [12:57:46<100:28:22, 46.96s/it]
+
+
+ 18%|████████████████▊ | 1676/9378 [12:58:14<88:12:57, 41.23s/it]
+
+
+ 18%|████████████████▊ | 1677/9378 [12:58:43<80:02:37, 37.42s/it]
+
+
+ 18%|████████████████▊ | 1678/9378 [12:59:11<73:49:25, 34.51s/it]
+
+ 18%|████████████████▊ | 1679/9378 [12:59:39<69:37:47, 32.56s/it]
+
+ 18%|████████████████▊ | 1680/9378 [13:00:07<66:32:27, 31.12s/it]
+{'loss': 1.075, 'grad_norm': 0.435546875, 'learning_rate': 7.977577177164436e-07, 'epoch': 0.54}
+
+
+ 18%|████████████████▊ | 1682/9378 [13:01:03<62:41:35, 29.33s/it]
+
+ 18%|████████████████▊ | 1683/9378 [13:01:30<61:46:31, 28.90s/it]
+
+ 18%|████████████████▊ | 1683/9378 [13:01:31<61:46:31, 28.90s/it]
+
+ 18%|████████████████▉ | 1684/9378 [13:01:59<61:30:42, 28.78s/it]
+
+ 18%|████████████████▉ | 1685/9378 [13:02:27<60:59:04, 28.54s/it]
+
+
+ 18%|████████████████▉ | 1687/9378 [13:03:21<59:48:47, 28.00s/it]
+
+ 18%|████████████████▉ | 1687/9378 [13:03:23<59:48:47, 28.00s/it]
+
+
+ 18%|████████████████▉ | 1689/9378 [13:04:17<59:32:02, 27.87s/it]
+
+
+ 18%|████████████████▉ | 1690/9378 [13:04:45<59:34:31, 27.90s/it]
+
+
+ 18%|████████████████▉ | 1691/9378 [13:05:13<59:33:54, 27.90s/it]
+
+
+ 18%|████████████████▉ | 1692/9378 [13:05:41<59:35:18, 27.91s/it]
+
+
+ 18%|████████████████▉ | 1693/9378 [13:06:09<59:40:31, 27.95s/it]
+
+
+ 18%|████████████████▉ | 1694/9378 [13:06:37<59:44:31, 27.99s/it]
+
+
+ 18%|████████████████▉ | 1695/9378 [13:07:05<59:45:35, 28.00s/it]
+
+
+ 18%|████████████████▉ | 1696/9378 [13:07:34<60:25:18, 28.32s/it]
+
+
+ 18%|█████████████████ | 1697/9378 [13:08:01<59:46:19, 28.01s/it]
+
+
+ 18%|█████████████████ | 1698/9378 [13:08:30<59:55:58, 28.09s/it]
+
+
+ 18%|█████████████████ | 1699/9378 [13:08:57<59:40:42, 27.98s/it]
+
+
+ 18%|█████████████████ | 1700/9378 [13:09:47<73:27:57, 34.45s/it]
+
+
+ 18%|█████████████████ | 1701/9378 [13:10:16<69:15:04, 32.47s/it]
+
+ 18%|█████████████████ | 1702/9378 [13:10:44<66:19:35, 31.11s/it]
+
+ 18%|█████████████████ | 1703/9378 [13:11:12<64:30:29, 30.26s/it]
+
+ 18%|█████████████████ | 1704/9378 [13:11:38<62:40:55, 29.41s/it]
+
+
+ 18%|█████████████████ | 1705/9378 [13:12:06<61:38:12, 28.92s/it]
+
+ 18%|█████████████████ | 1705/9378 [13:12:08<61:38:12, 28.92s/it]
+
+
+ 18%|█████████████████ | 1707/9378 [13:13:02<60:20:24, 28.32s/it]
+
+
+ 18%|█████████████████ | 1708/9378 [13:13:30<59:59:45, 28.16s/it]
+
+
+ 18%|█████████████████▏ | 1709/9378 [13:13:58<60:15:02, 28.28s/it]
+
+
+ 18%|█████████████████▏ | 1710/9378 [13:14:26<59:55:25, 28.13s/it]
+
+
+ 18%|█████████████████▏ | 1711/9378 [13:14:54<59:55:28, 28.14s/it]
+
+
+ 18%|█████████████████▏ | 1712/9378 [13:15:21<59:16:43, 27.84s/it]
+
+
+ 18%|█████████████████▏ | 1713/9378 [13:15:51<59:16:12, 27.84s/it]
+
+
+ 18%|█████████████████▏ | 1714/9378 [13:16:19<59:13:10, 27.82s/it]
+
+ 18%|█████████████████▏ | 1715/9378 [13:16:45<59:12:42, 27.82s/it]
+
+
+ 18%|█████████████████▏ | 1716/9378 [13:17:13<59:13:54, 27.83s/it]
+
+
+ 18%|█████████████████▏ | 1717/9378 [13:17:40<59:12:26, 27.82s/it]
+
+
+ 18%|█████████████████▏ | 1718/9378 [13:18:09<59:49:29, 28.12s/it]
+
+
+ 18%|█████████████████▏ | 1719/9378 [13:18:37<59:20:15, 27.89s/it]
+
+
+ 18%|█████████████████▏ | 1720/9378 [13:19:04<59:19:06, 27.89s/it]
+
+
+ 18%|█████████████████▎ | 1721/9378 [13:19:32<59:17:56, 27.88s/it]
+
+
+ 18%|█████████████████▎ | 1722/9378 [13:20:02<59:16:20, 27.87s/it]
+
+ 18%|█████████████████▎ | 1723/9378 [13:20:30<59:11:40, 27.84s/it]
+
+
+ 18%|█████████████████▎ | 1724/9378 [13:20:57<59:13:05, 27.85s/it]
+
+
+ 18%|█████████████████▎ | 1725/9378 [13:21:25<59:10:46, 27.84s/it]
+
+
+ 18%|█████████████████▎ | 1726/9378 [13:21:53<59:10:18, 27.84s/it]
+
+ 18%|█████████████████▎ | 1727/9378 [13:22:21<59:44:19, 28.11s/it]
+
+
+ 18%|█████████████████▎ | 1728/9378 [13:22:49<59:17:59, 27.91s/it]
+
+ 18%|█████████████████▎ | 1729/9378 [13:23:15<59:12:24, 27.87s/it]
+
+ 18%|█████████████████▎ | 1729/9378 [13:23:17<59:12:24, 27.87s/it]'(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 0a2cfb58-ce94-4049-b53a-8db0c766b276)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/81934053cfe9ff5ea842f204f6d138f19dd060e81cc180eb5499c349967a372a?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T233518Z&X-Amz-Expires=86400&X-Amz-Signature=fb83a740246e8cfffd6236483efe7dbf9955966b9bcfb36b0867c161c8ddae3a&X-Amz-SignedHeaders=host&partNumber=1&uploadId=iqIRdjKJkOCqtF8awUS2XubFoOZmH5WBHge7m56l8iADgdhlvDBDWvD7B1wGtp43Nit6_daYz3ulnzbCB7m9QmJSpTZHTgCs_perN6lyXan2IfJkLV9kpb3p.klnZHch&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+[2024-05-12 23:55:51,954] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 0a2cfb58-ce94-4049-b53a-8db0c766b276)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/81934053cfe9ff5ea842f204f6d138f19dd060e81cc180eb5499c349967a372a?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240512%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240512T233518Z&X-Amz-Expires=86400&X-Amz-Signature=fb83a740246e8cfffd6236483efe7dbf9955966b9bcfb36b0867c161c8ddae3a&X-Amz-SignedHeaders=host&partNumber=1&uploadId=iqIRdjKJkOCqtF8awUS2XubFoOZmH5WBHge7m56l8iADgdhlvDBDWvD7B1wGtp43Nit6_daYz3ulnzbCB7m9QmJSpTZHTgCs_perN6lyXan2IfJkLV9kpb3p.klnZHch&x-id=UploadPart
+[2024-05-12 23:55:51,955] [WARNING] [huggingface_hub.utils._http.http_backoff:302] [PID:44754] Retrying in 1s [Retry 1/5].
+ 18%|█████████████████▎ | 1730/9378 [13:23:44<59:44:28, 28.12s/it]
+
+
+ 18%|█████████████████▎ | 1731/9378 [13:24:11<58:59:33, 27.77s/it]
+
+
+ 18%|█████████████████▎ | 1732/9378 [13:24:39<59:08:05, 27.84s/it]
+
+
+ 18%|█████████████████▎ | 1733/9378 [13:25:07<59:09:41, 27.86s/it]
+
+
+ 18%|█████████████████▍ | 1734/9378 [13:25:35<59:15:06, 27.91s/it]
+
+
+ 19%|█████████████████▍ | 1735/9378 [13:26:03<59:11:08, 27.88s/it]
+
+
+ 19%|█████████████████▍ | 1736/9378 [13:26:32<59:11:01, 27.88s/it]
+
+
+ 19%|█████████████████▍ | 1737/9378 [13:27:00<59:07:25, 27.86s/it]
+
+
+ 19%|█████████████████▍ | 1738/9378 [13:27:28<59:07:30, 27.86s/it]
+
+
+ 19%|█████████████████▍ | 1739/9378 [13:27:56<59:08:09, 27.87s/it]
+
+
+ 19%|█████████████████▍ | 1740/9378 [13:28:24<59:03:59, 27.84s/it]
+
+ 19%|█████████████████▍ | 1741/9378 [13:28:50<58:59:27, 27.81s/it]
+
+
+
+ 19%|█████████████████▍ | 1742/9378 [13:29:20<59:09:25, 27.89s/it]
+
+ 19%|█████████████████▍ | 1743/9378 [13:29:46<59:10:08, 27.90s/it]
+
+
+ 19%|█████████████████▍ | 1744/9378 [13:30:14<59:25:48, 28.03s/it]
+
+
+ 19%|█████████████████▍ | 1745/9378 [13:30:42<59:08:43, 27.90s/it]
+
+
+ 19%|█████████████████▌ | 1746/9378 [13:31:10<59:08:36, 27.90s/it]
+
+
+ 19%|█████████████████▌ | 1747/9378 [13:31:37<59:05:53, 27.88s/it]
+
+
+ 19%|█████████████████▌ | 1748/9378 [13:32:05<59:09:58, 27.92s/it]
+
+
+ 19%|█████████████████▌ | 1749/9378 [13:32:34<59:44:32, 28.19s/it]
+
+
+
+ 19%|█████████████████▌ | 1750/9378 [13:33:24<72:31:43, 34.23s/it]
+
+
+ 19%|█████████████████▌ | 1751/9378 [13:33:52<68:29:31, 32.33s/it]
+
+ 19%|█████████████████▌ | 1752/9378 [13:34:18<65:44:07, 31.03s/it]
+
+
+ 19%|█████████████████▌ | 1753/9378 [13:34:46<63:37:49, 30.04s/it]
+
+
+ 19%|█████████████████▌ | 1754/9378 [13:35:14<62:14:41, 29.39s/it]
+
+
+ 19%|█████████████████▌ | 1755/9378 [13:35:43<61:48:47, 29.19s/it]
+
+
+ 19%|█████████████████▌ | 1756/9378 [13:36:10<60:29:54, 28.57s/it]
+
+
+ 19%|█████████████████▌ | 1757/9378 [13:36:38<60:12:33, 28.44s/it]
+
+
+ 19%|█████████████████▌ | 1758/9378 [13:37:06<59:47:00, 28.24s/it]
+
+
+ 19%|█████████████████▋ | 1759/9378 [13:37:34<59:32:53, 28.14s/it]
+
+
+ 19%|█████████████████▋ | 1760/9378 [13:38:03<59:22:26, 28.06s/it]
+
+
+ 19%|█████████████████▋ | 1761/9378 [13:38:31<59:09:08, 27.96s/it]
+
+
+ 19%|█████████████████▋ | 1762/9378 [13:38:59<59:10:08, 27.97s/it]
+{'loss': 1.039, 'grad_norm': 0.4453125, 'learning_rate': 8.665018423966997e-08, 'epoch': 0.56}
+
+
+ 19%|█████████████████▋ | 1764/9378 [13:39:53<59:00:31, 27.90s/it]
+
+
+ 19%|█████████████████▋ | 1765/9378 [13:40:21<58:57:50, 27.88s/it]
+
+ 19%|█████████████████▋ | 1765/9378 [13:40:23<58:57:50, 27.88s/it]
+
+
+ 19%|█████████████████▋ | 1767/9378 [13:41:17<58:58:22, 27.89s/it]
+
+
+ 19%|█████████████████▋ | 1768/9378 [13:41:45<58:52:20, 27.85s/it]
+
+
+ 19%|█████████████████▋ | 1769/9378 [13:42:13<58:53:00, 27.86s/it]
+
+
+ 19%|█████████████████▋ | 1770/9378 [13:42:41<58:58:53, 27.91s/it]
+
+
+ 19%|█████████████████▊ | 1771/9378 [13:43:09<59:24:46, 28.12s/it]
+
+
+ 19%|█████████████████▊ | 1772/9378 [13:43:38<58:55:04, 27.89s/it]
+
+
+ 19%|█████████████████▊ | 1773/9378 [13:44:06<58:42:51, 27.79s/it]
+
+ 19%|█████████████████▊ | 1774/9378 [13:44:34<59:23:50, 28.12s/it]
+
+
+ 19%|█████████████████▊ | 1775/9378 [13:45:02<58:51:51, 27.87s/it]
+
+ 19%|█████████████████▊ | 1776/9378 [13:45:30<58:59:37, 27.94s/it]
+
+ 19%|█████████████████▊ | 1777/9378 [13:45:58<59:23:48, 28.13s/it]
+
+
+ 19%|█████████████████▊ | 1778/9378 [13:46:26<58:41:20, 27.80s/it]
+
+
+ 19%|█████████████████▊ | 1779/9378 [13:46:54<58:47:47, 27.85s/it]
+
+
+ 19%|█████████████████▊ | 1780/9378 [13:47:22<59:10:47, 28.04s/it]
+
+ 19%|█████████████████▊ | 1781/9378 [13:47:48<58:34:01, 27.75s/it]
+
+
+ 19%|█████████████████▊ | 1782/9378 [13:48:16<59:05:55, 28.01s/it]
+
+
+ 19%|█████████████████▊ | 1783/9378 [13:48:44<59:01:15, 27.98s/it]
+
+
+ 19%|█████████████████▉ | 1784/9378 [13:49:12<58:58:14, 27.96s/it]
+
+
+ 19%|█████████████████▉ | 1785/9378 [13:49:40<58:57:41, 27.95s/it]
+
+
+ 19%|█████████████████▉ | 1786/9378 [13:50:07<58:27:58, 27.72s/it]
+
+
+ 19%|█████████████████▉ | 1787/9378 [13:50:37<58:26:44, 27.72s/it]
+
+ 19%|█████████████████▉ | 1788/9378 [13:51:05<58:40:47, 27.83s/it]
+
+ 19%|█████████████████▉ | 1789/9378 [13:51:33<59:05:31, 28.03s/it]
+
+
+ 19%|█████████████████▉ | 1790/9378 [13:52:01<58:41:46, 27.85s/it]
+
+
+ 19%|█████████████████▉ | 1791/9378 [13:52:28<58:40:06, 27.84s/it]
+
+
+ 19%|█████████████████▉ | 1792/9378 [13:52:56<58:33:28, 27.79s/it]
+
+
+ 19%|█████████████████▉ | 1793/9378 [13:53:24<58:43:32, 27.87s/it]
+
+
+ 19%|█████████████████▉ | 1794/9378 [13:53:52<58:44:09, 27.88s/it]
+
+
+ 19%|█████████████████▉ | 1795/9378 [13:54:20<58:50:31, 27.94s/it]
+
+
+ 19%|██████████████████ | 1796/9378 [13:54:48<59:03:28, 28.04s/it]
+
+ 19%|██████████████████ | 1797/9378 [13:55:14<58:40:15, 27.86s/it]
+
+
+ 19%|██████████████████ | 1798/9378 [13:55:42<58:41:53, 27.88s/it]
+
+
+ 19%|██████████████████ | 1799/9378 [13:56:11<59:20:09, 28.18s/it]
+
+
+ 19%|██████████████████ | 1800/9378 [13:57:00<72:25:12, 34.40s/it]
+
+
+ 19%|██████████████████ | 1801/9378 [13:57:28<68:28:30, 32.53s/it]
+
+
+
+ 19%|██████████████████ | 1802/9378 [13:57:57<65:12:55, 30.99s/it]
+
+
+ 19%|██████████████████ | 1803/9378 [13:58:25<63:02:13, 29.96s/it]
+
+ 19%|██████████████████ | 1804/9378 [13:58:51<61:26:38, 29.20s/it]
+
+
+ 19%|██████████████████ | 1805/9378 [13:59:20<60:51:28, 28.93s/it]
+
+ 19%|██████████████████ | 1806/9378 [13:59:47<59:35:37, 28.33s/it]
+
+
+ 19%|██████████████████ | 1807/9378 [14:00:15<59:22:29, 28.23s/it]
+
+ 19%|██████████████████ | 1808/9378 [14:00:43<59:20:54, 28.22s/it]
+
+
+ 19%|██████████████████▏ | 1809/9378 [14:01:11<58:56:17, 28.03s/it]
+
+
+ 19%|██████████████████▏ | 1810/9378 [14:01:39<58:56:12, 28.04s/it]
+
+
+ 19%|██████████████████▏ | 1811/9378 [14:02:07<58:44:42, 27.95s/it]
+
+
+ 19%|██████████████████▏ | 1812/9378 [14:02:35<58:45:17, 27.96s/it]
+
+
+ 19%|██████████████████▏ | 1813/9378 [14:03:03<58:51:59, 28.01s/it]
+
+
+ 19%|██████████████████▏ | 1814/9378 [14:03:31<58:40:32, 27.93s/it]
+
+ 19%|██████████████████▏ | 1815/9378 [14:03:57<58:44:38, 27.96s/it]
+
+
+ 19%|██████████████████▏ | 1816/9378 [14:04:25<58:41:30, 27.94s/it]
+
+
+ 19%|██████████████████▏ | 1817/9378 [14:04:53<58:36:37, 27.91s/it]
+
+
+ 19%|██████████████████▏ | 1818/9378 [14:05:21<58:33:54, 27.89s/it]
+
+
+ 19%|██████████████████▏ | 1819/9378 [14:05:49<58:37:53, 27.92s/it]
+
+
+ 19%|██████████████████▏ | 1820/9378 [14:06:18<58:31:43, 27.88s/it]
+
+ 19%|██████████████████▎ | 1821/9378 [14:06:46<58:31:22, 27.88s/it]
+
+
+ 19%|██████████████████▎ | 1822/9378 [14:07:14<58:28:25, 27.86s/it]
+
+
+ 19%|██████████████████▎ | 1823/9378 [14:07:42<58:26:42, 27.85s/it]
+
+
+ 19%|██████████████████▎ | 1824/9378 [14:08:10<58:31:33, 27.89s/it]
+
+
+ 19%|██████████████████▎ | 1825/9378 [14:08:37<58:23:14, 27.83s/it]
+
+
+ 19%|██████████████████▎ | 1826/9378 [14:09:05<58:29:09, 27.88s/it]
+
+ 19%|██████████████████▎ | 1827/9378 [14:09:32<58:27:23, 27.87s/it]
+
+
+ 19%|██████████████████▎ | 1828/9378 [14:10:01<58:22:12, 27.83s/it]
+
+ 20%|██████████████████▎ | 1829/9378 [14:10:29<58:23:28, 27.85s/it]
+
+ 20%|██████████████████▎ | 1830/9378 [14:10:57<58:52:42, 28.08s/it]
+
+
+ 20%|██████████████████▎ | 1831/9378 [14:11:25<58:12:31, 27.77s/it]
+
+
+ 20%|██████████████████▎ | 1832/9378 [14:11:52<58:11:25, 27.76s/it]
+
+
+ 20%|██████████████████▎ | 1833/9378 [14:12:20<58:12:26, 27.77s/it]
+
+
+ 20%|██████████████████▍ | 1834/9378 [14:12:48<58:14:47, 27.80s/it]
+
+
+ 20%|██████████████████▍ | 1835/9378 [14:13:16<58:17:43, 27.82s/it]
+
+
+ 20%|██████████████████▍ | 1836/9378 [14:13:44<58:20:06, 27.84s/it]
+
+ 20%|██████████████████▍ | 1837/9378 [14:14:10<58:19:25, 27.84s/it]
+
+
+ 20%|██████████████████▍ | 1838/9378 [14:14:38<58:23:12, 27.88s/it]
+
+
+ 20%|██████████████████▍ | 1839/9378 [14:15:08<58:58:51, 28.16s/it]
+
+ 20%|██████████████████▍ | 1840/9378 [14:15:35<58:18:42, 27.85s/it]
+
+ 20%|██████████████████▍ | 1841/9378 [14:16:03<58:42:24, 28.04s/it]
+
+ 20%|██████████████████▍ | 1842/9378 [14:16:31<58:21:05, 27.87s/it]
+
+
+ 20%|██████████████████▍ | 1843/9378 [14:16:59<58:08:04, 27.78s/it]
+
+
+ 20%|██████████████████▍ | 1844/9378 [14:17:27<58:12:41, 27.82s/it]
+
+
+ 20%|██████████████████▍ | 1845/9378 [14:17:55<58:25:05, 27.92s/it]
+
+
+ 20%|██████████████████▌ | 1846/9378 [14:18:23<58:27:05, 27.94s/it]
+
+
+ 20%|██████████████████▌ | 1847/9378 [14:18:51<58:29:10, 27.96s/it]
+
+
+ 20%|██████████████████▌ | 1848/9378 [14:19:19<58:20:16, 27.89s/it]
+
+
+ 20%|██████████████████▌ | 1849/9378 [14:19:47<58:15:09, 27.85s/it]
+
+ 20%|██████████████████▌ | 1850/9378 [14:20:34<70:25:38, 33.68s/it]
+
+ 20%|██████████████████▌ | 1851/9378 [14:21:02<67:18:39, 32.19s/it]
+
+ 20%|██████████████████▌ | 1852/9378 [14:21:30<64:33:36, 30.88s/it]
+
+
+ 20%|██████████████████▌ | 1853/9378 [14:21:57<62:14:05, 29.77s/it]
+
+ 20%|██████████████████▌ | 1854/9378 [14:22:25<61:25:42, 29.39s/it]
+
+
+ 20%|██████████████████▌ | 1855/9378 [14:22:53<60:26:31, 28.92s/it]
+
+
+ 20%|██████████████████▌ | 1856/9378 [14:23:21<59:25:06, 28.44s/it]
+
+
+ 20%|██████████████████▌ | 1857/9378 [14:23:49<59:00:30, 28.24s/it]
+
+
+ 20%|██████████████████▌ | 1858/9378 [14:24:17<58:44:37, 28.12s/it]
+
+ 20%|██████████████████▋ | 1859/9378 [14:24:43<58:40:43, 28.09s/it]
+
+
+ 20%|██████████████████▋ | 1860/9378 [14:25:12<58:34:17, 28.05s/it]
+
+ 20%|██████████████████▋ | 1861/9378 [14:25:40<58:43:38, 28.13s/it]
+
+ 20%|██████████████████▋ | 1862/9378 [14:26:08<58:32:11, 28.04s/it]
+
+
+ 20%|██████████████████▋ | 1863/9378 [14:26:36<57:55:17, 27.75s/it]
+
+
+ 20%|██████████████████▋ | 1864/9378 [14:27:04<58:04:58, 27.83s/it]
+
+
+ 20%|██████████████████▋ | 1865/9378 [14:27:32<58:10:59, 27.88s/it]
+
+
+ 20%|██████████████████▋ | 1866/9378 [14:28:00<58:15:02, 27.92s/it]
+
+ 20%|██████████████████▋ | 1867/9378 [14:28:28<58:38:31, 28.11s/it]
+
+
+ 20%|██████████████████▋ | 1868/9378 [14:28:56<57:58:44, 27.79s/it]
+
+
+ 20%|██████████████████▋ | 1869/9378 [14:29:24<58:06:20, 27.86s/it]
+
+
+ 20%|██████████████████▋ | 1870/9378 [14:29:52<58:09:52, 27.89s/it]
+
+
+ 20%|██████████████████▊ | 1871/9378 [14:30:20<58:40:15, 28.14s/it]
+
+
+ 20%|██████████████████▊ | 1872/9378 [14:30:48<58:37:31, 28.12s/it]
+
+
+ 20%|██████████████████▊ | 1873/9378 [14:31:16<58:27:15, 28.04s/it]
+
+ 20%|██████████████████▊ | 1874/9378 [14:31:43<58:24:59, 28.03s/it]
+
+
+ 20%|██████████████████▊ | 1875/9378 [14:32:12<58:11:46, 27.92s/it]
+
+
+ 20%|██████████████████▊ | 1876/9378 [14:32:39<57:52:27, 27.77s/it]
+
+
+ 20%|██████████████████▊ | 1877/9378 [14:33:07<57:57:49, 27.82s/it]
+
+ 20%|██████████████████▊ | 1878/9378 [14:33:35<57:57:13, 27.82s/it]
+
+
+ 20%|██████████████████▊ | 1879/9378 [14:34:03<57:56:20, 27.81s/it]
+
+
+ 20%|██████████████████▊ | 1880/9378 [14:34:31<57:57:10, 27.82s/it]
+
+
+ 20%|██████████████████▊ | 1881/9378 [14:34:59<57:59:05, 27.84s/it]
+{'loss': 1.0608, 'grad_norm': 0.484375, 'learning_rate': 3.332280594009529e-07, 'epoch': 0.6}
+[2024-05-13 01:07:52,845] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-1881/pytorch_model_fsdp.bin
+[2024-05-13 01:08:04,983] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-1881/pytorch_model_fsdp.bin
+[2024-05-13 01:08:29,767] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-1881/optimizer.bin
+ 20%|██████████████████▊ | 1881/9378 [14:34:59<57:59:05, 27.84s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 20%|██████████████████▋ | 1882/9378 [14:37:36<139:06:06, 66.80s/it]
+
+ 20%|██████████████████▋ | 1883/9378 [14:38:03<114:44:08, 55.11s/it]
+
+
+ 20%|██████████████████▉ | 1884/9378 [14:38:31<97:13:55, 46.71s/it]
+
+ 20%|██████████████████▉ | 1885/9378 [14:38:59<85:54:57, 41.28s/it]
+
+
+ 20%|██████████████████▉ | 1886/9378 [14:39:27<77:13:58, 37.11s/it]
+
+
+ 20%|██████████████████▉ | 1887/9378 [14:39:55<71:11:31, 34.21s/it]
+
+
+ 20%|██████████████████▉ | 1888/9378 [14:40:23<67:26:29, 32.42s/it]
+
+
+ 20%|██████████████████▉ | 1889/9378 [14:40:51<64:39:30, 31.08s/it]
+
+
+ 20%|██████████████████▉ | 1890/9378 [14:41:19<62:48:48, 30.20s/it]
+
+
+ 20%|██████████████████▉ | 1891/9378 [14:41:47<61:19:25, 29.49s/it]
+
+
+ 20%|██████████████████▉ | 1892/9378 [14:42:15<60:29:09, 29.09s/it]
+
+
+ 20%|██████████████████▉ | 1893/9378 [14:42:44<60:09:04, 28.93s/it]
+
+ 20%|██████████████████▉ | 1894/9378 [14:43:12<59:42:19, 28.72s/it]
+
+ 20%|██████████████████▉ | 1895/9378 [14:43:40<59:16:34, 28.52s/it]
+
+
+ 20%|███████████████████ | 1896/9378 [14:44:08<58:54:12, 28.34s/it]
+
+
+ 20%|███████████████████ | 1897/9378 [14:44:36<58:42:07, 28.25s/it]
+
+
+ 20%|███████████████████ | 1898/9378 [14:45:04<58:28:57, 28.15s/it]
+
+
+ 20%|███████████████████ | 1899/9378 [14:45:32<58:26:34, 28.13s/it]
+
+
+ 20%|███████████████████ | 1900/9378 [14:46:22<72:02:53, 34.68s/it]
+
+
+ 20%|███████████████████ | 1901/9378 [14:46:50<67:54:11, 32.69s/it]
+
+
+ 20%|███████████████████ | 1902/9378 [14:47:18<65:07:18, 31.36s/it]
+
+
+ 20%|███████████████████ | 1903/9378 [14:47:45<62:40:14, 30.18s/it]
+
+
+ 20%|███████████████████ | 1904/9378 [14:48:14<61:17:51, 29.53s/it]
+
+
+ 20%|███████████████████ | 1905/9378 [14:48:42<60:26:51, 29.12s/it]
+
+
+ 20%|███████████████████ | 1906/9378 [14:49:10<59:38:11, 28.73s/it]
+
+
+ 20%|███████████████████ | 1907/9378 [14:49:38<59:14:19, 28.54s/it]
+
+
+ 20%|███████████████████ | 1908/9378 [14:50:06<58:51:16, 28.36s/it]
+
+ 20%|███████████████████▏ | 1909/9378 [14:50:32<58:28:03, 28.18s/it]
+
+
+ 20%|███████████████████▏ | 1910/9378 [14:51:00<58:21:09, 28.13s/it]
+
+
+ 20%|███████████████████▏ | 1911/9378 [14:51:29<58:13:06, 28.07s/it]
+{'loss': 1.2289, 'grad_norm': 0.498046875, 'learning_rate': 6.338246209904891e-07, 'epoch': 0.61}
+
+
+ 20%|███████████████████▏ | 1913/9378 [14:52:23<58:00:38, 27.98s/it]
+
+
+ 20%|███████████████████▏ | 1914/9378 [14:52:51<57:53:40, 27.92s/it]
+
+ 20%|███████████████████▏ | 1914/9378 [14:52:53<57:53:40, 27.92s/it]
+
+
+ 20%|███████████████████▏ | 1916/9378 [14:53:47<57:53:33, 27.93s/it]
+
+
+ 20%|███████████████████▏ | 1917/9378 [14:54:15<57:58:43, 27.98s/it]
+
+
+ 20%|███████████████████▏ | 1918/9378 [14:54:43<57:50:07, 27.91s/it]
+
+
+ 20%|███████████████████▏ | 1919/9378 [14:55:11<57:53:33, 27.94s/it]
+
+
+ 20%|███████████████████▏ | 1920/9378 [14:55:39<57:54:51, 27.96s/it]
+
+
+ 20%|███████████████████▎ | 1921/9378 [14:56:07<58:00:02, 28.00s/it]
+
+
+ 20%|███████████████████▎ | 1922/9378 [14:56:35<57:54:29, 27.96s/it]
+
+
+ 21%|███████████████████▎ | 1923/9378 [14:57:03<57:50:22, 27.93s/it]
+
+
+ 21%|███████████████████▎ | 1924/9378 [14:57:31<57:46:25, 27.90s/it]
+
+
+ 21%|███████████████████▎ | 1925/9378 [14:57:59<57:40:17, 27.86s/it]
+
+
+ 21%|███████████████████▎ | 1926/9378 [14:58:27<57:44:54, 27.90s/it]
+
+
+ 21%|███████████████████▎ | 1927/9378 [14:58:54<57:45:09, 27.90s/it]
+
+
+ 21%|███████████████████▎ | 1928/9378 [14:59:22<57:42:57, 27.89s/it]
+
+
+
+ 21%|███████████████████▎ | 1929/9378 [14:59:52<57:38:35, 27.86s/it]
+
+ 21%|███████████████████▎ | 1930/9378 [15:00:20<57:46:57, 27.93s/it]
+
+ 21%|███████████████████▎ | 1931/9378 [15:00:46<57:41:40, 27.89s/it]
+
+
+ 21%|███████████████████▎ | 1932/9378 [15:01:14<57:44:35, 27.92s/it]
+
+
+ 21%|███████████████████▍ | 1933/9378 [15:01:42<57:43:29, 27.91s/it]
+
+ 21%|███████████████████▍ | 1933/9378 [15:01:43<57:43:29, 27.91s/it]
+[2024-05-13 01:33:57,906] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 61c8194b-44e6-4870-9cfc-419ac9d7b90b)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/3bc96ab500694c84fd4e6598dc105ba82aea8a141bc92c688a703ac239f88f27?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T011630Z&X-Amz-Expires=86400&X-Amz-Signature=9808dcdf27ea1180569c6992fc46196a630f45b030a52d5e82b8f84d65bbc56f&X-Amz-SignedHeaders=host&partNumber=1&uploadId=XTUj6q15M25OQqAYZA5tRJx6ySofFY3VzqVub0EY7OnlRpnp3qMvtoKNQx2kNYm9C75NUDPoUFfGA1NPT3KJNj1jl1Cm0PvODnAmY7JwPCVSXxagRvsHhuFhn2pzzefo&x-id=UploadPart
+ 21%|███████████████████▍ | 1933/9378 [15:01:43<57:43:29, 27.91s/it]'(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 61c8194b-44e6-4870-9cfc-419ac9d7b90b)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/3bc96ab500694c84fd4e6598dc105ba82aea8a141bc92c688a703ac239f88f27?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T011630Z&X-Amz-Expires=86400&X-Amz-Signature=9808dcdf27ea1180569c6992fc46196a630f45b030a52d5e82b8f84d65bbc56f&X-Amz-SignedHeaders=host&partNumber=1&uploadId=XTUj6q15M25OQqAYZA5tRJx6ySofFY3VzqVub0EY7OnlRpnp3qMvtoKNQx2kNYm9C75NUDPoUFfGA1NPT3KJNj1jl1Cm0PvODnAmY7JwPCVSXxagRvsHhuFhn2pzzefo&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+ 21%|███████████████████▍ | 1934/9378 [15:02:10<57:43:42, 27.92s/it]
+
+
+ 21%|███████████████████▍ | 1935/9378 [15:02:38<57:40:03, 27.89s/it]
+
+
+ 21%|███████████████████▍ | 1936/9378 [15:03:05<57:33:30, 27.84s/it]
+
+
+ 21%|███████████████████▍ | 1937/9378 [15:03:33<57:42:15, 27.92s/it]
+
+
+
+ 21%|███████████████████▍ | 1938/9378 [15:04:03<57:49:06, 27.98s/it]
+
+
+ 21%|███████████████████▍ | 1939/9378 [15:04:31<57:59:01, 28.06s/it]
+
+
+ 21%|███████████████████▍ | 1940/9378 [15:04:59<57:45:11, 27.95s/it]
+
+
+ 21%|███████████████████▍ | 1941/9378 [15:05:27<57:43:26, 27.94s/it]
+
+
+ 21%|███████████████████▍ | 1942/9378 [15:05:55<57:41:31, 27.93s/it]
+
+
+ 21%|███████████████████▍ | 1943/9378 [15:06:23<57:43:46, 27.95s/it]
+
+
+ 21%|███████████████████▍ | 1944/9378 [15:06:51<57:33:21, 27.87s/it]
+
+ 21%|███████████████████▍ | 1945/9378 [15:07:19<57:34:52, 27.89s/it]
+
+
+ 21%|███████████████████▌ | 1946/9378 [15:07:47<57:42:29, 27.95s/it]
+
+ 21%|███████████████████▌ | 1947/9378 [15:08:13<57:50:25, 28.02s/it]
+
+
+ 21%|███████████████████▌ | 1948/9378 [15:08:42<58:07:38, 28.16s/it]
+
+ 21%|███████████████████▌ | 1948/9378 [15:08:43<58:07:38, 28.16s/it]
+
+
+ 21%|███████████████████▌ | 1949/9378 [15:09:11<57:42:39, 27.97s/it]
+
+
+ 21%|███████████████████▌ | 1951/9378 [15:10:27<66:56:13, 32.45s/it]
+
+
+
+ 21%|███████████████████▌ | 1952/9378 [15:10:56<64:00:53, 31.03s/it]
+
+
+ 21%|███████████████████▌ | 1953/9378 [15:11:24<62:01:56, 30.08s/it]
+
+
+ 21%|███████████████████▌ | 1954/9378 [15:11:52<60:46:41, 29.47s/it]
+
+
+ 21%|███████████████████▌ | 1955/9378 [15:12:20<59:43:52, 28.97s/it]
+
+
+ 21%|███████████████████▌ | 1956/9378 [15:12:48<59:03:24, 28.65s/it]
+
+ 21%|███████████████████▌ | 1957/9378 [15:13:15<59:01:04, 28.63s/it]
+
+ 21%|███████████████████▋ | 1958/9378 [15:13:42<58:10:58, 28.23s/it]
+
+ 21%|███████████████████▋ | 1958/9378 [15:13:43<58:10:58, 28.23s/it]
+
+
+ 21%|███████████████████▋ | 1959/9378 [15:14:11<57:53:14, 28.09s/it]
+
+ 21%|███████████████████▋ | 1960/9378 [15:14:39<57:40:40, 27.99s/it]
+
+ 21%|███████████████████▋ | 1961/9378 [15:15:07<58:04:18, 28.19s/it]
+
+ 21%|███████████████████▋ | 1962/9378 [15:15:35<57:54:54, 28.11s/it]
+
+ 21%|███████████████████▋ | 1963/9378 [15:16:03<57:45:39, 28.04s/it]
+
+
+ 21%|███████████████████▋ | 1965/9378 [15:16:57<57:11:43, 27.78s/it]
+
+
+ 21%|███████████████████▋ | 1966/9378 [15:17:25<57:19:09, 27.84s/it]
+
+
+ 21%|███████████████████▋ | 1967/9378 [15:17:53<57:12:41, 27.79s/it]
+
+
+ 21%|███████████████████▋ | 1968/9378 [15:18:21<57:15:04, 27.81s/it]
+
+
+ 21%|███████████████████▋ | 1969/9378 [15:18:48<57:13:48, 27.81s/it]
+
+
+ 21%|███████████████████▋ | 1970/9378 [15:19:16<57:21:55, 27.88s/it]
+
+ 21%|███████████████████▋ | 1970/9378 [15:19:18<57:21:55, 27.88s/it]
+[2024-05-13 01:51:44,914] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 00917022-cd58-4dd0-8f07-0f07b356048a)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/3bc96ab500694c84fd4e6598dc105ba82aea8a141bc92c688a703ac239f88f27?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T011630Z&X-Amz-Expires=86400&X-Amz-Signature=9808dcdf27ea1180569c6992fc46196a630f45b030a52d5e82b8f84d65bbc56f&X-Amz-SignedHeaders=host&partNumber=1&uploadId=XTUj6q15M25OQqAYZA5tRJx6ySofFY3VzqVub0EY7OnlRpnp3qMvtoKNQx2kNYm9C75NUDPoUFfGA1NPT3KJNj1jl1Cm0PvODnAmY7JwPCVSXxagRvsHhuFhn2pzzefo&x-id=UploadPart
+ 21%|███████████████████▋ | 1970/9378 [15:19:18<57:21:55, 27.88s/it]'(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 00917022-cd58-4dd0-8f07-0f07b356048a)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/3bc96ab500694c84fd4e6598dc105ba82aea8a141bc92c688a703ac239f88f27?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T011630Z&X-Amz-Expires=86400&X-Amz-Signature=9808dcdf27ea1180569c6992fc46196a630f45b030a52d5e82b8f84d65bbc56f&X-Amz-SignedHeaders=host&partNumber=1&uploadId=XTUj6q15M25OQqAYZA5tRJx6ySofFY3VzqVub0EY7OnlRpnp3qMvtoKNQx2kNYm9C75NUDPoUFfGA1NPT3KJNj1jl1Cm0PvODnAmY7JwPCVSXxagRvsHhuFhn2pzzefo&x-id=UploadPart
+Retrying in 2s [Retry 2/5].
+ 21%|███████████████████▊ | 1971/9378 [15:19:44<57:20:22, 27.87s/it]
+
+
+ 21%|███████████████████▊ | 1972/9378 [15:20:13<57:58:26, 28.18s/it]
+
+
+ 21%|███████████████████▊ | 1973/9378 [15:20:41<57:27:19, 27.93s/it]
+
+
+ 21%|███████████████████▊ | 1974/9378 [15:21:09<57:36:14, 28.01s/it]
+
+
+ 21%|███████████████████▊ | 1975/9378 [15:21:37<57:28:15, 27.95s/it]
+
+
+ 21%|███████████████████▊ | 1976/9378 [15:22:05<57:43:27, 28.07s/it]
+
+
+
+ 21%|███████████████████▊ | 1977/9378 [15:22:35<58:20:02, 28.37s/it]
+
+
+ 21%|███████████████████▊ | 1978/9378 [15:23:03<57:38:38, 28.04s/it]
+
+
+ 21%|███████████████████▊ | 1979/9378 [15:23:31<57:38:20, 28.04s/it]
+
+
+ 21%|███████████████████▊ | 1980/9378 [15:23:59<57:38:06, 28.05s/it]
+
+
+ 21%|███████████████████▊ | 1981/9378 [15:24:27<57:48:54, 28.14s/it]
+
+ 21%|███████████████████▊ | 1982/9378 [15:24:55<58:05:41, 28.28s/it]
+
+
+ 21%|███████████████████▉ | 1983/9378 [15:25:23<57:24:51, 27.95s/it]
+
+ 21%|███████████████████▉ | 1984/9378 [15:25:49<57:25:28, 27.96s/it]
+
+
+
+ 21%|███████████████████▉ | 1985/9378 [15:26:19<57:25:16, 27.96s/it]
+
+
+ 21%|███████████████████▉ | 1986/9378 [15:26:47<57:22:27, 27.94s/it]
+
+ 21%|███████████████████▉ | 1987/9378 [15:27:13<57:16:30, 27.90s/it]
+
+
+ 21%|███████████████████▉ | 1988/9378 [15:27:42<57:42:38, 28.11s/it]
+
+
+ 21%|███████████████████▉ | 1989/9378 [15:28:09<57:12:10, 27.87s/it]
+
+
+ 21%|███████████████████▉ | 1990/9378 [15:28:38<57:11:53, 27.87s/it]
+{'loss': 1.1091, 'grad_norm': 0.51171875, 'learning_rate': 1.8721381299765112e-06, 'epoch': 0.64}
+
+ 21%|███████████████████▉ | 1991/9378 [15:29:06<57:32:37, 28.04s/it]
+
+
+ 21%|███████████████████▉ | 1993/9378 [15:30:02<57:41:51, 28.13s/it]
+
+
+ 21%|███████████████████▉ | 1994/9378 [15:30:29<57:22:25, 27.97s/it]
+
+
+ 21%|███████████████████▉ | 1995/9378 [15:30:58<57:37:32, 28.10s/it]
+
+
+ 21%|████████████████████ | 1996/9378 [15:31:26<57:48:56, 28.20s/it]
+
+
+ 21%|████████████████████ | 1997/9378 [15:31:55<57:57:13, 28.27s/it]
+
+ 21%|████████████████████ | 1997/9378 [15:31:56<57:57:13, 28.27s/it]
+
+
+ 21%|████████████████████ | 1999/9378 [15:32:52<58:07:43, 28.36s/it]
+
+
+ 21%|████████████████████ | 2000/9378 [15:33:41<71:01:42, 34.66s/it]
+
+
+ 21%|████████████████████ | 2001/9378 [15:34:09<66:53:14, 32.64s/it]
+
+
+ 21%|████████████████████ | 2002/9378 [15:34:37<64:09:03, 31.31s/it]
+
+
+ 21%|████████████████████ | 2003/9378 [15:35:06<62:22:58, 30.45s/it]
+
+
+ 21%|████████████████████ | 2004/9378 [15:35:34<61:01:36, 29.79s/it]
+
+
+ 21%|████████████████████ | 2005/9378 [15:36:02<60:10:55, 29.38s/it]
+
+
+ 21%|████████████████████ | 2006/9378 [15:36:31<59:30:31, 29.06s/it]
+
+
+ 21%|████████████████████ | 2007/9378 [15:36:59<59:19:36, 28.98s/it]
+
+
+ 21%|████████████████████▏ | 2008/9378 [15:37:28<59:07:50, 28.88s/it]
+
+
+ 21%|████████████████████▏ | 2009/9378 [15:37:57<58:59:26, 28.82s/it]
+
+
+ 21%|████████████████████▏ | 2010/9378 [15:38:25<58:28:30, 28.57s/it]
+
+
+ 21%|████████████████████▏ | 2011/9378 [15:38:53<58:28:46, 28.58s/it]
+
+
+ 21%|████████████████████▏ | 2012/9378 [15:39:22<58:35:20, 28.63s/it]
+
+
+ 21%|████████████████████▏ | 2013/9378 [15:39:50<58:08:50, 28.42s/it]
+
+
+ 21%|████████████████████▏ | 2014/9378 [15:40:20<58:04:08, 28.39s/it]
+
+ 21%|████████████████████▏ | 2015/9378 [15:40:47<58:12:35, 28.46s/it]
+
+
+ 21%|████████████████████▏ | 2016/9378 [15:41:15<58:08:39, 28.43s/it]
+
+
+ 22%|████████████████████▏ | 2017/9378 [15:41:44<58:11:21, 28.46s/it]
+
+
+ 22%|████████████████████▏ | 2018/9378 [15:42:12<58:13:47, 28.48s/it]
+
+
+ 22%|████████████████████▏ | 2019/9378 [15:42:41<58:33:38, 28.65s/it]
+
+
+ 22%|████████████████████▏ | 2020/9378 [15:43:10<58:20:10, 28.54s/it]
+
+
+ 22%|████████████████████▎ | 2021/9378 [15:43:37<57:51:24, 28.31s/it]
+
+
+ 22%|████████████████████▎ | 2022/9378 [15:44:06<58:00:37, 28.39s/it]
+
+
+ 22%|████████████████████▎ | 2023/9378 [15:44:34<58:03:06, 28.41s/it]
+
+
+ 22%|████████████████████▎ | 2024/9378 [15:45:03<58:18:16, 28.54s/it]
+
+
+ 22%|████████████████████▎ | 2025/9378 [15:45:31<57:46:37, 28.29s/it]
+
+
+ 22%|████████████████████▎ | 2026/9378 [15:45:59<57:45:57, 28.29s/it]
+
+
+ 22%|████████████████████▎ | 2027/9378 [15:46:28<57:44:14, 28.28s/it]
+
+
+ 22%|████████████████████▎ | 2028/9378 [15:46:56<58:02:02, 28.42s/it]
+
+
+ 22%|████████████████████▎ | 2029/9378 [15:47:25<57:58:58, 28.40s/it]
+
+
+ 22%|████████████████████▎ | 2030/9378 [15:47:55<57:56:07, 28.38s/it]
+
+ 22%|████████████████████▎ | 2031/9378 [15:48:22<58:08:03, 28.49s/it]
+
+
+ 22%|████████████████████▎ | 2032/9378 [15:48:50<58:02:11, 28.44s/it]
+
+
+ 22%|████████████████████▍ | 2033/9378 [15:49:18<58:01:25, 28.44s/it]
+
+
+ 22%|████████████████████▍ | 2034/9378 [15:49:49<58:03:04, 28.46s/it]
+
+ 22%|████████████████████▍ | 2035/9378 [15:50:15<58:02:06, 28.45s/it]
+
+
+ 22%|████████████████████▍ | 2036/9378 [15:50:44<57:58:26, 28.43s/it]
+
+
+ 22%|████████████████████▍ | 2037/9378 [15:51:12<57:49:28, 28.36s/it]
+
+
+ 22%|████████████████████▍ | 2038/9378 [15:51:40<57:48:35, 28.35s/it]
+
+ 22%|████████████████████▍ | 2038/9378 [15:51:42<57:48:35, 28.35s/it]
+
+
+ 22%|████████████████████▍ | 2040/9378 [15:52:38<58:08:12, 28.52s/it]
+
+
+ 22%|████████████████████▍ | 2041/9378 [15:53:07<58:26:21, 28.67s/it]
+
+
+ 22%|████████████████████▍ | 2042/9378 [15:53:34<57:41:35, 28.31s/it]
+
+
+ 22%|████████████████████▍ | 2043/9378 [15:54:03<57:47:38, 28.37s/it]
+
+
+ 22%|████████████████████▍ | 2044/9378 [15:54:31<57:45:29, 28.35s/it]
+
+
+ 22%|████████████████████▍ | 2045/9378 [15:55:01<57:44:09, 28.34s/it]
+
+ 22%|████████████████████▌ | 2046/9378 [15:55:28<57:42:00, 28.33s/it]
+
+
+ 22%|████████████████████▌ | 2047/9378 [15:55:57<58:10:28, 28.57s/it]
+
+
+ 22%|████████████████████▌ | 2048/9378 [15:56:24<57:27:27, 28.22s/it]
+
+
+ 22%|████████████████████▌ | 2049/9378 [15:56:53<57:34:44, 28.28s/it]
+
+
+ 22%|████████████████████▌ | 2050/9378 [15:57:43<70:46:47, 34.77s/it]
+
+
+ 22%|████████████████████▌ | 2051/9378 [15:58:11<66:46:59, 32.81s/it]
+
+
+ 22%|████████████████████▌ | 2052/9378 [15:58:40<64:19:08, 31.61s/it]
+
+
+ 22%|████████████████████▌ | 2053/9378 [15:59:10<62:42:21, 30.82s/it]
+
+ 22%|████████████████████▌ | 2054/9378 [15:59:37<61:13:23, 30.09s/it]
+
+
+ 22%|████████████████████▌ | 2055/9378 [16:00:05<59:41:04, 29.34s/it]
+
+
+ 22%|████████████████████▌ | 2056/9378 [16:00:33<59:24:42, 29.21s/it]
+
+
+ 22%|████████████████████▌ | 2057/9378 [16:01:02<58:52:19, 28.95s/it]
+
+ 22%|████████████████████▌ | 2057/9378 [16:01:03<58:52:19, 28.95s/it]
+
+ 22%|████████████████████▋ | 2058/9378 [16:01:31<58:27:08, 28.75s/it]
+
+
+ 22%|████████████████████▋ | 2060/9378 [16:02:28<57:38:53, 28.36s/it]
+
+ 22%|████████████████████▋ | 2061/9378 [16:02:54<57:40:26, 28.38s/it]
+
+
+ 22%|████████████████████▋ | 2062/9378 [16:03:23<57:35:08, 28.34s/it]
+
+
+ 22%|████████████████████▋ | 2063/9378 [16:03:51<57:34:44, 28.34s/it]
+
+
+ 22%|████████████████████▋ | 2064/9378 [16:04:19<57:33:59, 28.33s/it]
+
+ 22%|████████████████████▋ | 2064/9378 [16:04:21<57:33:59, 28.33s/it]
+
+ 22%|████████████████████▋ | 2065/9378 [16:04:50<58:00:39, 28.56s/it]
+
+
+ 22%|████████████████████▋ | 2067/9378 [16:05:46<57:24:25, 28.27s/it]
+
+ 22%|████████████████████▋ | 2068/9378 [16:06:14<57:54:09, 28.52s/it]
+
+
+ 22%|████████████████████▋ | 2069/9378 [16:06:41<57:24:54, 28.28s/it]
+
+
+ 22%|████████████████████▋ | 2070/9378 [16:07:10<57:25:44, 28.29s/it]
+
+
+ 22%|████████████████████▊ | 2071/9378 [16:07:38<57:28:10, 28.31s/it]
+
+
+ 22%|████████████████████▊ | 2072/9378 [16:08:06<57:29:26, 28.33s/it]
+
+
+ 22%|████████████████████▊ | 2073/9378 [16:08:35<57:28:19, 28.32s/it]
+
+
+ 22%|████████████████████▊ | 2074/9378 [16:09:03<57:30:26, 28.34s/it]
+
+
+ 22%|████████████████████▊ | 2075/9378 [16:09:32<57:56:35, 28.56s/it]
+
+
+ 22%|████████████████████▊ | 2076/9378 [16:10:00<57:26:52, 28.32s/it]
+
+
+ 22%|████████████████████▊ | 2077/9378 [16:10:28<57:32:21, 28.37s/it]
+
+
+
+ 22%|████████████████████▊ | 2078/9378 [16:10:59<57:28:44, 28.35s/it]
+
+ 22%|████████████████████▊ | 2079/9378 [16:11:25<57:31:44, 28.37s/it]
+
+
+ 22%|████████████████████▊ | 2080/9378 [16:11:54<57:35:20, 28.41s/it]
+
+
+ 22%|████████████████████▊ | 2081/9378 [16:12:22<57:28:02, 28.35s/it]
+
+
+ 22%|████████████████████▊ | 2082/9378 [16:12:50<57:22:18, 28.31s/it]
+
+
+ 22%|████████████████████▉ | 2083/9378 [16:13:18<57:24:33, 28.33s/it]
+
+ 22%|████████████████████▉ | 2083/9378 [16:13:20<57:24:33, 28.33s/it]
+
+
+
+ 22%|████████████████████▉ | 2085/9378 [16:14:17<57:13:33, 28.25s/it]
+
+ 22%|████████████████████▉ | 2086/9378 [16:14:44<57:43:03, 28.49s/it]
+
+
+ 22%|████████████████████▉ | 2087/9378 [16:15:13<57:41:46, 28.49s/it]
+
+
+ 22%|████████████████████▉ | 2088/9378 [16:15:41<57:19:49, 28.31s/it]
+
+
+ 22%|████████████████████▉ | 2089/9378 [16:16:10<57:48:50, 28.55s/it]
+
+
+ 22%|████████████████████▉ | 2090/9378 [16:16:37<57:07:46, 28.22s/it]
+
+ 22%|████████████████████▉ | 2090/9378 [16:16:39<57:07:46, 28.22s/it]
+[2024-05-13 02:49:32,425] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-2090/pytorch_model_fsdp.bin
+[2024-05-13 02:49:44,091] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-2090/pytorch_model_fsdp.bin
+[2024-05-13 02:50:09,071] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-2090/optimizer.bin
+ 22%|████████████████████▉ | 2090/9378 [16:16:39<57:07:46, 28.22s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+
+ 22%|████████████████████▋ | 2091/9378 [16:19:15<134:57:19, 66.67s/it]
+
+ 22%|████████████████████▋ | 2092/9378 [16:19:42<111:55:12, 55.30s/it]
+
+
+ 22%|████████████████████▉ | 2093/9378 [16:20:11<95:47:10, 47.33s/it]
+
+
+ 22%|████████████████████▉ | 2094/9378 [16:20:39<84:03:33, 41.54s/it]
+
+
+ 22%|████████████████████▉ | 2095/9378 [16:21:08<76:08:33, 37.64s/it]
+
+
+ 22%|█████████████████████ | 2096/9378 [16:21:37<70:27:49, 34.84s/it]
+
+ 22%|█████████████████████ | 2097/9378 [16:22:04<66:31:55, 32.90s/it]
+
+
+ 22%|█████████████████████ | 2098/9378 [16:22:33<63:50:01, 31.57s/it]
+
+
+ 22%|█████████████████████ | 2099/9378 [16:23:02<62:11:44, 30.76s/it]
+
+
+ 22%|█████████████████████ | 2100/9378 [16:23:54<74:06:08, 36.65s/it]
+
+ 22%|█████████████████████ | 2101/9378 [16:24:20<69:01:53, 34.15s/it]
+
+
+ 22%|█████████████████████ | 2102/9378 [16:24:49<65:58:35, 32.64s/it]
+
+
+ 22%|█████████████████████ | 2103/9378 [16:25:17<62:52:02, 31.11s/it]
+
+
+ 22%|█████████████████████ | 2104/9378 [16:25:45<61:10:48, 30.28s/it]
+
+ 22%|█████████████████████ | 2104/9378 [16:25:47<61:10:48, 30.28s/it]
+
+ 22%|█████████████████████ | 2105/9378 [16:26:15<60:24:45, 29.90s/it]
+
+
+ 22%|█████████████████████ | 2107/9378 [16:27:12<58:43:01, 29.07s/it]
+
+ 22%|█████████████████████▏ | 2108/9378 [16:27:40<58:30:44, 28.97s/it]
+
+
+ 22%|█████████████████████▏ | 2109/9378 [16:28:08<58:07:58, 28.79s/it]
+
+
+ 22%|█████████████████████▏ | 2110/9378 [16:28:36<57:27:53, 28.46s/it]
+
+
+ 23%|█████████████████████▏ | 2111/9378 [16:29:04<57:26:58, 28.46s/it]
+
+
+ 23%|█████████████████████▏ | 2112/9378 [16:29:32<57:15:30, 28.37s/it]
+
+ 23%|█████████████████████▏ | 2112/9378 [16:29:34<57:15:30, 28.37s/it]
+
+
+
+ 23%|█████████████████████▏ | 2114/9378 [16:30:29<57:09:34, 28.33s/it]
+
+
+ 23%|█████████████████████▏ | 2115/9378 [16:30:58<57:39:03, 28.58s/it]
+
+
+ 23%|█████████████████████▏ | 2116/9378 [16:31:26<57:02:37, 28.28s/it]
+
+
+ 23%|█████████████████████▏ | 2117/9378 [16:31:54<57:06:22, 28.31s/it]
+
+ 23%|█████████████████████▏ | 2117/9378 [16:31:56<57:06:22, 28.31s/it]
+
+
+ 23%|█████████████████████▏ | 2119/9378 [16:32:52<57:32:06, 28.53s/it]
+
+
+ 23%|█████████████████████▏ | 2120/9378 [16:33:20<57:07:42, 28.34s/it]
+
+
+ 23%|█████████████████████▎ | 2121/9378 [16:33:49<57:28:40, 28.51s/it]
+
+
+ 23%|█████████████████████▎ | 2122/9378 [16:34:16<56:55:12, 28.24s/it]
+
+ 23%|█████████████████████▎ | 2122/9378 [16:34:18<56:55:12, 28.24s/it]
+
+
+ 23%|█████████████████████▎ | 2124/9378 [16:35:13<56:42:10, 28.14s/it]
+
+ 23%|█████████████████████▎ | 2124/9378 [16:35:14<56:42:10, 28.14s/it]
+
+
+ 23%|█████████████████████▎ | 2126/9378 [16:36:09<56:59:26, 28.29s/it]
+
+
+ 23%|█████████████████████▎ | 2127/9378 [16:36:38<57:21:11, 28.47s/it]
+
+
+ 23%|█████████████████████▎ | 2128/9378 [16:37:06<57:00:04, 28.30s/it]
+
+
+ 23%|█████████████████████▎ | 2129/9378 [16:37:35<57:23:40, 28.50s/it]
+
+
+ 23%|█████████████████████▎ | 2130/9378 [16:38:06<57:52:37, 28.75s/it]
+
+ 23%|█████████████████████▎ | 2131/9378 [16:38:33<57:40:18, 28.65s/it]
+
+
+ 23%|█████████████████████▎ | 2132/9378 [16:39:01<57:27:06, 28.54s/it]
+
+
+ 23%|█████████████████████▍ | 2133/9378 [16:39:29<56:47:30, 28.22s/it]
+
+
+ 23%|█████████████████████▍ | 2134/9378 [16:39:57<56:49:34, 28.24s/it]
+
+
+ 23%|█████████████████████▍ | 2135/9378 [16:40:25<56:51:01, 28.26s/it]
+
+
+
+ 23%|█████████████████████▍ | 2136/9378 [16:40:56<56:56:21, 28.30s/it]
+
+ 23%|█████████████████████▍ | 2137/9378 [16:41:24<57:03:55, 28.37s/it]
+
+ 23%|█████████████████████▍ | 2138/9378 [16:41:51<57:05:20, 28.39s/it]
+
+
+ 23%|█████████████████████▍ | 2139/9378 [16:42:19<57:10:01, 28.43s/it]
+
+
+ 23%|█████████████████████▍ | 2140/9378 [16:42:48<57:08:27, 28.42s/it]
+
+
+
+ 23%|█████████████████████▍ | 2141/9378 [16:43:18<57:01:03, 28.36s/it]
+
+
+ 23%|█████████████████████▍ | 2142/9378 [16:43:46<56:59:35, 28.35s/it]
+
+ 23%|█████████████████████▍ | 2143/9378 [16:44:13<57:03:33, 28.39s/it]
+
+
+ 23%|█████████████████████▍ | 2144/9378 [16:44:41<57:07:42, 28.43s/it]
+
+
+ 23%|█████████████████████▌ | 2145/9378 [16:45:10<57:34:37, 28.66s/it]
+
+
+ 23%|█████████████████████▌ | 2146/9378 [16:45:40<57:29:24, 28.62s/it]
+
+ 23%|█████████████████████▌ | 2147/9378 [16:46:08<56:52:34, 28.32s/it]
+
+ 23%|█████████████████████▌ | 2148/9378 [16:46:35<56:48:45, 28.29s/it]
+
+
+ 23%|█████████████████████▌ | 2149/9378 [16:47:04<57:07:34, 28.45s/it]
+
+
+ 23%|█████████████████████▌ | 2150/9378 [16:47:54<70:04:50, 34.90s/it]
+
+
+ 23%|█████████████████████▌ | 2151/9378 [16:48:22<65:57:30, 32.86s/it]
+
+
+ 23%|█████████████████████▌ | 2152/9378 [16:48:50<63:12:59, 31.49s/it]
+
+
+
+ 23%|█████████████████████▌ | 2153/9378 [16:49:20<61:16:37, 30.53s/it]
+
+
+ 23%|█████████████████████▌ | 2154/9378 [16:49:48<59:57:54, 29.88s/it]
+
+ 23%|█████████████████████▌ | 2155/9378 [16:50:17<59:01:40, 29.42s/it]
+
+ 23%|█████████████████████▌ | 2156/9378 [16:50:43<58:24:51, 29.12s/it]
+
+
+ 23%|█████████████████████▌ | 2157/9378 [16:51:12<57:58:13, 28.90s/it]
+
+
+ 23%|█████████████████████▋ | 2158/9378 [16:51:40<57:31:55, 28.69s/it]
+
+
+ 23%|█████████████████████▋ | 2159/9378 [16:52:08<57:20:21, 28.59s/it]
+
+
+ 23%|█████████████████████▋ | 2160/9378 [16:52:36<57:05:19, 28.47s/it]
+
+
+
+ 23%|█████████████████████▋ | 2161/9378 [16:53:07<57:01:34, 28.45s/it]
+
+
+ 23%|█████████████████████▋ | 2162/9378 [16:53:35<56:51:57, 28.37s/it]
+
+ 23%|█████████████████████▋ | 2163/9378 [16:54:02<57:01:00, 28.45s/it]
+
+
+ 23%|█████████████████████▋ | 2164/9378 [16:54:30<56:54:50, 28.40s/it]
+
+
+ 23%|█████████████████████▋ | 2165/9378 [16:54:59<57:18:26, 28.60s/it]
+
+
+ 23%|█████████████████████▋ | 2166/9378 [16:55:27<56:42:33, 28.31s/it]
+
+
+
+ 23%|█████████████████████▋ | 2167/9378 [16:55:57<56:38:22, 28.28s/it]
+
+ 23%|█████████████████████▋ | 2168/9378 [16:56:25<57:09:01, 28.54s/it]
+
+ 23%|█████████████████████▋ | 2169/9378 [16:56:53<56:31:42, 28.23s/it]
+
+ 23%|█████████████████████▊ | 2170/9378 [16:57:20<56:32:00, 28.24s/it]
+
+
+ 23%|█████████████████████▊ | 2171/9378 [16:57:48<56:36:27, 28.28s/it]
+
+
+ 23%|█████████████████████▊ | 2172/9378 [16:58:17<56:44:18, 28.35s/it]
+
+
+ 23%|█████████████████████▊ | 2173/9378 [16:58:47<57:13:44, 28.59s/it]
+
+
+ 23%|█████████████████████▊ | 2174/9378 [16:59:15<56:34:21, 28.27s/it]
+
+ 23%|█████████████████████▊ | 2175/9378 [16:59:44<56:53:44, 28.44s/it]
+
+ 23%|█████████████████████▊ | 2176/9378 [17:00:11<56:50:41, 28.41s/it]
+
+
+ 23%|█████████████████████▊ | 2177/9378 [17:00:39<56:58:09, 28.48s/it]
+
+
+ 23%|█████████████████████▊ | 2178/9378 [17:01:07<56:48:40, 28.41s/it]
+
+
+ 23%|█████████████████████▊ | 2179/9378 [17:01:35<56:30:00, 28.25s/it]
+
+
+ 23%|█████████████████████▊ | 2180/9378 [17:02:03<56:14:17, 28.13s/it]
+
+
+
+ 23%|█████████████████████▊ | 2181/9378 [17:02:33<56:29:46, 28.26s/it]
+
+
+ 23%|█████████████████████▊ | 2182/9378 [17:03:02<56:24:39, 28.22s/it]
+
+ 23%|█████████████████████▉ | 2183/9378 [17:03:30<56:58:51, 28.51s/it]
+
+ 23%|█████████████████████▉ | 2184/9378 [17:03:58<56:23:40, 28.22s/it]
+
+ 23%|█████████████████████▉ | 2185/9378 [17:04:25<56:26:51, 28.25s/it]
+
+
+ 23%|█████████████████████▉ | 2186/9378 [17:04:53<56:26:13, 28.25s/it]
+
+
+ 23%|█████████████████████▉ | 2187/9378 [17:05:22<56:53:52, 28.48s/it]
+
+
+
+ 23%|█████████████████████▉ | 2188/9378 [17:05:51<56:28:44, 28.28s/it]
+
+
+ 23%|█████████████████████▉ | 2189/9378 [17:06:20<56:31:22, 28.30s/it]
+
+
+ 23%|█████████████████████▉ | 2190/9378 [17:06:48<56:26:08, 28.26s/it]
+
+
+ 23%|█████████████████████▉ | 2191/9378 [17:07:16<56:20:09, 28.22s/it]
+
+ 23%|█████████████████████▉ | 2192/9378 [17:07:43<56:33:09, 28.33s/it]
+
+
+ 23%|█████████████████████▉ | 2193/9378 [17:08:12<56:33:09, 28.34s/it]
+
+
+
+ 23%|█████████████████████▉ | 2194/9378 [17:08:42<56:27:49, 28.29s/it]
+
+
+ 23%|██████████████████████ | 2195/9378 [17:09:10<56:33:57, 28.35s/it]
+
+
+ 23%|██████████████████████ | 2196/9378 [17:09:38<56:37:20, 28.38s/it]
+
+
+ 23%|██████████████████████ | 2197/9378 [17:10:07<56:30:55, 28.33s/it]
+
+ 23%|██████████████████████ | 2198/9378 [17:10:34<57:04:40, 28.62s/it]
+
+
+ 23%|██████████████████████ | 2199/9378 [17:11:02<56:52:26, 28.52s/it]
+
+
+ 23%|██████████████████████ | 2200/9378 [17:11:52<69:32:01, 34.87s/it]
+
+
+ 23%|██████████████████████ | 2201/9378 [17:12:20<65:23:36, 32.80s/it]
+
+
+
+ 23%|██████████████████████ | 2202/9378 [17:12:50<62:39:12, 31.43s/it]
+
+
+ 23%|██████████████████████ | 2203/9378 [17:13:19<60:48:12, 30.51s/it]
+
+ 24%|██████████████████████ | 2204/9378 [17:13:47<59:55:50, 30.07s/it]
+
+ 24%|██████████████████████ | 2205/9378 [17:14:15<58:26:24, 29.33s/it]
+
+ 24%|██████████████████████ | 2206/9378 [17:14:42<57:51:56, 29.05s/it]
+
+
+ 24%|██████████████████████ | 2207/9378 [17:15:11<57:49:31, 29.03s/it]
+
+
+ 24%|██████████████████████▏ | 2208/9378 [17:15:39<57:24:07, 28.82s/it]
+
+
+
+ 24%|██████████████████████▏ | 2209/9378 [17:16:08<57:00:25, 28.63s/it]
+
+ 24%|██████████████████████▏ | 2210/9378 [17:16:37<56:55:24, 28.59s/it]
+
+ 24%|██████████████████████▏ | 2211/9378 [17:17:05<56:43:54, 28.50s/it]
+
+ 24%|██████████████████████▏ | 2212/9378 [17:17:33<56:52:08, 28.57s/it]
+
+
+ 24%|██████████████████████▏ | 2213/9378 [17:18:01<56:20:30, 28.31s/it]
+
+
+
+ 24%|██████████████████████▏ | 2214/9378 [17:18:31<56:18:37, 28.30s/it]
+
+
+ 24%|██████████████████████▏ | 2215/9378 [17:18:59<56:18:52, 28.30s/it]
+
+ 24%|██████████████████████▏ | 2216/9378 [17:19:27<56:47:36, 28.55s/it]
+
+ 24%|██████████████████████▏ | 2217/9378 [17:19:56<56:44:17, 28.52s/it]
+
+ 24%|██████████████████████▏ | 2218/9378 [17:20:22<56:06:38, 28.21s/it]
+
+
+ 24%|██████████████████████▏ | 2219/9378 [17:20:51<56:16:46, 28.30s/it]
+
+
+ 24%|██████████████████████▎ | 2220/9378 [17:21:19<56:33:48, 28.45s/it]
+
+
+
+ 24%|██████████████████████▎ | 2221/9378 [17:21:49<56:08:08, 28.24s/it]
+
+
+ 24%|██████████████████████▎ | 2222/9378 [17:22:18<56:18:40, 28.33s/it]
+
+
+ 24%|██████████████████████▎ | 2223/9378 [17:22:46<56:16:20, 28.31s/it]
+
+ 24%|██████████████████████▎ | 2224/9378 [17:23:12<56:21:30, 28.36s/it]
+
+
+ 24%|██████████████████████▎ | 2225/9378 [17:23:41<56:26:05, 28.40s/it]
+
+
+
+ 24%|██████████████████████▎ | 2226/9378 [17:24:11<56:25:34, 28.40s/it]
+
+
+ 24%|██████████████████████▎ | 2227/9378 [17:24:39<56:19:58, 28.36s/it]
+
+
+ 24%|██████████████████████▎ | 2228/9378 [17:25:08<56:13:09, 28.31s/it]
+
+
+ 24%|██████████████████████▎ | 2229/9378 [17:25:36<56:14:05, 28.32s/it]
+
+
+ 24%|██████████████████████▎ | 2230/9378 [17:26:04<56:14:32, 28.33s/it]
+
+
+ 24%|██████████████████████▎ | 2231/9378 [17:26:32<56:06:05, 28.26s/it]
+
+
+ 24%|██████████████████████▎ | 2232/9378 [17:27:01<56:08:39, 28.28s/it]
+
+ 24%|██████████████████████▍ | 2233/9378 [17:27:27<56:10:57, 28.31s/it]
+
+
+ 24%|██████████████████████▍ | 2234/9378 [17:27:56<56:26:19, 28.44s/it]
+
+
+
+ 24%|██████████████████████▍ | 2235/9378 [17:28:26<56:39:14, 28.55s/it]
+
+
+ 24%|██████████████████████▍ | 2236/9378 [17:28:54<56:01:36, 28.24s/it]
+
+
+ 24%|██████████████████████▍ | 2237/9378 [17:29:23<56:07:06, 28.29s/it]
+
+ 24%|██████████████████████▍ | 2238/9378 [17:29:51<56:38:57, 28.56s/it]
+
+ 24%|██████████████████████▍ | 2239/9378 [17:30:19<56:16:16, 28.38s/it]
+
+ 24%|██████████████████████▍ | 2240/9378 [17:30:46<56:00:18, 28.25s/it]
+
+
+
+ 24%|██████████████████████▍ | 2241/9378 [17:31:16<56:16:36, 28.39s/it]
+
+
+ 24%|██████████████████████▍ | 2242/9378 [17:31:44<56:01:13, 28.26s/it]
+
+ 24%|██████████████████████▍ | 2243/9378 [17:32:13<56:32:58, 28.53s/it]
+
+
+ 24%|██████████████████████▍ | 2244/9378 [17:32:41<55:59:44, 28.26s/it]
+
+ 24%|██████████████████████▌ | 2245/9378 [17:33:09<56:00:02, 28.26s/it]
+
+ 24%|██████████████████████▌ | 2246/9378 [17:33:36<56:07:52, 28.33s/it]
+
+
+ 24%|██████████████████████▌ | 2247/9378 [17:34:04<56:02:57, 28.30s/it]
+
+
+
+ 24%|██████████████████████▌ | 2248/9378 [17:34:34<55:58:03, 28.26s/it]
+
+
+ 24%|██████████████████████▌ | 2249/9378 [17:35:03<56:06:52, 28.34s/it]
+
+ 24%|██████████████████████▌ | 2250/9378 [17:35:50<68:35:08, 34.64s/it]
+
+
+
+ 24%|██████████████████████▌ | 2251/9378 [17:36:20<64:51:23, 32.76s/it]
+
+
+ 24%|██████████████████████▌ | 2252/9378 [17:36:49<62:11:28, 31.42s/it]
+
+
+ 24%|██████████████████████▌ | 2253/9378 [17:37:17<60:22:49, 30.51s/it]
+
+ 24%|██████████████████████▌ | 2254/9378 [17:37:45<59:32:46, 30.09s/it]
+
+
+ 24%|██████████████████████▌ | 2255/9378 [17:38:14<57:54:59, 29.27s/it]
+
+ 24%|██████████████████████▌ | 2256/9378 [17:38:40<57:26:11, 29.03s/it]
+
+
+ 24%|██████████████████████▌ | 2257/9378 [17:39:09<57:07:33, 28.88s/it]
+
+
+
+ 24%|██████████████████████▋ | 2258/9378 [17:39:39<57:11:28, 28.92s/it]
+
+
+ 24%|██████████████████████▋ | 2259/9378 [17:40:07<56:35:47, 28.62s/it]
+
+ 24%|██████████████████████▋ | 2260/9378 [17:40:36<56:39:53, 28.66s/it]
+
+
+ 24%|██████████████████████▋ | 2261/9378 [17:41:04<56:02:00, 28.34s/it]
+
+ 24%|██████████████████████▋ | 2262/9378 [17:41:31<56:07:11, 28.39s/it]
+
+
+ 24%|██████████████████████▋ | 2263/9378 [17:41:59<55:59:37, 28.33s/it]
+
+
+ 24%|██████████████████████▋ | 2264/9378 [17:42:27<55:54:20, 28.29s/it]
+
+
+
+ 24%|██████████████████████▋ | 2265/9378 [17:42:57<56:25:31, 28.56s/it]
+
+
+ 24%|██████████████████████▋ | 2266/9378 [17:43:26<55:49:51, 28.26s/it]
+
+
+ 24%|██████████████████████▋ | 2267/9378 [17:43:54<55:47:47, 28.25s/it]
+
+ 24%|██████████████████████▋ | 2268/9378 [17:44:22<56:15:54, 28.49s/it]
+
+ 24%|██████████████████████▋ | 2269/9378 [17:44:49<55:52:07, 28.29s/it]
+
+
+ 24%|██████████████████████▊ | 2270/9378 [17:45:17<55:53:29, 28.31s/it]
+
+
+
+ 24%|██████████████████████▊ | 2271/9378 [17:45:47<55:57:13, 28.34s/it]
+
+
+ 24%|██████████████████████▊ | 2272/9378 [17:46:16<55:57:34, 28.35s/it]
+
+
+ 24%|██████████████████████▊ | 2273/9378 [17:46:44<56:01:05, 28.38s/it]
+
+ 24%|██████████████████████▊ | 2274/9378 [17:47:13<56:23:24, 28.58s/it]
+
+ 24%|██████████████████████▊ | 2275/9378 [17:47:41<55:53:30, 28.33s/it]
+
+ 24%|██████████████████████▊ | 2276/9378 [17:48:09<56:16:51, 28.53s/it]
+
+
+ 24%|██████████████████████▊ | 2277/9378 [17:48:38<55:45:45, 28.27s/it]
+
+
+ 24%|██████████████████████▊ | 2278/9378 [17:49:06<55:42:08, 28.24s/it]
+
+
+ 24%|██████████████████████▊ | 2279/9378 [17:49:34<55:47:28, 28.29s/it]
+
+ 24%|██████████████████████▊ | 2280/9378 [17:50:03<56:17:26, 28.55s/it]
+
+
+ 24%|██████████████████████▊ | 2281/9378 [17:50:31<55:39:40, 28.23s/it]
+
+ 24%|██████████████████████▊ | 2282/9378 [17:50:59<55:45:30, 28.29s/it]
+
+
+ 24%|██████████████████████▉ | 2283/9378 [17:51:28<55:48:22, 28.32s/it]
+
+
+ 24%|██████████████████████▉ | 2284/9378 [17:51:56<55:53:58, 28.37s/it]
+
+
+ 24%|██████████████████████▉ | 2285/9378 [17:52:25<55:54:00, 28.37s/it]
+
+ 24%|██████████████████████▉ | 2286/9378 [17:52:53<56:18:54, 28.59s/it]
+
+
+ 24%|██████████████████████▉ | 2287/9378 [17:53:21<55:34:31, 28.21s/it]
+
+ 24%|██████████████████████▉ | 2288/9378 [17:53:49<55:52:13, 28.37s/it]
+
+ 24%|██████████████████████▉ | 2289/9378 [17:54:18<55:36:39, 28.24s/it]
+
+
+ 24%|██████████████████████▉ | 2290/9378 [17:54:46<56:02:31, 28.46s/it]
+
+
+ 24%|██████████████████████▉ | 2291/9378 [17:55:14<55:30:40, 28.20s/it]
+
+
+ 24%|██████████████████████▉ | 2292/9378 [17:55:43<55:34:39, 28.24s/it]
+
+
+ 24%|██████████████████████▉ | 2293/9378 [17:56:11<56:00:18, 28.46s/it]
+
+
+ 24%|██████████████████████▉ | 2294/9378 [17:56:39<55:26:23, 28.17s/it]
+
+
+ 24%|███████████████████████ | 2295/9378 [17:57:07<55:31:57, 28.22s/it]
+
+
+ 24%|███████████████████████ | 2296/9378 [17:57:36<55:29:28, 28.21s/it]
+
+ 24%|███████████████████████ | 2297/9378 [17:58:04<56:00:49, 28.48s/it]
+
+
+ 25%|███████████████████████ | 2298/9378 [17:58:32<55:59:21, 28.47s/it]
+
+
+ 25%|███████████████████████ | 2299/9378 [17:59:01<55:58:43, 28.47s/it]
+{'loss': 1.3085, 'grad_norm': 0.52734375, 'learning_rate': 1.213272986389423e-05, 'epoch': 0.74}
+[2024-05-13 04:31:53,884] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-2299/pytorch_model_fsdp.bin
+[2024-05-13 04:32:05,743] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-2299/pytorch_model_fsdp.bin
+[2024-05-13 04:32:27,230] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-2299/optimizer.bin
+ 25%|███████████████████████ | 2299/9378 [17:59:01<55:58:43, 28.47s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+
+ 25%|██████████████████████▊ | 2300/9378 [18:01:57<142:50:10, 72.65s/it]
+
+
+ 25%|██████████████████████▊ | 2301/9378 [18:02:25<116:21:03, 59.19s/it]
+
+
+ 25%|███████████████████████ | 2302/9378 [18:02:53<98:05:22, 49.90s/it]
+
+
+ 25%|███████████████████████ | 2303/9378 [18:03:22<85:24:22, 43.46s/it]
+
+ 25%|███████████████████████ | 2304/9378 [18:03:50<76:44:24, 39.05s/it]
+
+
+ 25%|███████████████████████ | 2305/9378 [18:04:18<70:05:16, 35.67s/it]
+
+
+ 25%|███████████████████████ | 2306/9378 [18:04:47<65:52:35, 33.53s/it]
+
+
+ 25%|███████████████████████ | 2307/9378 [18:05:15<62:49:59, 31.99s/it]
+
+ 25%|███████████████████████▏ | 2308/9378 [18:05:42<60:47:04, 30.95s/it]
+
+
+ 25%|███████████████████████▏ | 2309/9378 [18:06:11<59:19:42, 30.21s/it]
+
+
+ 25%|███████████████████████▏ | 2310/9378 [18:06:39<58:08:49, 29.62s/it]
+
+
+ 25%|███████████████████████▏ | 2311/9378 [18:07:09<57:25:47, 29.26s/it]
+
+ 25%|███████████████████████▏ | 2312/9378 [18:07:36<57:18:48, 29.20s/it]
+
+
+ 25%|███████████████████████▏ | 2313/9378 [18:08:04<56:35:33, 28.84s/it]
+
+
+ 25%|███████████████████████▏ | 2314/9378 [18:08:32<56:07:17, 28.60s/it]
+
+
+ 25%|███████████████████████▏ | 2315/9378 [18:09:01<55:56:13, 28.51s/it]
+
+ 25%|███████████████████████▏ | 2315/9378 [18:09:02<55:56:13, 28.51s/it]
+
+ 25%|███████████████████████▏ | 2316/9378 [18:09:31<56:20:05, 28.72s/it]
+
+
+ 25%|███████████████████████▏ | 2318/9378 [18:10:26<55:35:22, 28.35s/it]
+
+
+ 25%|███████████████████████▏ | 2319/9378 [18:10:55<55:55:14, 28.52s/it]
+
+ 25%|███████████████████████▎ | 2320/9378 [18:11:22<55:25:03, 28.27s/it]
+
+
+ 25%|███████████████████████▎ | 2321/9378 [18:11:51<55:27:07, 28.29s/it]
+
+
+ 25%|███████████████████████▎ | 2322/9378 [18:12:19<55:23:04, 28.26s/it]
+
+ 25%|███████████████████████▎ | 2322/9378 [18:12:20<55:23:04, 28.26s/it]
+
+
+ 25%|███████████████████████▎ | 2324/9378 [18:13:15<55:16:00, 28.21s/it]
+
+
+ 25%|███████████████████████▎ | 2325/9378 [18:13:44<55:33:56, 28.36s/it]
+
+
+ 25%|███████████████████████▎ | 2326/9378 [18:14:13<56:01:11, 28.60s/it]
+
+
+ 25%|███████████████████████▎ | 2327/9378 [18:14:41<55:21:02, 28.26s/it]
+
+
+ 25%|███████████████████████▎ | 2328/9378 [18:15:09<55:25:06, 28.30s/it]
+
+
+ 25%|███████████████████████▎ | 2329/9378 [18:15:37<55:28:15, 28.33s/it]
+
+
+ 25%|███████████████████████▎ | 2330/9378 [18:16:06<55:36:44, 28.41s/it]
+
+ 25%|███████████████████████▎ | 2330/9378 [18:16:08<55:36:44, 28.41s/it]
+
+
+ 25%|███████████████████████▎ | 2332/9378 [18:17:03<55:32:17, 28.38s/it]
+
+
+ 25%|███████████████████████▍ | 2333/9378 [18:17:31<55:34:10, 28.40s/it]
+
+
+ 25%|███████████████████████▍ | 2334/9378 [18:17:59<55:29:35, 28.36s/it]
+
+
+ 25%|███████████████████████▍ | 2335/9378 [18:18:28<55:22:17, 28.30s/it]
+
+ 25%|███████████████████████▍ | 2335/9378 [18:18:29<55:22:17, 28.30s/it]
+
+
+ 25%|███████████████████████▍ | 2337/9378 [18:19:24<55:11:36, 28.22s/it]
+
+
+ 25%|███████████████████████▍ | 2338/9378 [18:19:52<55:11:51, 28.23s/it]
+
+
+ 25%|███████████████████████▍ | 2339/9378 [18:20:22<55:13:22, 28.24s/it]
+
+ 25%|███████████████████████▍ | 2340/9378 [18:20:49<55:16:51, 28.28s/it]
+
+
+ 25%|███████████████████████▍ | 2341/9378 [18:21:17<55:11:12, 28.23s/it]
+
+
+ 25%|███████████████████████▍ | 2342/9378 [18:21:46<55:13:49, 28.26s/it]
+
+
+ 25%|███████████████████████▍ | 2343/9378 [18:22:14<55:07:05, 28.21s/it]
+
+
+ 25%|███████████████████████▍ | 2344/9378 [18:22:42<55:18:29, 28.31s/it]
+
+
+ 25%|███████████████████████▌ | 2345/9378 [18:23:10<55:16:20, 28.29s/it]
+
+ 25%|███████████████████████▌ | 2345/9378 [18:23:12<55:16:20, 28.29s/it]
+
+
+ 25%|███████████████████████▌ | 2347/9378 [18:24:09<55:40:03, 28.50s/it]
+
+ 25%|███████████████████████▌ | 2348/9378 [18:24:36<55:32:59, 28.45s/it]
+
+
+ 25%|███████████████████████▌ | 2349/9378 [18:25:04<55:21:18, 28.35s/it]
+
+
+ 25%|███████████████████████▌ | 2350/9378 [18:25:53<67:25:22, 34.54s/it]
+
+
+ 25%|███████████████████████▌ | 2351/9378 [18:26:22<63:43:49, 32.65s/it]
+
+
+ 25%|███████████████████████▌ | 2352/9378 [18:26:50<61:17:44, 31.41s/it]
+
+
+ 25%|███████████████████████▌ | 2353/9378 [18:27:18<59:20:56, 30.41s/it]
+
+
+ 25%|███████████████████████▌ | 2354/9378 [18:27:47<58:21:07, 29.91s/it]
+
+
+ 25%|███████████████████████▌ | 2355/9378 [18:28:17<57:52:39, 29.67s/it]
+
+ 25%|███████████████████████▌ | 2356/9378 [18:28:45<56:36:27, 29.02s/it]
+
+ 25%|███████████████████████▋ | 2357/9378 [18:29:12<56:05:18, 28.76s/it]
+
+
+ 25%|███████████████████████▋ | 2358/9378 [18:29:40<56:01:13, 28.73s/it]
+
+ 25%|███████████████████████▋ | 2358/9378 [18:29:42<56:01:13, 28.73s/it]
+
+
+ 25%|███████████████████████▋ | 2360/9378 [18:30:37<55:24:05, 28.42s/it]
+
+
+ 25%|███████████████████████▋ | 2361/9378 [18:31:05<55:15:51, 28.35s/it]
+
+
+ 25%|███████████████████████▋ | 2362/9378 [18:31:33<55:14:42, 28.35s/it]
+
+
+ 25%|███████████████████████▋ | 2363/9378 [18:32:03<55:08:38, 28.30s/it]
+
+ 25%|███████████████████████▋ | 2364/9378 [18:32:30<55:11:16, 28.33s/it]
+
+
+ 25%|███████████████████████▋ | 2365/9378 [18:32:59<55:32:50, 28.51s/it]
+
+
+ 25%|███████████████████████▋ | 2366/9378 [18:33:27<55:00:00, 28.24s/it]
+
+
+ 25%|███████████████████████▋ | 2367/9378 [18:33:55<55:26:04, 28.46s/it]
+
+
+ 25%|███████████████████████▋ | 2368/9378 [18:34:23<55:07:44, 28.31s/it]
+
+
+
+ 25%|███████████████████████▋ | 2369/9378 [18:34:53<55:05:56, 28.30s/it]
+
+ 25%|███████████████████████▊ | 2370/9378 [18:35:22<55:07:38, 28.32s/it]
+
+ 25%|███████████████████████▊ | 2371/9378 [18:35:48<55:06:18, 28.31s/it]
+
+
+ 25%|███████████████████████▊ | 2372/9378 [18:36:16<54:58:16, 28.25s/it]
+
+
+ 25%|███████████████████████▊ | 2373/9378 [18:36:45<54:58:45, 28.25s/it]
+
+
+ 25%|███████████████████████▊ | 2374/9378 [18:37:13<55:03:00, 28.30s/it]
+
+
+ 25%|███████████████████████▊ | 2375/9378 [18:37:42<55:06:15, 28.33s/it]
+
+
+ 25%|███████████████████████▊ | 2376/9378 [18:38:10<55:04:48, 28.32s/it]
+
+
+ 25%|███████████████████████▊ | 2377/9378 [18:38:38<55:09:08, 28.36s/it]
+
+
+
+ 25%|███████████████████████▊ | 2378/9378 [18:39:08<54:48:10, 28.18s/it]
+
+
+ 25%|███████████████████████▊ | 2379/9378 [18:39:36<54:47:03, 28.18s/it]
+
+ 25%|███████████████████████▊ | 2380/9378 [18:40:04<55:06:28, 28.35s/it]
+
+
+ 25%|███████████████████████▊ | 2381/9378 [18:40:32<54:44:31, 28.17s/it]
+
+ 25%|███████████████████████▉ | 2382/9378 [18:40:59<54:54:46, 28.26s/it]
+
+
+ 25%|███████████████████████▉ | 2383/9378 [18:41:28<54:57:33, 28.28s/it]
+
+
+ 25%|███████████████████████▉ | 2384/9378 [18:41:56<55:05:42, 28.36s/it]
+
+
+ 25%|███████████████████████▉ | 2385/9378 [18:42:24<54:54:43, 28.27s/it]
+
+
+
+ 25%|███████████████████████▉ | 2386/9378 [18:42:54<54:56:23, 28.29s/it]
+
+
+ 25%|███████████████████████▉ | 2387/9378 [18:43:23<55:06:24, 28.38s/it]
+
+ 25%|███████████████████████▉ | 2388/9378 [18:43:51<55:03:10, 28.35s/it]
+
+ 25%|███████████████████████▉ | 2389/9378 [18:44:18<54:58:03, 28.31s/it]
+
+
+ 25%|███████████████████████▉ | 2390/9378 [18:44:46<54:50:02, 28.25s/it]
+
+
+ 25%|███████████████████████▉ | 2391/9378 [18:45:14<54:46:52, 28.23s/it]
+
+
+ 26%|███████████████████████▉ | 2392/9378 [18:45:42<54:47:08, 28.23s/it]
+
+
+ 26%|███████████████████████▉ | 2393/9378 [18:46:10<54:49:36, 28.26s/it]
+
+
+
+ 26%|███████████████████████▉ | 2394/9378 [18:46:40<54:54:24, 28.30s/it]
+
+ 26%|████████████████████████ | 2395/9378 [18:47:09<55:20:07, 28.53s/it]
+
+ 26%|████████████████████████ | 2396/9378 [18:47:37<55:08:09, 28.43s/it]
+
+ 26%|████████████████████████ | 2397/9378 [18:48:05<54:59:12, 28.36s/it]
+
+ 26%|████████████████████████ | 2398/9378 [18:48:32<54:49:33, 28.28s/it]
+
+
+ 26%|████████████████████████ | 2399/9378 [18:49:01<54:47:01, 28.26s/it]
+
+
+ 26%|████████████████████████ | 2400/9378 [18:49:50<67:09:13, 34.65s/it]
+
+
+ 26%|████████████████████████ | 2401/9378 [18:50:18<63:22:17, 32.70s/it]
+
+
+ 26%|████████████████████████ | 2402/9378 [18:50:46<60:40:53, 31.31s/it]
+
+
+ 26%|████████████████████████ | 2403/9378 [18:51:15<58:51:30, 30.38s/it]
+
+
+ 26%|████████████████████████ | 2404/9378 [18:51:43<57:48:34, 29.84s/it]
+
+
+
+ 26%|████████████████████████ | 2405/9378 [18:52:13<56:49:59, 29.34s/it]
+
+ 26%|████████████████████████ | 2406/9378 [18:52:41<56:41:41, 29.27s/it]
+
+ 26%|████████████████████████▏ | 2407/9378 [18:53:10<56:06:51, 28.98s/it]
+
+ 26%|████████████████████████▏ | 2408/9378 [18:53:38<55:24:16, 28.62s/it]
+
+ 26%|████████████████████████▏ | 2409/9378 [18:54:04<55:00:11, 28.41s/it]
+
+
+ 26%|████████████████████████▏ | 2410/9378 [18:54:33<54:54:54, 28.37s/it]
+
+
+ 26%|████████████████████████▏ | 2411/9378 [18:55:02<55:13:58, 28.54s/it]
+
+
+ 26%|████████████████████████▏ | 2412/9378 [18:55:29<54:36:56, 28.23s/it]
+
+
+ 26%|████████████████████████▏ | 2413/9378 [18:55:57<54:32:54, 28.19s/it]
+
+
+ 26%|████████████████████████▏ | 2414/9378 [18:56:26<54:56:47, 28.40s/it]
+
+
+ 26%|████████████████████████▏ | 2415/9378 [18:56:54<54:25:15, 28.14s/it]
+
+
+ 26%|████████████████████████▏ | 2416/9378 [18:57:22<54:29:35, 28.18s/it]
+
+
+
+ 26%|████████████████████████▏ | 2417/9378 [18:57:52<54:29:12, 28.18s/it]
+
+
+ 26%|████████████████████████▏ | 2418/9378 [18:58:20<54:33:13, 28.22s/it]
+
+ 26%|████████████████████████▏ | 2419/9378 [18:58:48<54:28:11, 28.18s/it]
+
+ 26%|████████████████████████▎ | 2420/9378 [18:59:15<54:30:46, 28.20s/it]
+
+
+ 26%|████████████████████████▎ | 2421/9378 [18:59:43<54:30:58, 28.21s/it]
+
+
+ 26%|████████████████████████▎ | 2422/9378 [19:00:11<54:28:53, 28.20s/it]
+
+
+ 26%|████████████████████████▎ | 2423/9378 [19:00:40<54:54:28, 28.42s/it]
+
+
+ 26%|████████████████████████▎ | 2424/9378 [19:01:08<54:30:20, 28.22s/it]
+
+
+ 26%|████████████████████████▎ | 2425/9378 [19:01:36<54:30:56, 28.23s/it]
+
+
+ 26%|████████████████████████▎ | 2426/9378 [19:02:05<54:44:15, 28.35s/it]
+
+ 26%|████████████████████████▎ | 2426/9378 [19:02:06<54:44:15, 28.35s/it]
+
+
+
+ 26%|████████████████████████▎ | 2428/9378 [19:03:02<54:22:45, 28.17s/it]
+
+ 26%|████████████████████████▎ | 2429/9378 [19:03:31<54:22:15, 28.17s/it]
+
+ 26%|████████████████████████▎ | 2430/9378 [19:03:58<54:52:45, 28.43s/it]
+
+
+ 26%|████████████████████████▎ | 2431/9378 [19:04:26<54:46:10, 28.38s/it]
+
+
+ 26%|████████████████████████▍ | 2432/9378 [19:04:54<54:10:18, 28.08s/it]
+
+
+ 26%|████████████████████████▍ | 2433/9378 [19:05:22<54:18:12, 28.15s/it]
+
+
+ 26%|████████████████████████▍ | 2434/9378 [19:05:50<54:20:56, 28.18s/it]
+
+
+ 26%|████████████████████████▍ | 2435/9378 [19:06:19<54:27:35, 28.24s/it]
+
+
+
+ 26%|████████████████████████▍ | 2436/9378 [19:06:49<54:31:49, 28.28s/it]
+
+ 26%|████████████████████████▍ | 2437/9378 [19:07:17<54:35:56, 28.32s/it]
+
+ 26%|████████████████████████▍ | 2438/9378 [19:07:44<54:33:26, 28.30s/it]
+
+
+ 26%|████████████████████████▍ | 2439/9378 [19:08:13<54:53:07, 28.47s/it]
+
+
+ 26%|████████████████████████▍ | 2440/9378 [19:08:40<54:29:18, 28.27s/it]
+
+
+ 26%|████████████████████████▍ | 2441/9378 [19:09:08<54:24:02, 28.23s/it]
+
+
+ 26%|████████████████████████▍ | 2442/9378 [19:09:37<54:23:01, 28.23s/it]
+
+
+ 26%|████████████████████████▍ | 2443/9378 [19:10:05<54:24:29, 28.24s/it]
+
+
+
+ 26%|████████████████████████▍ | 2444/9378 [19:10:35<54:31:47, 28.31s/it]
+
+
+ 26%|████████████████████████▌ | 2445/9378 [19:11:04<54:37:58, 28.37s/it]
+
+ 26%|████████████████████████▌ | 2446/9378 [19:11:32<54:38:48, 28.38s/it]
+
+ 26%|████████████████████████▌ | 2447/9378 [19:11:59<54:32:07, 28.33s/it]
+
+
+ 26%|████████████████████████▌ | 2448/9378 [19:12:27<54:42:12, 28.42s/it]
+
+
+ 26%|████████████████████████▌ | 2449/9378 [19:12:55<54:27:48, 28.30s/it]
+
+
+ 26%|████████████████████████▌ | 2450/9378 [19:13:45<66:49:37, 34.73s/it]
+
+
+ 26%|████████████████████████▌ | 2451/9378 [19:14:13<63:01:03, 32.75s/it]
+
+
+ 26%|████████████████████████▌ | 2452/9378 [19:14:41<60:23:39, 31.39s/it]
+
+
+
+ 26%|████████████████████████▌ | 2453/9378 [19:15:11<58:35:46, 30.46s/it]
+
+
+ 26%|████████████████████████▌ | 2454/9378 [19:15:40<57:25:45, 29.86s/it]
+
+ 26%|████████████████████████▌ | 2455/9378 [19:16:08<56:53:49, 29.59s/it]
+
+ 26%|████████████████████████▌ | 2456/9378 [19:16:36<55:46:29, 29.01s/it]
+
+ 26%|████████████████████████▋ | 2457/9378 [19:17:04<55:16:24, 28.75s/it]
+
+ 26%|████████████████████████▋ | 2458/9378 [19:17:31<54:58:22, 28.60s/it]
+
+
+ 26%|████████████████████████▋ | 2459/9378 [19:18:00<55:02:08, 28.64s/it]
+
+
+
+ 26%|████████████████████████▋ | 2460/9378 [19:18:29<55:01:18, 28.63s/it]
+
+
+ 26%|████████████████████████▋ | 2461/9378 [19:18:58<54:24:45, 28.32s/it]
+
+
+ 26%|████████████████████████▋ | 2462/9378 [19:19:26<54:27:08, 28.34s/it]
+
+
+ 26%|████████████████████████▋ | 2463/9378 [19:19:54<54:23:45, 28.32s/it]
+
+ 26%|████████████████████████▋ | 2464/9378 [19:20:22<54:44:29, 28.50s/it]
+
+
+ 26%|████████████████████████▋ | 2465/9378 [19:20:50<54:31:22, 28.39s/it]
+
+
+
+ 26%|████████████████████████▋ | 2466/9378 [19:21:20<54:28:54, 28.38s/it]
+
+
+ 26%|████████████████████████▋ | 2467/9378 [19:21:48<54:26:49, 28.36s/it]
+
+
+ 26%|████████████████████████▋ | 2468/9378 [19:22:16<54:25:40, 28.36s/it]
+
+
+ 26%|████████████████████████▋ | 2469/9378 [19:22:45<54:19:06, 28.30s/it]
+
+ 26%|████████████████████████▊ | 2470/9378 [19:23:13<54:15:38, 28.28s/it]
+
+ 26%|████████████████████████▊ | 2471/9378 [19:23:41<54:10:02, 28.23s/it]
+
+ 26%|████████████████████████▊ | 2472/9378 [19:24:07<54:10:10, 28.24s/it]
+
+
+ 26%|████████████████████████▊ | 2473/9378 [19:24:36<54:07:10, 28.22s/it]
+
+
+ 26%|████████████████████████▊ | 2474/9378 [19:25:04<54:07:28, 28.22s/it]
+
+
+
+ 26%|████████████████████████▊ | 2475/9378 [19:25:34<54:09:56, 28.25s/it]
+
+
+ 26%|████████████████████████▊ | 2476/9378 [19:26:02<54:14:41, 28.29s/it]
+
+
+ 26%|████████████████████████▊ | 2477/9378 [19:26:31<54:12:11, 28.28s/it]
+
+
+ 26%|████████████████████████▊ | 2478/9378 [19:26:59<54:12:43, 28.28s/it]
+
+
+ 26%|████████████████████████▊ | 2479/9378 [19:27:27<54:05:46, 28.23s/it]
+
+ 26%|████████████████████████▊ | 2480/9378 [19:27:55<54:08:06, 28.25s/it]
+
+ 26%|████████████████████████▊ | 2481/9378 [19:28:23<54:17:32, 28.34s/it]
+
+ 26%|████████████████████████▉ | 2482/9378 [19:28:50<54:06:28, 28.25s/it]
+
+
+ 26%|████████████████████████▉ | 2483/9378 [19:29:18<54:06:13, 28.25s/it]
+
+
+
+ 26%|████████████████████████▉ | 2484/9378 [19:29:48<54:07:01, 28.26s/it]
+
+
+ 26%|████████████████████████▉ | 2485/9378 [19:30:17<54:20:21, 28.38s/it]
+
+
+ 27%|████████████████████████▉ | 2486/9378 [19:30:45<54:13:53, 28.33s/it]
+
+ 27%|████████████████████████▉ | 2487/9378 [19:31:13<54:11:23, 28.31s/it]
+
+ 27%|████████████████████████▉ | 2488/9378 [19:31:42<54:07:49, 28.28s/it]
+
+ 27%|████████████████████████▉ | 2489/9378 [19:32:09<54:19:31, 28.39s/it]
+
+
+ 27%|████████████████████████▉ | 2490/9378 [19:32:37<54:21:49, 28.41s/it]
+
+
+
+ 27%|████████████████████████▉ | 2491/9378 [19:33:07<54:04:28, 28.27s/it]
+
+
+ 27%|████████████████████████▉ | 2492/9378 [19:33:35<54:01:52, 28.25s/it]
+
+
+ 27%|████████████████████████▉ | 2493/9378 [19:34:03<54:04:00, 28.27s/it]
+
+
+ 27%|████████████████████████▉ | 2494/9378 [19:34:31<54:00:54, 28.25s/it]
+
+
+ 27%|█████████████████████████ | 2495/9378 [19:35:00<53:59:25, 28.24s/it]
+
+
+ 27%|█████████████████████████ | 2496/9378 [19:35:28<54:01:51, 28.26s/it]
+
+
+ 27%|█████████████████████████ | 2497/9378 [19:35:56<54:03:05, 28.28s/it]
+
+ 27%|█████████████████████████ | 2498/9378 [19:36:25<54:03:17, 28.28s/it]
+
+ 27%|█████████████████████████ | 2499/9378 [19:36:52<54:16:09, 28.40s/it]
+
+
+
+ 27%|█████████████████████████ | 2500/9378 [19:37:43<66:29:32, 34.80s/it]
+
+
+ 27%|█████████████████████████ | 2501/9378 [19:38:11<62:52:11, 32.91s/it]
+
+
+ 27%|█████████████████████████ | 2502/9378 [19:38:40<60:18:41, 31.58s/it]
+
+
+ 27%|█████████████████████████ | 2503/9378 [19:39:08<58:25:53, 30.60s/it]
+
+ 27%|█████████████████████████ | 2504/9378 [19:39:37<57:18:37, 30.01s/it]
+
+ 27%|█████████████████████████ | 2505/9378 [19:40:05<56:01:12, 29.34s/it]
+
+ 27%|█████████████████████████ | 2506/9378 [19:40:31<55:27:33, 29.05s/it]
+
+
+
+ 27%|█████████████████████████▏ | 2507/9378 [19:41:01<54:57:29, 28.79s/it]
+
+
+ 27%|█████████████████████████▏ | 2508/9378 [19:41:30<54:50:40, 28.74s/it]
+{'loss': 1.0998, 'grad_norm': 0.47265625, 'learning_rate': 2.2322643839092165e-05, 'epoch': 0.8}
+[2024-05-13 06:14:23,052] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-2508/pytorch_model_fsdp.bin
+[2024-05-13 06:14:34,759] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-2508/pytorch_model_fsdp.bin
+[2024-05-13 06:14:55,968] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-2508/optimizer.bin
+ 27%|█████████████████████████▏ | 2508/9378 [19:41:30<54:50:40, 28.74s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+
+ 27%|████████████████████████▉ | 2509/9378 [19:44:02<125:45:43, 65.91s/it]
+
+
+ 27%|████████████████████████▉ | 2510/9378 [19:44:31<104:04:20, 54.55s/it]
+
+ 27%|█████████████████████████▏ | 2511/9378 [19:44:59<89:34:03, 46.96s/it]
+
+ 27%|█████████████████████████▏ | 2512/9378 [19:45:26<78:30:50, 41.17s/it]
+
+
+ 27%|█████████████████████████▏ | 2513/9378 [19:45:55<71:32:57, 37.52s/it]
+
+
+
+ 27%|█████████████████████████▏ | 2514/9378 [19:46:24<65:56:07, 34.58s/it]
+
+ 27%|█████████████████████████▏ | 2515/9378 [19:46:53<62:48:43, 32.95s/it]
+
+
+ 27%|█████████████████████████▏ | 2516/9378 [19:47:21<59:48:50, 31.38s/it]
+
+ 27%|█████████████████████████▏ | 2517/9378 [19:47:48<58:14:52, 30.56s/it]
+
+
+ 27%|█████████████████████████▏ | 2518/9378 [19:48:17<57:21:45, 30.10s/it]
+
+
+ 27%|█████████████████████████▏ | 2519/9378 [19:48:45<56:22:14, 29.59s/it]
+
+
+
+ 27%|█████████████████████████▎ | 2520/9378 [19:49:15<55:08:58, 28.95s/it]
+
+
+ 27%|█████████████████████████▎ | 2521/9378 [19:49:43<54:50:36, 28.79s/it]
+
+ 27%|█████████████████████████▎ | 2522/9378 [19:50:11<54:39:59, 28.70s/it]
+
+ 27%|█████████████████████████▎ | 2523/9378 [19:50:40<54:25:41, 28.58s/it]
+
+
+ 27%|█████████████████████████▎ | 2524/9378 [19:51:09<54:34:21, 28.66s/it]
+
+
+ 27%|█████████████████████████▎ | 2525/9378 [19:51:37<54:27:46, 28.61s/it]
+ 27%|█████████████████████████▎ | 2525/9378 [19:51:37<54:27:46, 28.61s/it]'(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/bdfc9278437b32ed7975db20c674c49d5bff415ebeedadafecc1274426a6d445?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=70377557290721e94b9bd8dd5e77ceae4e6a53268aa6b1ace714d89b76e37962&X-Amz-SignedHeaders=host&partNumber=1&uploadId=uFQntnS3dD6qfV5sfbJeUgmhQ7yQdH_kWu2xzX01ZoTSbUgxazocWGULeb_NzTak6cuLLaPYBk2OgtWragK_zvTl_C.aXirmtf2XSK6pzg2aciv2VQhj4fWAeNFRFpN7&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: 34b194dd-bec8-4950-ae26-bba36200cdf9)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/bdfc9278437b32ed7975db20c674c49d5bff415ebeedadafecc1274426a6d445?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=70377557290721e94b9bd8dd5e77ceae4e6a53268aa6b1ace714d89b76e37962&X-Amz-SignedHeaders=host&partNumber=1&uploadId=uFQntnS3dD6qfV5sfbJeUgmhQ7yQdH_kWu2xzX01ZoTSbUgxazocWGULeb_NzTak6cuLLaPYBk2OgtWragK_zvTl_C.aXirmtf2XSK6pzg2aciv2VQhj4fWAeNFRFpN7&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+[2024-05-13 06:23:49,801] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/bdfc9278437b32ed7975db20c674c49d5bff415ebeedadafecc1274426a6d445?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=70377557290721e94b9bd8dd5e77ceae4e6a53268aa6b1ace714d89b76e37962&X-Amz-SignedHeaders=host&partNumber=1&uploadId=uFQntnS3dD6qfV5sfbJeUgmhQ7yQdH_kWu2xzX01ZoTSbUgxazocWGULeb_NzTak6cuLLaPYBk2OgtWragK_zvTl_C.aXirmtf2XSK6pzg2aciv2VQhj4fWAeNFRFpN7&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: 34b194dd-bec8-4950-ae26-bba36200cdf9)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/bdfc9278437b32ed7975db20c674c49d5bff415ebeedadafecc1274426a6d445?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=70377557290721e94b9bd8dd5e77ceae4e6a53268aa6b1ace714d89b76e37962&X-Amz-SignedHeaders=host&partNumber=1&uploadId=uFQntnS3dD6qfV5sfbJeUgmhQ7yQdH_kWu2xzX01ZoTSbUgxazocWGULeb_NzTak6cuLLaPYBk2OgtWragK_zvTl_C.aXirmtf2XSK6pzg2aciv2VQhj4fWAeNFRFpN7&x-id=UploadPart
+[2024-05-13 06:23:49,813] [WARNING] [huggingface_hub.utils._http.http_backoff:302] [PID:44754] Retrying in 1s [Retry 1/5].
+[2024-05-13 06:23:54,621] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/49fd5aedddac63650a48c15001efc051b05981fa0f567f1eae40afc062a8b1e5?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=bd5730089f9ea2ca39b24bf3639dcdcd2f5616cf25a3dcfb1b224e32b54cbbad&X-Amz-SignedHeaders=host&partNumber=1&uploadId=BOvlGre32Aj9.Nj6Kv.ZisHVzaVGs_AKK7Mh08BVFDbpYuuEw48I9QxDD.od.EgIf9bHXQ4_y9njV4_RyFLOEx5nYmXTWbb1qLwxyb_xIQLQhPgZdJMFYDFsNAWewegZ&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: d63b853f-897c-44f1-a01a-246832f012df)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/49fd5aedddac63650a48c15001efc051b05981fa0f567f1eae40afc062a8b1e5?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=bd5730089f9ea2ca39b24bf3639dcdcd2f5616cf25a3dcfb1b224e32b54cbbad&X-Amz-SignedHeaders=host&partNumber=1&uploadId=BOvlGre32Aj9.Nj6Kv.ZisHVzaVGs_AKK7Mh08BVFDbpYuuEw48I9QxDD.od.EgIf9bHXQ4_y9njV4_RyFLOEx5nYmXTWbb1qLwxyb_xIQLQhPgZdJMFYDFsNAWewegZ&x-id=UploadPart
+[2024-05-13 06:23:54,622] [WARNING] [huggingface_hub.utils._http.http_backoff:302] [PID:44754] Retrying in 1s [Retry 1/5].
+'(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/49fd5aedddac63650a48c15001efc051b05981fa0f567f1eae40afc062a8b1e5?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=bd5730089f9ea2ca39b24bf3639dcdcd2f5616cf25a3dcfb1b224e32b54cbbad&X-Amz-SignedHeaders=host&partNumber=1&uploadId=BOvlGre32Aj9.Nj6Kv.ZisHVzaVGs_AKK7Mh08BVFDbpYuuEw48I9QxDD.od.EgIf9bHXQ4_y9njV4_RyFLOEx5nYmXTWbb1qLwxyb_xIQLQhPgZdJMFYDFsNAWewegZ&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: d63b853f-897c-44f1-a01a-246832f012df)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/49fd5aedddac63650a48c15001efc051b05981fa0f567f1eae40afc062a8b1e5?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=bd5730089f9ea2ca39b24bf3639dcdcd2f5616cf25a3dcfb1b224e32b54cbbad&X-Amz-SignedHeaders=host&partNumber=1&uploadId=BOvlGre32Aj9.Nj6Kv.ZisHVzaVGs_AKK7Mh08BVFDbpYuuEw48I9QxDD.od.EgIf9bHXQ4_y9njV4_RyFLOEx5nYmXTWbb1qLwxyb_xIQLQhPgZdJMFYDFsNAWewegZ&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+
+ 27%|█████████████████████████▎ | 2526/9378 [19:52:05<54:16:06, 28.51s/it]
+
+ 27%|█████████████████████████▎ | 2527/9378 [19:52:34<54:19:39, 28.55s/it]
+
+ 27%|█████████████████████████▎ | 2528/9378 [19:53:01<54:34:45, 28.68s/it]
+
+
+ 27%|█████████████████████████▎ | 2529/9378 [19:53:29<53:58:14, 28.37s/it]
+
+
+
+ 27%|█████████████████████████▎ | 2530/9378 [19:53:59<54:18:33, 28.55s/it]
+
+
+ 27%|█████████████████████████▎ | 2531/9378 [19:54:27<54:07:35, 28.46s/it]
+
+
+ 27%|█████████████████████████▍ | 2532/9378 [19:54:56<53:35:04, 28.18s/it]
+
+
+ 27%|█████████████████████████▍ | 2533/9378 [19:55:24<53:36:52, 28.20s/it]
+
+
+ 27%|█████████████████████████▍ | 2534/9378 [19:55:52<53:42:05, 28.25s/it]
+
+ 27%|█████████████████████████▍ | 2535/9378 [19:56:19<53:58:52, 28.40s/it]
+
+
+
+ 27%|█████████████████████████▍ | 2536/9378 [19:56:49<53:57:06, 28.39s/it]
+
+
+ 27%|█████████████████████████▍ | 2537/9378 [19:57:18<53:55:55, 28.38s/it]
+
+
+ 27%|█████████████████████████▍ | 2538/9378 [19:57:46<53:46:38, 28.30s/it]
+
+
+ 27%|█████████████████████████▍ | 2539/9378 [19:58:14<53:47:22, 28.31s/it]
+
+ 27%|█████████████████████████▍ | 2540/9378 [19:58:43<53:57:25, 28.41s/it]
+
+ 27%|█████████████████████████▍ | 2541/9378 [19:59:10<54:17:33, 28.59s/it]
+
+
+ 27%|█████████████████████████▍ | 2542/9378 [19:59:37<53:37:28, 28.24s/it]
+
+
+
+ 27%|█████████████████████████▍ | 2543/9378 [20:00:07<53:41:08, 28.28s/it]
+
+
+ 27%|█████████████████████████▍ | 2544/9378 [20:00:36<54:03:22, 28.48s/it]
+
+
+ 27%|█████████████████████████▌ | 2545/9378 [20:01:04<53:32:34, 28.21s/it]
+
+
+ 27%|█████████████████████████▌ | 2546/9378 [20:01:32<53:38:15, 28.26s/it]
+
+
+ 27%|█████████████████████████▌ | 2547/9378 [20:02:01<53:42:17, 28.30s/it]
+
+ 27%|█████████████████████████▌ | 2548/9378 [20:02:29<54:04:03, 28.50s/it]
+
+
+ 27%|█████████████████████████▌ | 2549/9378 [20:02:57<53:25:41, 28.17s/it]
+
+
+ 27%|█████████████████████████▌ | 2550/9378 [20:03:46<65:32:35, 34.56s/it]
+
+ 27%|█████████████████████████▌ | 2551/9378 [20:04:15<62:19:22, 32.86s/it]
+
+ 27%|█████████████████████████▌ | 2552/9378 [20:04:43<59:41:21, 31.48s/it]
+
+ 27%|█████████████████████████▌ | 2553/9378 [20:05:10<57:40:30, 30.42s/it]
+
+
+
+ 27%|█████████████████████████▌ | 2554/9378 [20:05:40<56:40:11, 29.90s/it]
+
+
+ 27%|█████████████████████████▌ | 2555/9378 [20:06:09<56:09:06, 29.63s/it]
+
+
+ 27%|█████████████████████████▌ | 2556/9378 [20:06:37<54:52:48, 28.96s/it]
+
+
+ 27%|█████████████████████████▋ | 2557/9378 [20:07:05<54:35:07, 28.81s/it]
+
+
+ 27%|█████████████████████████▋ | 2558/9378 [20:07:34<54:19:35, 28.68s/it]
+
+
+ 27%|█████████████████████████▋ | 2559/9378 [20:08:02<54:10:28, 28.60s/it]
+
+
+ 27%|█████████████████████████▋ | 2560/9378 [20:08:30<54:26:17, 28.74s/it]
+
+
+ 27%|█████████████████████████▋ | 2561/9378 [20:08:59<54:11:56, 28.62s/it]
+ 27%|█████████████████████████▋ | 2561/9378 [20:08:59<54:11:56, 28.62s/it]'(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 5d4edeee-bf3a-437d-8f64-7ee2a338127c)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/49fd5aedddac63650a48c15001efc051b05981fa0f567f1eae40afc062a8b1e5?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=bd5730089f9ea2ca39b24bf3639dcdcd2f5616cf25a3dcfb1b224e32b54cbbad&X-Amz-SignedHeaders=host&partNumber=1&uploadId=BOvlGre32Aj9.Nj6Kv.ZisHVzaVGs_AKK7Mh08BVFDbpYuuEw48I9QxDD.od.EgIf9bHXQ4_y9njV4_RyFLOEx5nYmXTWbb1qLwxyb_xIQLQhPgZdJMFYDFsNAWewegZ&x-id=UploadPart
+Retrying in 2s [Retry 2/5].
+[2024-05-13 06:41:32,434] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: 5d4edeee-bf3a-437d-8f64-7ee2a338127c)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/49fd5aedddac63650a48c15001efc051b05981fa0f567f1eae40afc062a8b1e5?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T062240Z&X-Amz-Expires=86400&X-Amz-Signature=bd5730089f9ea2ca39b24bf3639dcdcd2f5616cf25a3dcfb1b224e32b54cbbad&X-Amz-SignedHeaders=host&partNumber=1&uploadId=BOvlGre32Aj9.Nj6Kv.ZisHVzaVGs_AKK7Mh08BVFDbpYuuEw48I9QxDD.od.EgIf9bHXQ4_y9njV4_RyFLOEx5nYmXTWbb1qLwxyb_xIQLQhPgZdJMFYDFsNAWewegZ&x-id=UploadPart
+[2024-05-13 06:41:32,435] [WARNING] [huggingface_hub.utils._http.http_backoff:302] [PID:44754] Retrying in 2s [Retry 2/5].
+
+ 27%|█████████████████████████▋ | 2562/9378 [20:09:27<53:32:39, 28.28s/it]
+
+
+ 27%|█████████████████████████▋ | 2563/9378 [20:09:55<53:31:44, 28.28s/it]
+
+
+ 27%|█████████████████████████▋ | 2564/9378 [20:10:24<53:34:32, 28.31s/it]
+
+
+ 27%|█████████████████████████▋ | 2565/9378 [20:10:52<53:35:09, 28.31s/it]
+
+
+ 27%|█████████████████████████▋ | 2566/9378 [20:11:20<53:44:08, 28.40s/it]
+
+
+ 27%|█████████████████████████▋ | 2567/9378 [20:11:49<53:41:33, 28.38s/it]
+
+
+ 27%|█████████████████████████▋ | 2568/9378 [20:12:17<53:58:34, 28.53s/it]
+
+
+ 27%|█████████████████████████▊ | 2569/9378 [20:12:45<53:22:37, 28.22s/it]
+
+
+ 27%|█████████████████████████▊ | 2570/9378 [20:13:14<53:32:03, 28.31s/it]
+
+ 27%|█████████████████████████▊ | 2571/9378 [20:13:42<53:54:32, 28.51s/it]
+
+
+ 27%|█████████████████████████▊ | 2572/9378 [20:14:10<53:16:39, 28.18s/it]
+
+ 27%|█████████████████████████▊ | 2573/9378 [20:14:38<53:23:32, 28.25s/it]
+
+
+ 27%|█████████████████████████▊ | 2574/9378 [20:15:07<53:21:01, 28.23s/it]
+
+
+ 27%|█████████████████████████▊ | 2575/9378 [20:15:35<53:27:49, 28.29s/it]
+
+
+ 27%|█████████████████████████▊ | 2576/9378 [20:16:03<53:52:29, 28.51s/it]
+
+
+ 27%|█████████████████████████▊ | 2577/9378 [20:16:32<53:33:05, 28.35s/it]
+
+
+ 27%|█████████████████████████▊ | 2578/9378 [20:17:00<53:13:26, 28.18s/it]
+
+ 28%|█████████████████████████▊ | 2579/9378 [20:17:28<53:43:09, 28.44s/it]
+
+ 28%|█████████████████████████▊ | 2580/9378 [20:17:57<53:41:18, 28.43s/it]
+
+
+ 28%|█████████████████████████▊ | 2581/9378 [20:18:25<53:41:11, 28.43s/it]
+
+
+ 28%|█████████████████████████▉ | 2582/9378 [20:18:53<53:09:17, 28.16s/it]
+
+
+ 28%|█████████████████████████▉ | 2583/9378 [20:19:22<53:41:07, 28.44s/it]
+
+
+ 28%|█████████████████████████▉ | 2584/9378 [20:19:50<53:10:06, 28.17s/it]
+
+
+ 28%|█████████████████████████▉ | 2585/9378 [20:20:18<53:17:56, 28.25s/it]
+
+ 28%|█████████████████████████▉ | 2586/9378 [20:20:45<53:18:48, 28.26s/it]
+
+ 28%|█████████████████████████▉ | 2586/9378 [20:20:47<53:18:48, 28.26s/it]
+
+
+ 28%|█████████████████████████▉ | 2588/9378 [20:21:42<53:21:59, 28.29s/it]
+
+
+ 28%|█████████████████████████▉ | 2589/9378 [20:22:10<53:22:28, 28.30s/it]
+
+
+ 28%|█████████████████████████▉ | 2590/9378 [20:22:38<53:19:43, 28.28s/it]
+
+
+ 28%|█████████████████████████▉ | 2591/9378 [20:23:07<53:46:01, 28.52s/it]
+
+ 28%|█████████████████████████▉ | 2591/9378 [20:23:08<53:46:01, 28.52s/it]
+
+
+ 28%|█████████████████████████▉ | 2593/9378 [20:24:03<53:05:36, 28.17s/it]
+
+
+ 28%|██████████████████████████ | 2594/9378 [20:24:31<53:08:18, 28.20s/it]
+
+ 28%|██████████████████████████ | 2594/9378 [20:24:33<53:08:18, 28.20s/it]
+
+
+ 28%|██████████████████████████ | 2596/9378 [20:25:28<53:10:10, 28.22s/it]
+
+
+ 28%|██████████████████████████ | 2597/9378 [20:25:57<53:31:40, 28.42s/it]
+
+
+ 28%|██████████████████████████ | 2598/9378 [20:26:25<53:27:04, 28.38s/it]
+
+
+ 28%|██████████████████████████ | 2599/9378 [20:26:53<53:21:29, 28.34s/it]
+
+
+ 28%|██████████████████████████ | 2600/9378 [20:27:41<64:15:55, 34.13s/it]
+
+
+ 28%|██████████████████████████ | 2601/9378 [20:28:09<60:58:46, 32.39s/it]
+
+
+ 28%|██████████████████████████ | 2602/9378 [20:28:38<58:37:30, 31.15s/it]
+
+
+ 28%|██████████████████████████ | 2603/9378 [20:29:08<57:03:48, 30.32s/it]
+
+ 28%|██████████████████████████ | 2604/9378 [20:29:35<56:20:45, 29.94s/it]
+
+
+ 28%|██████████████████████████ | 2605/9378 [20:30:03<55:18:25, 29.40s/it]
+
+
+ 28%|██████████████████████████ | 2606/9378 [20:30:31<54:11:59, 28.81s/it]
+
+
+ 28%|██████████████████████████▏ | 2607/9378 [20:31:00<54:22:43, 28.91s/it]
+
+
+ 28%|██████████████████████████▏ | 2608/9378 [20:31:27<53:34:47, 28.49s/it]
+
+ 28%|██████████████████████████▏ | 2608/9378 [20:31:29<53:34:47, 28.49s/it]
+
+
+ 28%|██████████████████████████▏ | 2610/9378 [20:32:24<53:16:57, 28.34s/it]
+
+
+ 28%|██████████████████████████▏ | 2611/9378 [20:32:52<53:20:34, 28.38s/it]
+
+
+ 28%|██████████████████████████▏ | 2612/9378 [20:33:21<53:17:22, 28.35s/it]
+
+
+ 28%|██████████████████████████▏ | 2613/9378 [20:33:50<53:45:31, 28.61s/it]
+
+
+ 28%|██████████████████████████▏ | 2614/9378 [20:34:17<53:08:34, 28.28s/it]
+
+ 28%|██████████████████████████▏ | 2614/9378 [20:34:19<53:08:34, 28.28s/it]
+
+ 28%|██████████████████████████▏ | 2615/9378 [20:34:47<53:28:23, 28.46s/it]
+
+
+ 28%|██████████████████████████▏ | 2617/9378 [20:35:44<53:23:19, 28.43s/it]
+
+ 28%|██████████████████████████▏ | 2618/9378 [20:36:12<52:46:41, 28.11s/it]
+
+ 28%|██████████████████████████▎ | 2619/9378 [20:36:39<52:52:57, 28.17s/it]
+
+
+ 28%|██████████████████████████▎ | 2620/9378 [20:37:07<52:55:44, 28.20s/it]
+
+
+
+ 28%|██████████████████████████▎ | 2621/9378 [20:37:37<53:22:02, 28.43s/it]
+
+
+ 28%|██████████████████████████▎ | 2622/9378 [20:38:05<52:56:03, 28.21s/it]
+
+
+ 28%|██████████████████████████▎ | 2623/9378 [20:38:34<53:09:20, 28.33s/it]
+
+ 28%|██████████████████████████▎ | 2624/9378 [20:39:02<53:34:17, 28.55s/it]
+
+ 28%|██████████████████████████▎ | 2625/9378 [20:39:31<52:57:39, 28.23s/it]
+
+ 28%|██████████████████████████▎ | 2626/9378 [20:39:57<53:03:54, 28.29s/it]
+
+
+
+ 28%|██████████████████████████▎ | 2627/9378 [20:40:27<52:56:21, 28.23s/it]
+
+
+ 28%|██████████████████████████▎ | 2628/9378 [20:40:55<53:01:07, 28.28s/it]
+
+
+ 28%|██████████████████████████▎ | 2629/9378 [20:41:24<53:07:48, 28.34s/it]
+
+
+ 28%|██████████████████████████▎ | 2630/9378 [20:41:52<53:08:02, 28.35s/it]
+
+
+ 28%|██████████████████████████▎ | 2631/9378 [20:42:21<53:04:54, 28.32s/it]
+
+ 28%|██████████████████████████▍ | 2632/9378 [20:42:48<53:19:06, 28.45s/it]
+
+
+
+ 28%|██████████████████████████▍ | 2633/9378 [20:43:17<53:13:22, 28.41s/it]
+
+
+ 28%|██████████████████████████▍ | 2634/9378 [20:43:46<52:58:43, 28.28s/it]
+
+
+ 28%|██████████████████████████▍ | 2635/9378 [20:44:14<53:09:45, 28.38s/it]
+
+
+ 28%|██████████████████████████▍ | 2636/9378 [20:44:43<53:06:45, 28.36s/it]
+
+
+ 28%|██████████████████████████▍ | 2637/9378 [20:45:11<52:59:53, 28.30s/it]
+
+ 28%|██████████████████████████▍ | 2638/9378 [20:45:39<53:00:44, 28.32s/it]
+
+ 28%|██████████████████████████▍ | 2639/9378 [20:46:06<52:59:12, 28.31s/it]
+
+
+
+ 28%|██████████████████████████▍ | 2640/9378 [20:46:36<53:26:58, 28.56s/it]
+
+
+ 28%|██████████████████████████▍ | 2641/9378 [20:47:04<52:54:42, 28.27s/it]
+
+
+ 28%|██████████████████████████▍ | 2642/9378 [20:47:32<52:54:28, 28.28s/it]
+
+
+ 28%|██████████████████████████▍ | 2643/9378 [20:48:01<52:52:50, 28.27s/it]
+
+ 28%|██████████████████████████▌ | 2644/9378 [20:48:29<53:19:33, 28.51s/it]
+
+
+ 28%|██████████████████████████▌ | 2645/9378 [20:48:57<52:44:48, 28.20s/it]
+
+ 28%|██████████████████████████▌ | 2646/9378 [20:49:25<53:12:31, 28.45s/it]
+
+ 28%|██████████████████████████▌ | 2647/9378 [20:49:53<53:03:10, 28.37s/it]
+
+
+ 28%|██████████████████████████▌ | 2648/9378 [20:50:21<53:01:58, 28.37s/it]
+
+
+ 28%|██████████████████████████▌ | 2649/9378 [20:50:49<52:31:30, 28.10s/it]
+
+
+ 28%|██████████████████████████▌ | 2650/9378 [20:51:36<63:26:07, 33.94s/it]
+
+
+ 28%|██████████████████████████▌ | 2651/9378 [20:52:04<60:12:54, 32.22s/it]
+
+
+ 28%|██████████████████████████▌ | 2652/9378 [20:52:33<58:14:32, 31.17s/it]
+
+
+ 28%|██████████████████████████▌ | 2653/9378 [20:53:01<56:29:34, 30.24s/it]
+
+
+ 28%|██████████████████████████▌ | 2654/9378 [20:53:29<55:23:36, 29.66s/it]
+
+ 28%|██████████████████████████▌ | 2654/9378 [20:53:31<55:23:36, 29.66s/it]
+
+ 28%|██████████████████████████▌ | 2655/9378 [20:54:00<55:04:23, 29.49s/it]
+
+
+ 28%|██████████████████████████▋ | 2657/9378 [20:54:54<53:39:39, 28.74s/it]
+
+
+ 28%|██████████████████████████▋ | 2658/9378 [20:55:23<53:33:49, 28.69s/it]
+
+
+ 28%|██████████████████████████▋ | 2659/9378 [20:55:51<53:17:52, 28.56s/it]
+
+
+ 28%|██████████████████████████▋ | 2660/9378 [20:56:20<53:09:54, 28.49s/it]
+
+
+ 28%|██████████████████████████▋ | 2661/9378 [20:56:48<53:10:24, 28.50s/it]
+
+
+ 28%|██████████████████████████▋ | 2662/9378 [20:57:17<53:36:25, 28.74s/it]
+
+
+ 28%|██████████████████████████▋ | 2663/9378 [20:57:45<52:53:34, 28.36s/it]
+
+
+ 28%|██████████████████████████▋ | 2664/9378 [20:58:13<52:52:39, 28.35s/it]
+
+
+ 28%|██████████████████████████▋ | 2665/9378 [20:58:42<52:57:32, 28.40s/it]
+
+
+ 28%|██████████████████████████▋ | 2666/9378 [20:59:10<52:48:56, 28.33s/it]
+
+
+ 28%|██████████████████████████▋ | 2667/9378 [20:59:38<52:54:37, 28.38s/it]
+
+ 28%|██████████████████████████▋ | 2667/9378 [20:59:40<52:54:37, 28.38s/it]
+
+
+ 28%|██████████████████████████▊ | 2669/9378 [21:00:35<52:38:05, 28.24s/it]
+
+
+ 28%|██████████████████████████▊ | 2670/9378 [21:01:05<53:10:21, 28.54s/it]
+
+ 28%|██████████████████████████▊ | 2671/9378 [21:01:33<53:07:01, 28.51s/it]
+
+ 28%|██████████████████████████▊ | 2672/9378 [21:02:00<52:33:38, 28.22s/it]
+
+
+ 29%|██████████████████████████▊ | 2673/9378 [21:02:29<52:40:34, 28.28s/it]
+
+
+ 29%|██████████████████████████▊ | 2674/9378 [21:02:57<52:53:47, 28.41s/it]
+
+
+ 29%|██████████████████████████▊ | 2675/9378 [21:03:27<53:07:22, 28.53s/it]
+
+ 29%|██████████████████████████▊ | 2676/9378 [21:03:55<52:32:57, 28.23s/it]
+
+ 29%|██████████████████████████▊ | 2677/9378 [21:04:22<52:35:31, 28.25s/it]
+
+
+ 29%|██████████████████████████▊ | 2678/9378 [21:04:50<52:35:15, 28.26s/it]
+
+
+ 29%|██████████████████████████▊ | 2679/9378 [21:05:19<53:04:55, 28.53s/it]
+
+
+ 29%|██████████████████████████▊ | 2680/9378 [21:05:47<52:37:05, 28.28s/it]
+
+
+ 29%|██████████████████████████▊ | 2681/9378 [21:06:17<53:04:23, 28.53s/it]
+
+ 29%|██████████████████████████▉ | 2682/9378 [21:06:45<52:57:35, 28.47s/it]
+
+ 29%|██████████████████████████▉ | 2683/9378 [21:07:12<52:25:32, 28.19s/it]
+
+
+ 29%|██████████████████████████▉ | 2684/9378 [21:07:40<52:26:56, 28.21s/it]
+
+
+ 29%|██████████████████████████▉ | 2685/9378 [21:08:08<52:28:58, 28.23s/it]
+
+
+ 29%|██████████████████████████▉ | 2686/9378 [21:08:37<52:35:14, 28.29s/it]
+
+
+ 29%|██████████████████████████▉ | 2687/9378 [21:09:05<52:37:50, 28.32s/it]
+
+
+ 29%|██████████████████████████▉ | 2688/9378 [21:09:33<52:33:10, 28.28s/it]
+
+
+ 29%|██████████████████████████▉ | 2689/9378 [21:10:03<52:53:15, 28.46s/it]
+
+ 29%|██████████████████████████▉ | 2690/9378 [21:10:32<52:24:09, 28.21s/it]
+
+ 29%|██████████████████████████▉ | 2691/9378 [21:11:00<52:47:36, 28.42s/it]
+
+ 29%|██████████████████████████▉ | 2692/9378 [21:11:26<52:16:43, 28.15s/it]
+
+
+ 29%|██████████████████████████▉ | 2693/9378 [21:11:55<52:22:06, 28.20s/it]
+
+
+ 29%|███████████████████████████ | 2694/9378 [21:12:23<52:26:38, 28.25s/it]
+
+
+ 29%|███████████████████████████ | 2695/9378 [21:12:52<52:37:11, 28.35s/it]
+
+
+
+ 29%|███████████████████████████ | 2696/9378 [21:13:22<52:30:45, 28.29s/it]
+
+
+ 29%|███████████████████████████ | 2697/9378 [21:13:50<52:31:18, 28.30s/it]
+
+ 29%|███████████████████████████ | 2698/9378 [21:14:18<52:30:56, 28.30s/it]
+
+ 29%|███████████████████████████ | 2699/9378 [21:14:46<52:56:45, 28.54s/it]
+
+ 29%|███████████████████████████ | 2700/9378 [21:15:35<64:28:25, 34.76s/it]
+
+
+ 29%|███████████████████████████ | 2701/9378 [21:16:03<60:54:37, 32.84s/it]
+
+
+ 29%|███████████████████████████ | 2702/9378 [21:16:32<58:44:35, 31.68s/it]
+
+
+ 29%|███████████████████████████ | 2703/9378 [21:17:00<56:26:21, 30.44s/it]
+
+
+ 29%|███████████████████████████ | 2704/9378 [21:17:28<55:13:43, 29.79s/it]
+
+
+
+ 29%|███████████████████████████ | 2705/9378 [21:17:58<54:31:20, 29.41s/it]
+
+ 29%|███████████████████████████ | 2706/9378 [21:18:26<54:20:58, 29.33s/it]
+
+
+ 29%|███████████████████████████▏ | 2707/9378 [21:18:55<53:46:47, 29.02s/it]
+
+ 29%|███████████████████████████▏ | 2708/9378 [21:19:21<52:55:17, 28.56s/it]
+
+
+ 29%|███████████████████████████▏ | 2709/9378 [21:19:50<53:05:58, 28.66s/it]
+
+
+ 29%|███████████████████████████▏ | 2710/9378 [21:20:19<53:04:39, 28.66s/it]
+
+
+
+ 29%|███████████████████████████▏ | 2711/9378 [21:20:48<52:32:41, 28.37s/it]
+
+
+ 29%|███████████████████████████▏ | 2712/9378 [21:21:17<52:27:19, 28.33s/it]
+
+
+ 29%|███████████████████████████▏ | 2713/9378 [21:21:45<52:25:05, 28.31s/it]
+
+ 29%|███████████████████████████▏ | 2714/9378 [21:22:13<52:47:22, 28.52s/it]
+
+ 29%|███████████████████████████▏ | 2715/9378 [21:22:40<52:16:44, 28.25s/it]
+
+
+ 29%|███████████████████████████▏ | 2716/9378 [21:23:08<52:19:43, 28.28s/it]
+
+
+ 29%|███████████████████████████▏ | 2717/9378 [21:23:36<52:19:53, 28.28s/it]
+
+ 29%|███████████████████████████▏ | 2717/9378 [21:23:38<52:19:53, 28.28s/it]
+[2024-05-13 07:56:33,767] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-2717/pytorch_model_fsdp.bin
+[2024-05-13 07:56:45,476] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-2717/pytorch_model_fsdp.bin
+[2024-05-13 07:57:06,701] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-2717/optimizer.bin
+ 29%|███████████████████████████▏ | 2717/9378 [21:23:38<52:19:53, 28.28s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 29%|██████████████████████████▉ | 2718/9378 [21:26:13<123:32:27, 66.78s/it]
+
+
+ 29%|██████████████████████████▉ | 2719/9378 [21:26:41<102:11:15, 55.24s/it]
+
+
+ 29%|███████████████████████████▎ | 2720/9378 [21:27:11<87:20:35, 47.23s/it]
+
+ 29%|███████████████████████████▎ | 2721/9378 [21:27:38<76:50:10, 41.55s/it]
+
+
+ 29%|███████████████████████████▎ | 2722/9378 [21:28:07<69:31:25, 37.60s/it]
+
+
+ 29%|███████████████████████████▎ | 2723/9378 [21:28:35<64:18:34, 34.79s/it]
+
+
+ 29%|███████████████████████████▎ | 2724/9378 [21:29:03<60:40:57, 32.83s/it]
+
+ 29%|███████████████████████████▎ | 2724/9378 [21:29:05<60:40:57, 32.83s/it]
+
+
+
+ 29%|███████████████████████████▎ | 2726/9378 [21:30:02<56:23:27, 30.52s/it]
+
+ 29%|███████████████████████████▎ | 2727/9378 [21:30:29<55:20:57, 29.96s/it]
+
+
+ 29%|███████████████████████████▎ | 2728/9378 [21:30:57<54:36:01, 29.56s/it]
+
+
+ 29%|███████████████████████████▎ | 2729/9378 [21:31:25<53:43:28, 29.09s/it]
+
+
+ 29%|███████████████████████████▎ | 2730/9378 [21:31:54<53:20:21, 28.88s/it]
+
+
+
+ 29%|███████████████████████████▎ | 2731/9378 [21:32:24<53:01:08, 28.72s/it]
+
+ 29%|███████████████████████████▍ | 2732/9378 [21:32:52<52:56:48, 28.68s/it]
+
+ 29%|███████████████████████████▍ | 2733/9378 [21:33:19<52:43:59, 28.57s/it]
+
+
+ 29%|███████████████████████████▍ | 2734/9378 [21:33:47<52:38:12, 28.52s/it]
+
+ 29%|███████████████████████████▍ | 2734/9378 [21:33:49<52:38:12, 28.52s/it]
+
+ 29%|███████████████████████████▍ | 2735/9378 [21:34:17<52:58:22, 28.71s/it]
+
+
+ 29%|███████████████████████████▍ | 2737/9378 [21:35:14<52:18:10, 28.35s/it]
+
+ 29%|███████████████████████████▍ | 2738/9378 [21:35:41<52:20:16, 28.38s/it]
+
+
+ 29%|███████████████████████████▍ | 2739/9378 [21:36:09<52:16:51, 28.35s/it]
+
+
+ 29%|███████████████████████████▍ | 2740/9378 [21:36:38<52:26:26, 28.44s/it]
+
+
+ 29%|███████████████████████████▍ | 2741/9378 [21:37:06<52:26:07, 28.44s/it]
+
+
+ 29%|███████████████████████████▍ | 2742/9378 [21:37:36<52:25:55, 28.44s/it]
+
+ 29%|███████████████████████████▍ | 2743/9378 [21:38:05<52:35:12, 28.53s/it]
+
+ 29%|███████████████████████████▌ | 2744/9378 [21:38:31<52:13:55, 28.34s/it]
+
+
+ 29%|███████████████████████████▌ | 2745/9378 [21:39:00<52:05:44, 28.27s/it]
+
+
+ 29%|███████████████████████████▌ | 2746/9378 [21:39:28<52:07:12, 28.29s/it]
+
+
+ 29%|███████████████████████████▌ | 2747/9378 [21:39:56<52:10:57, 28.33s/it]
+
+
+
+ 29%|███████████████████████████▌ | 2748/9378 [21:40:26<52:10:47, 28.33s/it]
+
+ 29%|███████████████████████████▌ | 2749/9378 [21:40:55<52:28:28, 28.50s/it]
+
+ 29%|███████████████████████████▌ | 2750/9378 [21:41:44<64:19:17, 34.94s/it]
+
+
+ 29%|███████████████████████████▌ | 2751/9378 [21:42:12<60:49:06, 33.04s/it]
+
+
+ 29%|███████████████████████████▌ | 2752/9378 [21:42:41<58:15:20, 31.65s/it]
+
+ 29%|███████████████████████████▌ | 2752/9378 [21:42:42<58:15:20, 31.65s/it]
+
+
+ 29%|███████████████████████████▌ | 2754/9378 [21:43:39<55:24:07, 30.11s/it]
+
+
+ 29%|███████████████████████████▌ | 2755/9378 [21:44:07<53:55:34, 29.31s/it]
+
+
+ 29%|███████████████████████████▌ | 2756/9378 [21:44:35<53:22:33, 29.02s/it]
+
+
+ 29%|███████████████████████████▋ | 2757/9378 [21:45:04<52:54:59, 28.77s/it]
+
+ 29%|███████████████████████████▋ | 2758/9378 [21:45:30<52:44:24, 28.68s/it]
+
+
+ 29%|███████████████████████████▋ | 2759/9378 [21:45:59<52:30:14, 28.56s/it]
+
+
+ 29%|███████████████████████████▋ | 2760/9378 [21:46:27<52:25:51, 28.52s/it]
+
+
+
+ 29%|███████████████████████████▋ | 2761/9378 [21:46:57<52:18:38, 28.46s/it]
+
+
+ 29%|███████████████████████████▋ | 2762/9378 [21:47:25<52:12:46, 28.41s/it]
+
+
+ 29%|███████████████████████████▋ | 2763/9378 [21:47:54<52:07:04, 28.36s/it]
+
+ 29%|███████████████████████████▋ | 2764/9378 [21:48:22<52:03:44, 28.34s/it]
+
+ 29%|███████████████████████████▋ | 2765/9378 [21:48:49<52:04:01, 28.34s/it]
+
+
+ 29%|███████████████████████████▋ | 2766/9378 [21:49:17<52:03:19, 28.34s/it]
+
+
+ 30%|███████████████████████████▋ | 2767/9378 [21:49:46<52:12:45, 28.43s/it]
+
+
+
+ 30%|███████████████████████████▋ | 2768/9378 [21:50:15<52:04:51, 28.36s/it]
+
+
+ 30%|███████████████████████████▊ | 2769/9378 [21:50:44<52:01:20, 28.34s/it]
+
+ 30%|███████████████████████████▊ | 2770/9378 [21:51:12<52:03:59, 28.37s/it]
+
+ 30%|███████████████████████████▊ | 2771/9378 [21:51:39<52:02:20, 28.35s/it]
+
+
+ 30%|███████████████████████████▊ | 2772/9378 [21:52:07<52:00:24, 28.34s/it]
+
+
+
+ 30%|███████████████████████████▊ | 2773/9378 [21:52:37<52:02:04, 28.36s/it]
+
+
+ 30%|███████████████████████████▊ | 2774/9378 [21:53:05<52:22:36, 28.55s/it]
+
+
+ 30%|███████████████████████████▊ | 2775/9378 [21:53:34<51:56:21, 28.32s/it]
+
+ 30%|███████████████████████████▊ | 2776/9378 [21:54:02<51:59:21, 28.35s/it]
+
+ 30%|███████████████████████████▊ | 2777/9378 [21:54:29<51:58:22, 28.34s/it]
+
+
+ 30%|███████████████████████████▊ | 2778/9378 [21:54:57<51:52:03, 28.29s/it]
+
+
+ 30%|███████████████████████████▊ | 2779/9378 [21:55:26<51:58:13, 28.35s/it]
+
+
+
+ 30%|███████████████████████████▊ | 2780/9378 [21:55:56<51:54:45, 28.32s/it]
+
+
+ 30%|███████████████████████████▉ | 2781/9378 [21:56:24<51:52:49, 28.31s/it]
+
+
+ 30%|███████████████████████████▉ | 2782/9378 [21:56:52<51:54:52, 28.33s/it]
+
+
+ 30%|███████████████████████████▉ | 2783/9378 [21:57:21<51:51:22, 28.31s/it]
+
+ 30%|███████████████████████████▉ | 2784/9378 [21:57:47<51:55:54, 28.35s/it]
+
+
+ 30%|███████████████████████████▉ | 2785/9378 [21:58:16<51:58:36, 28.38s/it]
+
+
+
+ 30%|███████████████████████████▉ | 2786/9378 [21:58:46<51:52:32, 28.33s/it]
+
+
+ 30%|███████████████████████████▉ | 2787/9378 [21:59:14<52:04:56, 28.45s/it]
+
+ 30%|███████████████████████████▉ | 2788/9378 [21:59:43<52:06:18, 28.46s/it]
+
+ 30%|███████████████████████████▉ | 2789/9378 [22:00:09<51:59:37, 28.41s/it]
+
+
+ 30%|███████████████████████████▉ | 2790/9378 [22:00:38<52:06:36, 28.48s/it]
+
+
+ 30%|███████████████████████████▉ | 2791/9378 [22:01:06<51:47:52, 28.31s/it]
+
+
+
+ 30%|███████████████████████████▉ | 2792/9378 [22:01:36<52:15:48, 28.57s/it]
+
+ 30%|███████████████████████████▉ | 2793/9378 [22:02:05<52:11:18, 28.53s/it]
+
+
+ 30%|████████████████████████████ | 2794/9378 [22:02:33<51:42:48, 28.28s/it]
+
+ 30%|████████████████████████████ | 2795/9378 [22:03:01<51:36:36, 28.22s/it]
+
+ 30%|████████████████████████████ | 2796/9378 [22:03:28<51:48:43, 28.34s/it]
+
+
+ 30%|████████████████████████████ | 2797/9378 [22:03:56<51:47:34, 28.33s/it]
+
+
+
+ 30%|████████████████████████████ | 2798/9378 [22:04:27<51:57:29, 28.43s/it]
+
+
+ 30%|████████████████████████████ | 2799/9378 [22:04:55<51:57:37, 28.43s/it]
+
+ 30%|████████████████████████████ | 2800/9378 [22:05:41<62:24:39, 34.16s/it]
+
+
+ 30%|████████████████████████████ | 2801/9378 [22:06:11<59:27:57, 32.55s/it]
+
+
+ 30%|████████████████████████████ | 2802/9378 [22:06:39<56:48:58, 31.10s/it]
+
+ 30%|████████████████████████████ | 2803/9378 [22:07:07<55:48:30, 30.56s/it]
+
+
+ 30%|████████████████████████████ | 2804/9378 [22:07:35<54:29:30, 29.84s/it]
+
+
+ 30%|████████████████████████████ | 2805/9378 [22:08:03<53:23:21, 29.24s/it]
+
+
+
+ 30%|████████████████████████████▏ | 2806/9378 [22:08:33<52:59:40, 29.03s/it]
+
+
+ 30%|████████████████████████████▏ | 2807/9378 [22:09:02<52:41:34, 28.87s/it]
+
+
+ 30%|████████████████████████████▏ | 2808/9378 [22:09:30<52:19:17, 28.67s/it]
+
+
+ 30%|████████████████████████████▏ | 2809/9378 [22:09:58<52:08:55, 28.58s/it]
+
+ 30%|████████████████████████████▏ | 2810/9378 [22:10:25<52:06:55, 28.57s/it]
+
+
+ 30%|████████████████████████████▏ | 2811/9378 [22:10:53<52:00:13, 28.51s/it]
+
+
+
+ 30%|████████████████████████████▏ | 2812/9378 [22:11:23<52:00:05, 28.51s/it]
+
+
+ 30%|████████████████████████████▏ | 2813/9378 [22:11:52<51:51:07, 28.43s/it]
+
+
+ 30%|████████████████████████████▏ | 2814/9378 [22:12:20<51:45:09, 28.38s/it]
+
+
+ 30%|████████████████████████████▏ | 2815/9378 [22:12:48<51:42:27, 28.36s/it]
+
+ 30%|████████████████████████████▏ | 2816/9378 [22:13:15<51:42:12, 28.37s/it]
+
+
+ 30%|████████████████████████████▏ | 2817/9378 [22:13:43<51:42:50, 28.38s/it]
+
+
+ 30%|████████████████████████████▏ | 2818/9378 [22:14:12<51:40:46, 28.36s/it]
+
+
+
+ 30%|████████████████████████████▎ | 2819/9378 [22:14:42<51:41:34, 28.37s/it]
+
+ 30%|████████████████████████████▎ | 2820/9378 [22:15:10<52:05:34, 28.60s/it]
+
+
+ 30%|████████████████████████████▎ | 2821/9378 [22:15:39<51:31:39, 28.29s/it]
+
+ 30%|████████████████████████████▎ | 2822/9378 [22:16:07<51:32:59, 28.31s/it]
+
+ 30%|████████████████████████████▎ | 2823/9378 [22:16:34<51:35:07, 28.33s/it]
+
+
+ 30%|████████████████████████████▎ | 2824/9378 [22:17:02<51:37:43, 28.36s/it]
+
+
+
+ 30%|████████████████████████████▎ | 2825/9378 [22:17:32<51:38:06, 28.37s/it]
+
+
+ 30%|████████████████████████████▎ | 2826/9378 [22:18:00<51:33:46, 28.33s/it]
+
+
+ 30%|████████████████████████████▎ | 2827/9378 [22:18:29<51:31:09, 28.31s/it]
+
+ 30%|████████████████████████████▎ | 2828/9378 [22:18:56<52:04:22, 28.62s/it]
+
+
+ 30%|████████████████████████████▎ | 2829/9378 [22:19:24<51:30:19, 28.31s/it]
+
+
+ 30%|████████████████████████████▎ | 2830/9378 [22:19:52<51:27:54, 28.29s/it]
+
+
+
+ 30%|████████████████████████████▍ | 2831/9378 [22:20:22<51:36:20, 28.38s/it]
+
+
+ 30%|████████████████████████████▍ | 2832/9378 [22:20:50<51:27:47, 28.30s/it]
+
+
+ 30%|████████████████████████████▍ | 2833/9378 [22:21:19<51:25:54, 28.29s/it]
+
+ 30%|████████████████████████████▍ | 2834/9378 [22:21:47<51:34:53, 28.38s/it]
+
+ 30%|████████████████████████████▍ | 2835/9378 [22:22:14<51:37:02, 28.40s/it]
+
+
+
+ 30%|████████████████████████████▍ | 2836/9378 [22:22:44<51:39:07, 28.42s/it]
+
+
+ 30%|████████████████████████████▍ | 2837/9378 [22:23:13<51:46:09, 28.49s/it]
+
+
+ 30%|████████████████████████████▍ | 2838/9378 [22:23:41<51:36:25, 28.41s/it]
+
+ 30%|████████████████████████████▍ | 2839/9378 [22:24:09<51:34:50, 28.40s/it]
+
+ 30%|████████████████████████████▍ | 2840/9378 [22:24:36<51:27:19, 28.33s/it]
+
+
+ 30%|████████████████████████████▍ | 2841/9378 [22:25:04<51:29:47, 28.36s/it]
+
+
+
+ 30%|████████████████████████████▍ | 2842/9378 [22:25:35<51:46:20, 28.52s/it]
+
+
+ 30%|████████████████████████████▍ | 2843/9378 [22:26:03<51:23:58, 28.31s/it]
+
+
+ 30%|████████████████████████████▌ | 2844/9378 [22:26:31<51:18:39, 28.27s/it]
+
+
+ 30%|████████████████████████████▌ | 2845/9378 [22:26:59<51:25:25, 28.34s/it]
+
+
+ 30%|████████████████████████████▌ | 2846/9378 [22:27:28<51:20:21, 28.29s/it]
+
+ 30%|████████████████████████████▌ | 2847/9378 [22:27:54<51:28:19, 28.37s/it]
+
+
+
+ 30%|████████████████████████████▌ | 2848/9378 [22:28:25<51:29:28, 28.39s/it]
+
+
+ 30%|████████████████████████████▌ | 2849/9378 [22:28:53<51:59:16, 28.67s/it]
+
+
+ 30%|████████████████████████████▌ | 2850/9378 [22:29:43<63:14:26, 34.88s/it]
+
+
+ 30%|████████████████████████████▌ | 2851/9378 [22:30:12<59:37:56, 32.89s/it]
+
+
+ 30%|████████████████████████████▌ | 2852/9378 [22:30:40<57:02:35, 31.47s/it]
+
+
+ 30%|████████████████████████████▌ | 2853/9378 [22:31:08<55:25:09, 30.58s/it]
+
+ 30%|████████████████████████████▌ | 2854/9378 [22:31:35<54:20:12, 29.98s/it]
+
+
+
+ 30%|████████████████████████████▌ | 2855/9378 [22:32:05<53:28:38, 29.51s/it]
+
+
+ 30%|████████████████████████████▋ | 2856/9378 [22:32:34<52:49:57, 29.16s/it]
+
+
+ 30%|████████████████████████████▋ | 2857/9378 [22:33:02<52:21:58, 28.91s/it]
+
+
+ 30%|████████████████████████████▋ | 2858/9378 [22:33:30<52:07:09, 28.78s/it]
+
+
+ 30%|████████████████████████████▋ | 2859/9378 [22:33:59<51:54:33, 28.67s/it]
+
+
+ 30%|████████████████████████████▋ | 2860/9378 [22:34:27<51:47:22, 28.60s/it]
+
+
+ 31%|████████████████████████████▋ | 2861/9378 [22:34:56<52:04:03, 28.76s/it]
+
+
+ 31%|████████████████████████████▋ | 2862/9378 [22:35:24<51:25:22, 28.41s/it]
+
+
+ 31%|████████████████████████████▋ | 2863/9378 [22:35:53<51:35:48, 28.51s/it]
+
+ 31%|████████████████████████████▋ | 2864/9378 [22:36:21<51:32:03, 28.48s/it]
+
+
+ 31%|████████████████████████████▋ | 2865/9378 [22:36:50<51:52:55, 28.68s/it]
+
+
+ 31%|████████████████████████████▋ | 2866/9378 [22:37:18<51:18:27, 28.36s/it]
+
+
+ 31%|████████████████████████████▋ | 2867/9378 [22:37:46<51:18:31, 28.37s/it]
+
+ 31%|████████████████████████████▋ | 2868/9378 [22:38:15<51:43:15, 28.60s/it]
+
+
+ 31%|████████████████████████████▊ | 2869/9378 [22:38:43<51:06:47, 28.27s/it]
+
+ 31%|████████████████████████████▊ | 2870/9378 [22:39:11<51:12:54, 28.33s/it]
+
+
+ 31%|████████████████████████████▊ | 2871/9378 [22:39:40<51:12:27, 28.33s/it]
+
+
+ 31%|████████████████████████████▊ | 2872/9378 [22:40:08<51:15:48, 28.37s/it]
+
+
+ 31%|████████████████████████████▊ | 2873/9378 [22:40:37<51:14:25, 28.36s/it]
+
+
+ 31%|████████████████████████████▊ | 2874/9378 [22:41:05<51:27:06, 28.48s/it]
+
+ 31%|████████████████████████████▊ | 2875/9378 [22:41:33<51:24:51, 28.46s/it]
+
+ 31%|████████████████████████████▊ | 2876/9378 [22:42:00<51:08:03, 28.31s/it]
+
+
+
+ 31%|████████████████████████████▊ | 2877/9378 [22:42:30<51:12:26, 28.36s/it]
+
+
+ 31%|████████████████████████████▊ | 2878/9378 [22:42:58<51:21:21, 28.44s/it]
+
+
+ 31%|████████████████████████████▊ | 2879/9378 [22:43:27<51:08:55, 28.33s/it]
+
+
+ 31%|████████████████████████████▊ | 2880/9378 [22:43:55<51:08:31, 28.33s/it]
+
+
+ 31%|████████████████████████████▉ | 2881/9378 [22:44:24<51:10:34, 28.36s/it]
+
+
+ 31%|████████████████████████████▉ | 2882/9378 [22:44:52<51:14:10, 28.39s/it]
+
+
+ 31%|████████████████████████████▉ | 2883/9378 [22:45:20<51:35:48, 28.60s/it]
+
+
+ 31%|████████████████████████████▉ | 2884/9378 [22:45:49<51:06:44, 28.33s/it]
+
+
+ 31%|████████████████████████████▉ | 2885/9378 [22:46:17<51:06:27, 28.34s/it]
+
+
+ 31%|████████████████████████████▉ | 2886/9378 [22:46:46<51:12:55, 28.40s/it]
+
+
+ 31%|████████████████████████████▉ | 2887/9378 [22:47:14<51:05:13, 28.33s/it]
+
+
+ 31%|████████████████████████████▉ | 2888/9378 [22:47:42<51:36:11, 28.62s/it]
+
+
+ 31%|████████████████████████████▉ | 2889/9378 [22:48:11<51:01:33, 28.31s/it]
+
+
+ 31%|████████████████████████████▉ | 2890/9378 [22:48:39<51:05:13, 28.35s/it]
+
+
+ 31%|████████████████████████████▉ | 2891/9378 [22:49:08<51:07:00, 28.37s/it]
+
+
+ 31%|████████████████████████████▉ | 2892/9378 [22:49:36<51:08:07, 28.38s/it]
+
+
+ 31%|████████████████████████████▉ | 2893/9378 [22:50:05<51:12:26, 28.43s/it]
+
+
+ 31%|█████████████████████████████ | 2894/9378 [22:50:33<51:03:17, 28.35s/it]
+
+
+ 31%|█████████████████████████████ | 2895/9378 [22:51:01<51:07:42, 28.39s/it]
+
+
+ 31%|█████████████████████████████ | 2896/9378 [22:51:30<51:01:44, 28.34s/it]
+
+
+ 31%|█████████████████████████████ | 2897/9378 [22:51:58<51:04:06, 28.37s/it]
+
+ 31%|█████████████████████████████ | 2898/9378 [22:52:26<51:29:11, 28.60s/it]
+
+ 31%|█████████████████████████████ | 2899/9378 [22:52:55<50:53:20, 28.28s/it]
+
+
+ 31%|█████████████████████████████ | 2900/9378 [22:53:42<61:21:05, 34.09s/it]
+
+ 31%|█████████████████████████████ | 2901/9378 [22:54:11<58:14:18, 32.37s/it]
+
+
+ 31%|█████████████████████████████ | 2902/9378 [22:54:39<56:02:15, 31.15s/it]
+
+
+ 31%|█████████████████████████████ | 2903/9378 [22:55:07<54:35:17, 30.35s/it]
+
+
+ 31%|█████████████████████████████ | 2904/9378 [22:55:36<53:20:43, 29.66s/it]
+
+ 31%|█████████████████████████████ | 2905/9378 [22:56:02<52:40:42, 29.30s/it]
+
+
+ 31%|█████████████████████████████▏ | 2906/9378 [22:56:31<52:22:50, 29.14s/it]
+
+
+ 31%|█████████████████████████████▏ | 2907/9378 [22:56:59<52:02:29, 28.95s/it]
+
+
+ 31%|█████████████████████████████▏ | 2908/9378 [22:57:30<51:55:00, 28.89s/it]
+
+ 31%|█████████████████████████████▏ | 2909/9378 [22:57:56<51:19:40, 28.56s/it]
+
+
+ 31%|█████████████████████████████▏ | 2910/9378 [22:58:25<51:36:58, 28.73s/it]
+
+
+ 31%|█████████████████████████████▏ | 2911/9378 [22:58:53<50:54:23, 28.34s/it]
+
+
+ 31%|█████████████████████████████▏ | 2912/9378 [22:59:21<51:06:48, 28.46s/it]
+
+
+ 31%|█████████████████████████████▏ | 2913/9378 [22:59:49<50:52:09, 28.33s/it]
+
+
+
+ 31%|█████████████████████████████▏ | 2914/9378 [23:00:19<50:53:18, 28.34s/it]
+
+ 31%|█████████████████████████████▏ | 2915/9378 [23:00:46<50:58:16, 28.39s/it]
+
+
+ 31%|█████████████████████████████▏ | 2916/9378 [23:01:14<50:55:11, 28.37s/it]
+
+
+ 31%|█████████████████████████████▏ | 2917/9378 [23:01:43<50:59:39, 28.41s/it]
+
+
+ 31%|█████████████████████████████▏ | 2918/9378 [23:02:11<51:01:43, 28.44s/it]
+
+
+
+ 31%|█████████████████████████████▎ | 2919/9378 [23:02:42<50:57:45, 28.40s/it]
+
+
+ 31%|█████████████████████████████▎ | 2920/9378 [23:03:10<50:52:45, 28.36s/it]
+
+ 31%|█████████████████████████████▎ | 2921/9378 [23:03:37<50:55:53, 28.40s/it]
+
+
+ 31%|█████████████████████████████▎ | 2922/9378 [23:04:05<50:58:28, 28.42s/it]
+
+
+ 31%|█████████████████████████████▎ | 2923/9378 [23:04:34<51:02:41, 28.47s/it]
+
+
+ 31%|█████████████████████████████▎ | 2924/9378 [23:05:04<51:23:47, 28.67s/it]
+
+ 31%|█████████████████████████████▎ | 2925/9378 [23:05:32<50:49:33, 28.35s/it]
+
+ 31%|█████████████████████████████▎ | 2926/9378 [23:06:00<51:14:40, 28.59s/it]
+
+ 31%|█████████████████████████████▎ | 2926/9378 [23:06:01<51:14:40, 28.59s/it]
+[2024-05-13 09:38:54,145] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-2926/pytorch_model_fsdp.bin
+[2024-05-13 09:39:05,865] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-2926/pytorch_model_fsdp.bin
+[2024-05-13 09:39:28,729] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-2926/optimizer.bin
+ 31%|█████████████████████████████▎ | 2926/9378 [23:06:01<51:14:40, 28.59s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 31%|█████████████████████████████ | 2927/9378 [23:08:34<119:06:21, 66.47s/it]
+
+
+ 31%|█████████████████████████████▎ | 2928/9378 [23:09:02<98:13:29, 54.82s/it]
+
+
+ 31%|█████████████████████████████▎ | 2929/9378 [23:09:30<84:00:36, 46.90s/it]
+
+ 31%|█████████████████████████████▎ | 2929/9378 [23:09:32<84:00:36, 46.90s/it]
+
+
+ 31%|█████████████████████████████▍ | 2931/9378 [23:10:27<67:01:53, 37.43s/it]
+
+
+ 31%|█████████████████████████████▍ | 2932/9378 [23:10:56<62:11:19, 34.73s/it]
+
+
+ 31%|█████████████████████████████▍ | 2933/9378 [23:11:25<59:13:42, 33.08s/it]
+
+
+ 31%|█████████████████████████████▍ | 2934/9378 [23:11:52<56:17:48, 31.45s/it]
+
+ 31%|█████████████████████████████▍ | 2934/9378 [23:11:54<56:17:48, 31.45s/it]
+
+
+ 31%|█████████████████████████████▍ | 2936/9378 [23:12:49<53:22:07, 29.82s/it]
+
+
+ 31%|█████████████████████████████▍ | 2937/9378 [23:13:18<52:49:45, 29.53s/it]
+
+ 31%|█████████████████████████████▍ | 2937/9378 [23:13:20<52:49:45, 29.53s/it]
+
+
+ 31%|█████████████████████████████▍ | 2939/9378 [23:14:15<51:40:09, 28.89s/it]
+
+ 31%|█████████████████████████████▍ | 2939/9378 [23:14:16<51:40:09, 28.89s/it]
+
+
+ 31%|█████████████████████████████▍ | 2941/9378 [23:15:12<51:20:23, 28.71s/it]
+
+
+ 31%|█████████████████████████████▍ | 2942/9378 [23:15:40<51:03:37, 28.56s/it]
+
+
+ 31%|█████████████████████████████▍ | 2943/9378 [23:16:08<50:58:22, 28.52s/it]
+
+ 31%|█████████████████████████████▍ | 2943/9378 [23:16:10<50:58:22, 28.52s/it]
+
+
+ 31%|█████████████████████████████▌ | 2945/9378 [23:17:05<50:42:36, 28.38s/it]
+
+
+ 31%|█████████████████████████████▌ | 2946/9378 [23:17:34<51:05:37, 28.60s/it]
+
+
+ 31%|█████████████████████████████▌ | 2947/9378 [23:18:02<50:39:18, 28.36s/it]
+
+
+ 31%|█████████████████████████████▌ | 2948/9378 [23:18:31<50:40:20, 28.37s/it]
+
+
+ 31%|█████████████████████████████▌ | 2949/9378 [23:18:59<50:38:57, 28.36s/it]
+
+
+ 31%|█████████████████████████████▌ | 2950/9378 [23:19:47<61:11:47, 34.27s/it]
+
+ 31%|█████████████████████████████▌ | 2950/9378 [23:19:49<61:11:47, 34.27s/it]
+
+
+ 31%|█████████████████████████████▌ | 2952/9378 [23:20:45<56:09:22, 31.46s/it]
+
+
+ 31%|█████████████████████████████▌ | 2953/9378 [23:21:12<54:07:21, 30.33s/it]
+
+
+ 31%|█████████████████████████████▌ | 2954/9378 [23:21:41<53:06:15, 29.76s/it]
+
+ 31%|█████████████████████████████▌ | 2954/9378 [23:21:43<53:06:15, 29.76s/it]
+
+
+ 32%|█████████████████████████████▋ | 2956/9378 [23:22:38<52:00:24, 29.15s/it]
+
+
+ 32%|█████████████████████████████▋ | 2957/9378 [23:23:07<51:36:36, 28.94s/it]
+
+
+ 32%|█████████████████████████████▋ | 2958/9378 [23:23:35<51:23:35, 28.82s/it]
+
+ 32%|█████████████████████████████▋ | 2958/9378 [23:23:37<51:23:35, 28.82s/it]
+
+
+ 32%|█████████████████████████████▋ | 2960/9378 [23:24:34<50:58:07, 28.59s/it]
+
+
+ 32%|█████████████████████████████▋ | 2961/9378 [23:25:02<50:48:19, 28.50s/it]
+
+ 32%|█████████████████████████████▋ | 2962/9378 [23:25:30<50:45:50, 28.48s/it]
+
+ 32%|█████████████████████████████▋ | 2963/9378 [23:25:57<50:45:23, 28.48s/it]
+
+
+ 32%|█████████████████████████████▋ | 2964/9378 [23:26:26<50:45:01, 28.48s/it]
+
+
+ 32%|█████████████████████████████▋ | 2965/9378 [23:26:54<50:40:42, 28.45s/it]
+
+
+ 32%|█████████████████████████████▋ | 2966/9378 [23:27:24<50:49:15, 28.53s/it]
+
+ 32%|█████████████████████████████▋ | 2967/9378 [23:27:52<50:30:44, 28.36s/it]
+
+ 32%|█████████████████████████████▋ | 2968/9378 [23:28:19<50:28:07, 28.34s/it]
+
+
+ 32%|█████████████████████████████▊ | 2969/9378 [23:28:47<50:26:32, 28.33s/it]
+
+
+ 32%|█████████████████████████████▊ | 2970/9378 [23:29:16<50:31:25, 28.38s/it]
+
+
+
+ 32%|█████████████████████████████▊ | 2971/9378 [23:29:46<50:37:42, 28.45s/it]
+
+
+ 32%|█████████████████████████████▊ | 2972/9378 [23:30:14<50:35:40, 28.43s/it]
+
+ 32%|█████████████████████████████▊ | 2973/9378 [23:30:43<50:29:01, 28.37s/it]
+
+ 32%|█████████████████████████████▊ | 2974/9378 [23:31:09<50:28:17, 28.37s/it]
+
+
+ 32%|█████████████████████████████▊ | 2975/9378 [23:31:38<50:31:23, 28.41s/it]
+
+
+ 32%|█████████████████████████████▊ | 2976/9378 [23:32:06<50:32:55, 28.42s/it]
+
+
+
+ 32%|█████████████████████████████▊ | 2977/9378 [23:32:36<50:31:37, 28.42s/it]
+
+
+ 32%|█████████████████████████████▊ | 2978/9378 [23:33:05<50:22:53, 28.34s/it]
+
+ 32%|█████████████████████████████▊ | 2979/9378 [23:33:32<50:50:47, 28.61s/it]
+
+
+ 32%|█████████████████████████████▊ | 2980/9378 [23:34:01<50:44:19, 28.55s/it]
+
+
+ 32%|█████████████████████████████▉ | 2981/9378 [23:34:28<50:14:45, 28.28s/it]
+
+
+
+ 32%|█████████████████████████████▉ | 2982/9378 [23:34:59<50:20:56, 28.34s/it]
+
+
+ 32%|█████████████████████████████▉ | 2983/9378 [23:35:27<50:23:27, 28.37s/it]
+
+ 32%|█████████████████████████████▉ | 2984/9378 [23:35:55<50:41:52, 28.54s/it]
+
+ 32%|█████████████████████████████▉ | 2985/9378 [23:36:22<50:27:24, 28.41s/it]
+
+
+ 32%|█████████████████████████████▉ | 2986/9378 [23:36:51<50:38:27, 28.52s/it]
+
+
+ 32%|█████████████████████████████▉ | 2987/9378 [23:37:19<50:09:13, 28.25s/it]
+
+ 32%|█████████████████████████████▉ | 2987/9378 [23:37:20<50:09:13, 28.25s/it]
+
+
+
+ 32%|█████████████████████████████▉ | 2989/9378 [23:38:17<50:04:20, 28.21s/it]
+
+ 32%|█████████████████████████████▉ | 2990/9378 [23:38:45<50:08:04, 28.25s/it]
+
+ 32%|█████████████████████████████▉ | 2991/9378 [23:39:12<50:16:11, 28.33s/it]
+
+
+ 32%|█████████████████████████████▉ | 2992/9378 [23:39:41<50:39:07, 28.55s/it]
+
+
+ 32%|██████████████████████████████ | 2993/9378 [23:40:09<50:27:26, 28.45s/it]
+
+
+ 32%|██████████████████████████████ | 2994/9378 [23:40:37<50:01:00, 28.21s/it]
+
+
+
+ 32%|██████████████████████████████ | 2995/9378 [23:41:07<50:03:09, 28.23s/it]
+{'loss': 1.2864, 'grad_norm': 0.60546875, 'learning_rate': 4.462135158567193e-05, 'epoch': 0.96}
+
+
+ 32%|██████████████████████████████ | 2997/9378 [23:42:03<50:33:52, 28.53s/it]
+
+
+ 32%|██████████████████████████████ | 2998/9378 [23:42:31<50:08:24, 28.29s/it]
+
+
+ 32%|██████████████████████████████ | 2999/9378 [23:42:59<50:05:53, 28.27s/it]
+
+
+ 32%|██████████████████████████████ | 3000/9378 [23:43:49<61:32:21, 34.74s/it]
+
+
+ 32%|██████████████████████████████ | 3001/9378 [23:44:17<58:08:55, 32.83s/it]
+
+
+ 32%|██████████████████████████████ | 3002/9378 [23:44:45<55:46:56, 31.50s/it]
+
+
+ 32%|██████████████████████████████ | 3003/9378 [23:45:14<54:07:11, 30.56s/it]
+
+
+ 32%|██████████████████████████████ | 3004/9378 [23:45:42<52:51:44, 29.86s/it]
+
+
+ 32%|██████████████████████████████ | 3005/9378 [23:46:11<52:11:31, 29.48s/it]
+
+
+ 32%|██████████████████████████████▏ | 3006/9378 [23:46:39<51:36:20, 29.16s/it]
+
+
+ 32%|██████████████████████████████▏ | 3007/9378 [23:47:07<51:14:43, 28.96s/it]
+
+
+ 32%|██████████████████████████████▏ | 3008/9378 [23:47:36<51:07:10, 28.89s/it]
+
+
+ 32%|██████████████████████████████▏ | 3009/9378 [23:48:04<50:34:40, 28.59s/it]
+
+
+ 32%|██████████████████████████████▏ | 3010/9378 [23:48:34<50:27:37, 28.53s/it]
+
+ 32%|██████████████████████████████▏ | 3011/9378 [23:49:01<50:27:51, 28.53s/it]
+
+
+ 32%|██████████████████████████████▏ | 3012/9378 [23:49:29<50:18:41, 28.45s/it]
+
+
+ 32%|██████████████████████████████▏ | 3013/9378 [23:49:58<50:39:53, 28.66s/it]
+
+ 32%|██████████████████████████████▏ | 3013/9378 [23:49:59<50:39:53, 28.66s/it]
+
+ 32%|██████████████████████████████▏ | 3014/9378 [23:50:28<50:28:43, 28.55s/it]
+
+
+ 32%|██████████████████████████████▏ | 3016/9378 [23:51:24<50:18:14, 28.46s/it]
+
+ 32%|██████████████████████████████▏ | 3017/9378 [23:51:53<50:09:50, 28.39s/it]
+
+ 32%|██████████████████████████████▎ | 3018/9378 [23:52:20<50:11:13, 28.41s/it]
+
+
+ 32%|██████████████████████████████▎ | 3019/9378 [23:52:48<50:04:38, 28.35s/it]
+
+
+ 32%|██████████████████████████████▎ | 3020/9378 [23:53:17<50:07:58, 28.39s/it]
+
+
+ 32%|██████████████████████████████▎ | 3021/9378 [23:53:45<49:52:01, 28.24s/it]
+
+
+ 32%|██████████████████████████████▎ | 3022/9378 [23:54:13<49:39:34, 28.13s/it]
+
+
+ 32%|██████████████████████████████▎ | 3023/9378 [23:54:43<50:05:57, 28.38s/it]
+
+
+ 32%|██████████████████████████████▎ | 3024/9378 [23:55:11<49:48:39, 28.22s/it]
+
+
+ 32%|██████████████████████████████▎ | 3025/9378 [23:55:39<49:48:49, 28.23s/it]
+
+ 32%|██████████████████████████████▎ | 3026/9378 [23:56:06<49:57:25, 28.31s/it]
+
+
+ 32%|██████████████████████████████▎ | 3027/9378 [23:56:34<49:57:18, 28.32s/it]
+
+
+ 32%|██████████████████████████████▎ | 3028/9378 [23:57:03<49:55:45, 28.31s/it]
+
+
+
+ 32%|██████████████████████████████▎ | 3029/9378 [23:57:33<49:56:01, 28.31s/it]
+
+ 32%|██████████████████████████████▎ | 3030/9378 [23:58:02<50:12:43, 28.48s/it]
+
+ 32%|██████████████████████████████▍ | 3031/9378 [23:58:29<50:25:14, 28.60s/it]
+
+
+ 32%|██████████████████████████████▍ | 3032/9378 [23:58:57<49:58:51, 28.35s/it]
+
+
+ 32%|██████████████████████████████▍ | 3033/9378 [23:59:25<49:59:35, 28.36s/it]
+
+
+ 32%|██████████████████████████████▍ | 3034/9378 [23:59:53<49:51:13, 28.29s/it]
+
+
+
+ 32%|██████████████████████████████▍ | 3035/9378 [24:00:23<49:48:36, 28.27s/it]
+
+
+ 32%|██████████████████████████████▍ | 3036/9378 [24:00:52<49:56:00, 28.34s/it]
+
+ 32%|██████████████████████████████▍ | 3037/9378 [24:01:20<49:56:12, 28.35s/it]
+
+ 32%|██████████████████████████████▍ | 3038/9378 [24:01:47<50:20:29, 28.59s/it]
+
+
+ 32%|██████████████████████████████▍ | 3039/9378 [24:02:16<50:12:33, 28.51s/it]
+
+
+ 32%|██████████████████████████████▍ | 3040/9378 [24:02:43<49:43:34, 28.24s/it]
+
+
+
+ 32%|██████████████████████████████▍ | 3041/9378 [24:03:14<49:51:38, 28.33s/it]
+
+ 32%|██████████████████████████████▍ | 3042/9378 [24:03:42<50:13:08, 28.53s/it]
+
+
+ 32%|██████████████████████████████▌ | 3043/9378 [24:04:10<49:40:20, 28.23s/it]
+
+ 32%|██████████████████████████████▌ | 3044/9378 [24:04:37<49:46:05, 28.29s/it]
+
+
+ 32%|██████████████████████████████▌ | 3045/9378 [24:05:06<50:09:47, 28.52s/it]
+
+
+ 32%|██████████████████████████████▌ | 3046/9378 [24:05:34<50:06:58, 28.49s/it]
+
+
+
+ 32%|██████████████████████████████▌ | 3047/9378 [24:06:04<49:58:28, 28.42s/it]
+
+
+ 33%|██████████████████████████████▌ | 3048/9378 [24:06:32<49:31:58, 28.17s/it]
+
+ 33%|██████████████████████████████▌ | 3049/9378 [24:07:00<50:00:56, 28.45s/it]
+
+ 33%|██████████████████████████████▌ | 3050/9378 [24:07:50<61:09:10, 34.79s/it]
+
+ 33%|██████████████████████████████▌ | 3051/9378 [24:08:18<57:45:45, 32.87s/it]
+
+ 33%|██████████████████████████████▌ | 3052/9378 [24:08:45<55:10:53, 31.40s/it]
+
+
+ 33%|██████████████████████████████▌ | 3053/9378 [24:09:14<53:39:01, 30.54s/it]
+
+
+
+ 33%|██████████████████████████████▌ | 3054/9378 [24:09:44<52:29:57, 29.89s/it]
+
+ 33%|██████████████████████████████▌ | 3055/9378 [24:10:12<52:15:25, 29.75s/it]
+
+ 33%|██████████████████████████████▋ | 3056/9378 [24:10:41<51:07:23, 29.11s/it]
+
+ 33%|██████████████████████████████▋ | 3057/9378 [24:11:08<50:47:51, 28.93s/it]
+
+
+ 33%|██████████████████████████████▋ | 3058/9378 [24:11:36<50:27:40, 28.74s/it]
+
+
+ 33%|██████████████████████████████▋ | 3059/9378 [24:12:06<50:43:11, 28.90s/it]
+
+
+ 33%|██████████████████████████████▋ | 3060/9378 [24:12:35<50:07:15, 28.56s/it]
+
+ 33%|██████████████████████████████▋ | 3061/9378 [24:13:02<50:07:07, 28.56s/it]
+
+
+ 33%|██████████████████████████████▋ | 3062/9378 [24:13:30<49:57:00, 28.47s/it]
+
+
+
+ 33%|██████████████████████████████▋ | 3063/9378 [24:14:00<50:24:10, 28.73s/it]
+
+
+ 33%|██████████████████████████████▋ | 3064/9378 [24:14:29<50:01:01, 28.52s/it]
+
+
+ 33%|██████████████████████████████▋ | 3065/9378 [24:14:57<49:43:05, 28.35s/it]
+
+ 33%|██████████████████████████████▋ | 3066/9378 [24:15:25<50:05:54, 28.57s/it]
+
+ 33%|██████████████████████████████▋ | 3067/9378 [24:15:52<49:43:22, 28.36s/it]
+
+
+ 33%|██████████████████████████████▊ | 3068/9378 [24:16:20<49:41:31, 28.35s/it]
+
+
+ 33%|██████████████████████████████▊ | 3069/9378 [24:16:49<49:46:49, 28.41s/it]
+
+
+
+ 33%|██████████████████████████████▊ | 3070/9378 [24:17:19<49:26:46, 28.22s/it]
+
+
+ 33%|██████████████████████████████▊ | 3071/9378 [24:17:47<49:30:18, 28.26s/it]
+
+
+ 33%|██████████████████████████████▊ | 3072/9378 [24:18:15<49:32:08, 28.28s/it]
+
+ 33%|██████████████████████████████▊ | 3073/9378 [24:18:43<49:45:29, 28.41s/it]
+
+ 33%|██████████████████████████████▊ | 3074/9378 [24:19:10<49:29:54, 28.27s/it]
+
+
+ 33%|██████████████████████████████▊ | 3075/9378 [24:19:39<49:42:35, 28.39s/it]
+
+
+
+ 33%|██████████████████████████████▊ | 3076/9378 [24:20:09<50:05:16, 28.61s/it]
+
+ 33%|██████████████████████████████▊ | 3077/9378 [24:20:37<49:54:16, 28.51s/it]
+
+ 33%|██████████████████████████████▊ | 3078/9378 [24:21:06<49:52:06, 28.50s/it]
+
+ 33%|██████████████████████████████▊ | 3079/9378 [24:21:34<49:45:38, 28.44s/it]
+
+ 33%|██████████████████████████████▊ | 3080/9378 [24:22:00<49:17:16, 28.17s/it]
+
+
+ 33%|██████████████████████████████▉ | 3081/9378 [24:22:29<49:41:47, 28.41s/it]
+
+
+
+ 33%|██████████████████████████████▉ | 3082/9378 [24:22:59<49:22:48, 28.24s/it]
+
+
+ 33%|██████████████████████████████▉ | 3083/9378 [24:23:28<49:36:02, 28.37s/it]
+
+
+ 33%|██████████████████████████████▉ | 3084/9378 [24:23:56<49:28:28, 28.30s/it]
+
+ 33%|██████████████████████████████▉ | 3085/9378 [24:24:22<49:30:53, 28.33s/it]
+
+
+ 33%|██████████████████████████████▉ | 3086/9378 [24:24:51<49:31:38, 28.34s/it]
+
+
+ 33%|██████████████████████████████▉ | 3087/9378 [24:25:19<49:34:19, 28.37s/it]
+
+
+
+ 33%|██████████████████████████████▉ | 3088/9378 [24:25:50<49:41:39, 28.44s/it]
+
+
+ 33%|██████████████████████████████▉ | 3089/9378 [24:26:18<49:39:28, 28.43s/it]
+
+ 33%|██████████████████████████████▉ | 3090/9378 [24:26:46<49:31:59, 28.36s/it]
+
+ 33%|██████████████████████████████▉ | 3091/9378 [24:27:13<49:35:29, 28.40s/it]
+
+
+ 33%|██████████████████████████████▉ | 3092/9378 [24:27:42<49:56:11, 28.60s/it]
+
+
+ 33%|███████████████████████████████ | 3093/9378 [24:28:10<49:36:12, 28.41s/it]
+
+
+ 33%|███████████████████████████████ | 3094/9378 [24:28:38<49:18:55, 28.25s/it]
+
+
+
+ 33%|███████████████████████████████ | 3095/9378 [24:29:08<49:23:25, 28.30s/it]
+
+ 33%|███████████████████████████████ | 3096/9378 [24:29:36<49:32:28, 28.39s/it]
+
+ 33%|███████████████████████████████ | 3097/9378 [24:30:03<49:25:05, 28.32s/it]
+
+
+ 33%|███████████████████████████████ | 3098/9378 [24:30:31<49:21:32, 28.29s/it]
+
+
+ 33%|███████████████████████████████ | 3099/9378 [24:31:00<49:26:32, 28.35s/it]
+
+
+ 33%|███████████████████████████████ | 3100/9378 [24:31:47<59:35:07, 34.17s/it]
+
+
+ 33%|███████████████████████████████ | 3101/9378 [24:32:17<56:27:43, 32.38s/it]
+
+ 33%|███████████████████████████████ | 3102/9378 [24:32:44<54:17:47, 31.15s/it]
+
+
+ 33%|███████████████████████████████ | 3103/9378 [24:33:12<52:52:12, 30.33s/it]
+
+
+
+ 33%|███████████████████████████████ | 3104/9378 [24:33:43<51:49:32, 29.74s/it]
+
+ 33%|███████████████████████████████ | 3105/9378 [24:34:11<51:29:58, 29.55s/it]
+
+ 33%|███████████████████████████████▏ | 3106/9378 [24:34:39<50:46:05, 29.14s/it]
+
+
+ 33%|███████████████████████████████▏ | 3107/9378 [24:35:07<49:59:41, 28.70s/it]
+
+ 33%|███████████████████████████████▏ | 3108/9378 [24:35:34<49:49:46, 28.61s/it]
+
+
+ 33%|███████████████████████████████▏ | 3109/9378 [24:36:02<49:36:15, 28.49s/it]
+
+
+
+ 33%|███████████████████████████████▏ | 3110/9378 [24:36:33<49:56:49, 28.69s/it]
+
+
+ 33%|███████████████████████████████▏ | 3111/9378 [24:37:01<49:18:54, 28.33s/it]
+
+
+ 33%|███████████████████████████████▏ | 3112/9378 [24:37:29<49:22:51, 28.37s/it]
+
+
+ 33%|███████████████████████████████▏ | 3113/9378 [24:37:57<49:19:14, 28.34s/it]
+
+ 33%|███████████████████████████████▏ | 3114/9378 [24:38:26<49:23:04, 28.38s/it]
+
+ 33%|███████████████████████████████▏ | 3115/9378 [24:38:52<49:17:59, 28.34s/it]
+
+
+ 33%|███████████████████████████████▏ | 3116/9378 [24:39:22<49:43:08, 28.58s/it]
+
+
+
+ 33%|███████████████████████████████▏ | 3117/9378 [24:39:51<49:12:55, 28.30s/it]
+
+
+ 33%|███████████████████████████████▎ | 3118/9378 [24:40:19<49:09:33, 28.27s/it]
+
+
+ 33%|███████████████████████████████▎ | 3119/9378 [24:40:48<49:10:48, 28.29s/it]
+
+
+ 33%|███████████████████████████████▎ | 3120/9378 [24:41:16<49:15:49, 28.34s/it]
+
+ 33%|███████████████████████████████▎ | 3121/9378 [24:41:44<49:11:46, 28.31s/it]
+
+ 33%|███████████████████████████████▎ | 3122/9378 [24:42:11<49:10:33, 28.30s/it]
+
+
+ 33%|███████████████████████████████▎ | 3123/9378 [24:42:39<49:16:59, 28.36s/it]
+
+
+
+ 33%|███████████████████████████████▎ | 3124/9378 [24:43:09<49:12:45, 28.33s/it]
+
+
+ 33%|███████████████████████████████▎ | 3125/9378 [24:43:37<49:07:54, 28.29s/it]
+
+
+ 33%|███████████████████████████████▎ | 3126/9378 [24:44:06<49:19:42, 28.40s/it]
+
+
+ 33%|███████████████████████████████▎ | 3127/9378 [24:44:34<49:05:59, 28.28s/it]
+
+ 33%|███████████████████████████████▎ | 3128/9378 [24:45:01<49:17:43, 28.39s/it]
+
+
+
+ 33%|███████████████████████████████▎ | 3129/9378 [24:45:31<49:19:00, 28.41s/it]
+
+
+ 33%|███████████████████████████████▎ | 3130/9378 [24:45:59<49:39:13, 28.61s/it]
+
+
+ 33%|███████████████████████████████▍ | 3131/9378 [24:46:28<49:09:16, 28.33s/it]
+
+
+ 33%|███████████████████████████████▍ | 3132/9378 [24:46:56<49:09:16, 28.33s/it]
+
+
+ 33%|███████████████████████████████▍ | 3133/9378 [24:47:24<49:02:53, 28.27s/it]
+
+ 33%|███████████████████████████████▍ | 3134/9378 [24:47:51<49:05:58, 28.31s/it]
+
+
+ 33%|███████████████████████████████▍ | 3135/9378 [24:48:20<49:11:43, 28.37s/it]
+
+ 33%|███████████████████████████████▍ | 3135/9378 [24:48:21<49:11:43, 28.37s/it]
+[2024-05-13 11:21:14,731] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-3135/pytorch_model_fsdp.bin
+[2024-05-13 11:21:26,331] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-3135/pytorch_model_fsdp.bin
+[2024-05-13 11:21:51,886] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-3135/optimizer.bin
+ 33%|███████████████████████████████▍ | 3135/9378 [24:48:21<49:11:43, 28.37s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 33%|███████████████████████████████ | 3136/9378 [24:50:59<117:34:29, 67.81s/it]
+
+
+
+ 33%|███████████████████████████████▍ | 3137/9378 [24:51:30<97:30:06, 56.24s/it]
+
+
+ 33%|███████████████████████████████▍ | 3138/9378 [24:51:58<82:32:57, 47.62s/it]
+
+
+ 33%|███████████████████████████████▍ | 3139/9378 [24:52:26<72:32:13, 41.86s/it]
+
+
+ 33%|███████████████████████████████▍ | 3140/9378 [24:52:55<65:31:49, 37.82s/it]
+
+ 33%|███████████████████████████████▍ | 3141/9378 [24:53:23<60:42:36, 35.04s/it]
+
+ 34%|███████████████████████████████▍ | 3142/9378 [24:53:50<57:09:26, 33.00s/it]
+
+
+
+ 34%|███████████████████████████████▌ | 3143/9378 [24:54:20<55:07:59, 31.83s/it]
+
+
+ 34%|███████████████████████████████▌ | 3144/9378 [24:54:48<53:21:10, 30.81s/it]
+
+
+ 34%|███████████████████████████████▌ | 3145/9378 [24:55:17<51:38:43, 29.83s/it]
+
+
+ 34%|███████████████████████████████▌ | 3146/9378 [24:55:45<50:59:31, 29.46s/it]
+
+ 34%|███████████████████████████████▌ | 3147/9378 [24:56:12<50:26:25, 29.14s/it]
+
+
+ 34%|███████████████████████████████▌ | 3148/9378 [24:56:42<50:12:33, 29.01s/it]
+
+
+ 34%|███████████████████████████████▌ | 3149/9378 [24:57:10<50:02:45, 28.92s/it]
+
+
+ 34%|███████████████████████████████▌ | 3150/9378 [24:58:00<60:43:38, 35.10s/it]
+
+
+ 34%|███████████████████████████████▌ | 3151/9378 [24:58:29<57:10:50, 33.06s/it]
+
+
+ 34%|███████████████████████████████▌ | 3152/9378 [24:58:57<54:42:46, 31.64s/it]
+
+ 34%|███████████████████████████████▌ | 3153/9378 [24:59:26<53:34:15, 30.98s/it]
+
+ 34%|███████████████████████████████▌ | 3154/9378 [24:59:52<51:49:23, 29.97s/it]
+
+
+
+ 34%|███████████████████████████████▌ | 3155/9378 [25:00:23<51:11:20, 29.61s/it]
+
+
+ 34%|███████████████████████████████▋ | 3156/9378 [25:00:51<50:18:45, 29.11s/it]
+
+
+ 34%|███████████████████████████████▋ | 3157/9378 [25:01:19<49:57:03, 28.91s/it]
+
+
+ 34%|███████████████████████████████▋ | 3158/9378 [25:01:48<49:37:23, 28.72s/it]
+
+
+ 34%|███████████████████████████████▋ | 3159/9378 [25:02:16<49:28:12, 28.64s/it]
+
+ 34%|███████████████████████████████▋ | 3160/9378 [25:02:44<49:16:41, 28.53s/it]
+
+
+ 34%|███████████████████████████████▋ | 3161/9378 [25:03:13<49:25:20, 28.62s/it]
+
+
+ 34%|███████████████████████████████▋ | 3162/9378 [25:03:41<49:08:15, 28.46s/it]
+
+
+ 34%|███████████████████████████████▋ | 3163/9378 [25:04:09<49:03:26, 28.42s/it]
+
+
+ 34%|███████████████████████████████▋ | 3164/9378 [25:04:38<48:59:06, 28.38s/it]
+
+
+ 34%|███████████████████████████████▋ | 3165/9378 [25:05:06<49:03:35, 28.43s/it]
+
+
+ 34%|███████████████████████████████▋ | 3166/9378 [25:05:35<49:33:10, 28.72s/it]
+
+
+ 34%|███████████████████████████████▋ | 3167/9378 [25:06:03<48:56:47, 28.37s/it]
+
+
+ 34%|███████████████████████████████▊ | 3168/9378 [25:06:32<48:55:46, 28.36s/it]
+
+
+ 34%|███████████████████████████████▊ | 3169/9378 [25:07:00<49:09:24, 28.50s/it]
+
+
+ 34%|███████████████████████████████▊ | 3170/9378 [25:07:28<48:47:47, 28.30s/it]
+
+ 34%|███████████████████████████████▊ | 3171/9378 [25:07:57<48:51:58, 28.34s/it]
+
+
+ 34%|███████████████████████████████▊ | 3172/9378 [25:08:25<48:48:05, 28.31s/it]
+
+
+ 34%|███████████████████████████████▊ | 3173/9378 [25:08:54<49:17:57, 28.60s/it]
+
+
+ 34%|███████████████████████████████▊ | 3174/9378 [25:09:22<48:44:14, 28.28s/it]
+
+
+ 34%|███████████████████████████████▊ | 3175/9378 [25:09:50<48:51:07, 28.35s/it]
+
+
+ 34%|███████████████████████████████▊ | 3176/9378 [25:10:19<48:48:30, 28.33s/it]
+
+ 34%|███████████████████████████████▊ | 3177/9378 [25:10:47<48:48:40, 28.34s/it]
+
+
+ 34%|███████████████████████████████▊ | 3178/9378 [25:11:15<49:02:34, 28.48s/it]
+
+
+ 34%|███████████████████████████████▊ | 3179/9378 [25:11:44<49:08:49, 28.54s/it]
+{'loss': 1.103, 'grad_norm': 0.54296875, 'learning_rate': 4.89786768076155e-05, 'epoch': 1.02}
+[2024-05-13 11:44:05,069] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:44754] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+
+
+ 34%|███████████████████████████████▊ | 3180/9378 [25:12:16<50:25:12, 29.29s/it]
+
+
+ 34%|███████████████████████████████▉ | 3181/9378 [25:12:44<49:56:45, 29.02s/it]
+
+
+ 34%|███████████████████████████████▉ | 3182/9378 [25:13:12<49:35:24, 28.81s/it]
+
+ 34%|███████████████████████████████▉ | 3183/9378 [25:13:41<49:43:02, 28.89s/it]
+
+
+ 34%|███████████████████████████████▉ | 3184/9378 [25:14:09<49:43:49, 28.90s/it]
+
+
+ 34%|███████████████████████████████▉ | 3185/9378 [25:14:38<49:02:42, 28.51s/it]
+
+
+ 34%|███████████████████████████████▉ | 3186/9378 [25:15:06<48:54:41, 28.44s/it]
+
+
+ 34%|███████████████████████████████▉ | 3187/9378 [25:15:34<48:53:25, 28.43s/it]
+
+
+ 34%|███████████████████████████████▉ | 3188/9378 [25:16:03<48:48:49, 28.39s/it]
+
+
+ 34%|███████████████████████████████▉ | 3189/9378 [25:16:31<48:47:19, 28.38s/it]
+
+ 34%|███████████████████████████████▉ | 3190/9378 [25:17:00<48:46:33, 28.38s/it]
+
+ 34%|███████████████████████████████▉ | 3191/9378 [25:17:26<48:40:38, 28.32s/it]
+
+
+
+ 34%|███████████████████████████████▉ | 3192/9378 [25:17:56<48:47:40, 28.40s/it]
+
+
+ 34%|████████████████████████████████ | 3193/9378 [25:18:25<49:16:14, 28.68s/it]
+
+
+ 34%|████████████████████████████████ | 3194/9378 [25:18:53<48:47:44, 28.41s/it]
+
+
+ 34%|████████████████████████████████ | 3195/9378 [25:19:22<48:41:59, 28.36s/it]
+
+ 34%|████████████████████████████████ | 3196/9378 [25:19:50<48:34:45, 28.29s/it]
+
+ 34%|████████████████████████████████ | 3197/9378 [25:20:16<48:38:10, 28.33s/it]
+
+
+
+ 34%|████████████████████████████████ | 3198/9378 [25:20:47<48:38:00, 28.33s/it]
+
+ 34%|████████████████████████████████ | 3199/9378 [25:21:13<48:45:27, 28.41s/it]
+
+
+ 34%|████████████████████████████████ | 3200/9378 [25:22:01<58:47:40, 34.26s/it]
+
+
+ 34%|████████████████████████████████ | 3201/9378 [25:22:30<55:41:41, 32.46s/it]
+
+
+ 34%|████████████████████████████████ | 3202/9378 [25:22:58<53:32:28, 31.21s/it]
+
+
+ 34%|████████████████████████████████ | 3203/9378 [25:23:26<51:59:32, 30.31s/it]
+
+ 34%|████████████████████████████████ | 3203/9378 [25:23:28<51:59:32, 30.31s/it]
+
+
+ 34%|████████████████████████████████▏ | 3205/9378 [25:24:23<50:23:13, 29.38s/it]
+
+
+ 34%|████████████████████████████████▏ | 3206/9378 [25:24:52<50:05:39, 29.22s/it]
+
+
+ 34%|████████████████████████████████▏ | 3207/9378 [25:25:20<49:20:24, 28.78s/it]
+
+
+ 34%|████████████████████████████████▏ | 3208/9378 [25:25:48<49:18:02, 28.77s/it]
+
+
+ 34%|████████████████████████████████▏ | 3209/9378 [25:26:16<48:52:47, 28.52s/it]
+
+
+ 34%|████████████████████████████████▏ | 3210/9378 [25:26:45<48:47:38, 28.48s/it]
+
+
+ 34%|████████████████████████████████▏ | 3211/9378 [25:27:13<48:40:51, 28.42s/it]
+
+
+ 34%|████████████████████████████████▏ | 3212/9378 [25:27:43<48:42:08, 28.43s/it]
+
+ 34%|████████████████████████████████▏ | 3213/9378 [25:28:10<48:34:20, 28.36s/it]
+
+
+ 34%|████████████████████████████████▏ | 3214/9378 [25:28:38<48:36:34, 28.39s/it]
+
+
+ 34%|████████████████████████████████▏ | 3215/9378 [25:29:07<48:54:25, 28.57s/it]
+
+
+ 34%|████████████████████████████████▏ | 3216/9378 [25:29:35<48:35:32, 28.39s/it]
+
+
+ 34%|████████████████████████████████▏ | 3217/9378 [25:30:03<48:32:12, 28.36s/it]
+
+
+ 34%|████████████████████████████████▎ | 3218/9378 [25:30:33<48:30:48, 28.35s/it]
+
+ 34%|████████████████████████████████▎ | 3219/9378 [25:31:00<48:29:27, 28.34s/it]
+
+
+ 34%|████████████████████████████████▎ | 3220/9378 [25:31:28<48:23:59, 28.29s/it]
+
+
+ 34%|████████████████████████████████▎ | 3221/9378 [25:31:56<48:21:24, 28.27s/it]
+
+
+ 34%|████████████████████████████████▎ | 3222/9378 [25:32:25<48:22:12, 28.29s/it]
+
+
+ 34%|████████████████████████████████▎ | 3223/9378 [25:32:53<48:22:49, 28.30s/it]
+
+
+ 34%|████████████████████████████████▎ | 3224/9378 [25:33:21<48:23:03, 28.30s/it]
+
+
+ 34%|████████████████████████████████▎ | 3225/9378 [25:33:52<48:29:37, 28.37s/it]
+
+ 34%|████████████████████████████████▎ | 3226/9378 [25:34:18<48:26:35, 28.35s/it]
+
+
+ 34%|████████████████████████████████▎ | 3227/9378 [25:34:47<48:56:32, 28.64s/it]
+
+ 34%|████████████████████████████████▎ | 3227/9378 [25:34:48<48:56:32, 28.64s/it]
+
+
+ 34%|████████████████████████████████▎ | 3229/9378 [25:35:44<48:20:01, 28.30s/it]
+
+
+
+ 34%|████████████████████████████████▍ | 3230/9378 [25:36:14<48:21:22, 28.32s/it]
+
+ 34%|████████████████████████████████▍ | 3231/9378 [25:36:40<48:23:29, 28.34s/it]
+
+
+ 34%|████████████████████████████████▍ | 3232/9378 [25:37:09<48:19:44, 28.31s/it]
+
+
+ 34%|████████████████████████████████▍ | 3233/9378 [25:37:37<48:35:56, 28.47s/it]
+
+
+ 34%|████████████████████████████████▍ | 3234/9378 [25:38:05<48:22:03, 28.34s/it]
+
+
+ 34%|████████████████████████████████▍ | 3235/9378 [25:38:34<48:26:09, 28.39s/it]
+
+
+
+ 35%|████████████████████████████████▍ | 3236/9378 [25:39:04<48:21:02, 28.34s/it]
+
+ 35%|████████████████████████████████▍ | 3237/9378 [25:39:31<48:25:23, 28.39s/it]
+
+
+ 35%|████████████████████████████████▍ | 3238/9378 [25:39:59<48:28:44, 28.42s/it]
+
+
+ 35%|████████████████████████████████▍ | 3239/9378 [25:40:28<48:46:46, 28.61s/it]
+
+
+ 35%|████████████████████████████████▍ | 3240/9378 [25:40:56<48:27:29, 28.42s/it]
+
+
+
+ 35%|████████████████████████████████▍ | 3241/9378 [25:41:26<48:18:23, 28.34s/it]
+
+ 35%|████████████████████████████████▍ | 3242/9378 [25:41:53<48:23:52, 28.40s/it]
+
+
+ 35%|████████████████████████████████▌ | 3243/9378 [25:42:21<48:20:42, 28.37s/it]
+
+
+ 35%|████████████████████████████████▌ | 3244/9378 [25:42:50<48:20:29, 28.37s/it]
+
+
+ 35%|████████████████████████████████▌ | 3245/9378 [25:43:18<48:17:42, 28.35s/it]
+
+
+ 35%|████████████████████████████████▌ | 3246/9378 [25:43:46<48:17:46, 28.35s/it]
+
+
+ 35%|████████████████████████████████▌ | 3247/9378 [25:44:16<48:29:57, 28.48s/it]
+
+ 35%|████████████████████████████████▌ | 3248/9378 [25:44:45<48:11:49, 28.31s/it]
+
+ 35%|████████████████████████████████▌ | 3249/9378 [25:45:11<48:14:18, 28.33s/it]
+
+
+ 35%|████████████████████████████████▌ | 3250/9378 [25:46:02<59:29:05, 34.95s/it]
+
+
+ 35%|████████████████████████████████▌ | 3251/9378 [25:46:30<56:04:33, 32.95s/it]
+
+
+ 35%|████████████████████████████████▌ | 3252/9378 [25:46:58<53:48:02, 31.62s/it]
+
+
+ 35%|████████████████████████████████▌ | 3253/9378 [25:47:27<52:04:38, 30.61s/it]
+
+
+ 35%|████████████████████████████████▌ | 3254/9378 [25:47:55<51:02:43, 30.01s/it]
+
+
+ 35%|████████████████████████████████▋ | 3255/9378 [25:48:24<50:15:12, 29.55s/it]
+
+
+ 35%|████████████████████████████████▋ | 3256/9378 [25:48:53<49:59:11, 29.39s/it]
+
+
+ 35%|████████████████████████████████▋ | 3257/9378 [25:49:21<49:14:07, 28.96s/it]
+
+
+ 35%|████████████████████████████████▋ | 3258/9378 [25:49:49<48:53:29, 28.76s/it]
+
+
+ 35%|████████████████████████████████▋ | 3259/9378 [25:50:17<48:35:45, 28.59s/it]
+
+
+ 35%|████████████████████████████████▋ | 3260/9378 [25:50:46<48:27:19, 28.51s/it]
+
+
+ 35%|████████████████████████████████▋ | 3261/9378 [25:51:14<48:18:46, 28.43s/it]
+
+
+ 35%|████████████████████████████████▋ | 3262/9378 [25:51:42<48:16:33, 28.42s/it]
+
+
+ 35%|████████████████████████████████▋ | 3263/9378 [25:52:11<48:11:53, 28.38s/it]
+
+
+ 35%|████████████████████████████████▋ | 3264/9378 [25:52:39<48:12:49, 28.39s/it]
+
+
+ 35%|████████████████████████████████▋ | 3265/9378 [25:53:07<48:14:39, 28.41s/it]
+
+
+ 35%|████████████████████████████████▋ | 3266/9378 [25:53:36<48:16:27, 28.43s/it]
+
+
+ 35%|████████████████████████████████▋ | 3267/9378 [25:54:05<48:22:45, 28.50s/it]
+
+
+ 35%|████████████████████████████████▊ | 3268/9378 [25:54:33<48:13:30, 28.41s/it]
+
+
+ 35%|████████████████████████████████▊ | 3269/9378 [25:55:01<48:11:40, 28.40s/it]
+
+
+ 35%|████████████████████████████████▊ | 3270/9378 [25:55:30<48:10:42, 28.40s/it]
+
+
+ 35%|████████████████████████████████▊ | 3271/9378 [25:55:58<48:06:28, 28.36s/it]
+
+
+ 35%|████████████████████████████████▊ | 3272/9378 [25:56:28<48:04:12, 28.34s/it]
+
+ 35%|████████████████████████████████▊ | 3273/9378 [25:56:54<48:02:35, 28.33s/it]
+
+
+ 35%|████████████████████████████████▊ | 3274/9378 [25:57:23<48:02:10, 28.33s/it]
+
+
+ 35%|████████████████████████████████▊ | 3275/9378 [25:57:51<47:57:26, 28.29s/it]
+
+
+ 35%|████████████████████████████████▊ | 3276/9378 [25:58:19<47:57:25, 28.29s/it]
+
+
+ 35%|████████████████████████████████▊ | 3277/9378 [25:58:47<47:56:32, 28.29s/it]
+
+
+ 35%|████████████████████████████████▊ | 3278/9378 [25:59:18<48:09:57, 28.43s/it]
+
+ 35%|████████████████████████████████▊ | 3279/9378 [25:59:45<48:05:50, 28.39s/it]
+
+
+ 35%|████████████████████████████████▉ | 3280/9378 [26:00:13<48:18:51, 28.52s/it]
+
+
+ 35%|████████████████████████████████▉ | 3281/9378 [26:00:41<48:01:48, 28.36s/it]
+
+
+ 35%|████████████████████████████████▉ | 3282/9378 [26:01:10<48:01:30, 28.36s/it]
+
+ 35%|████████████████████████████████▉ | 3282/9378 [26:01:11<48:01:30, 28.36s/it]
+
+
+
+ 35%|████████████████████████████████▉ | 3284/9378 [26:02:08<47:58:15, 28.34s/it]
+
+ 35%|████████████████████████████████▉ | 3285/9378 [26:02:37<48:04:02, 28.40s/it]
+
+ 35%|████████████████████████████████▉ | 3286/9378 [26:03:03<47:56:30, 28.33s/it]
+
+
+ 35%|████████████████████████████████▉ | 3287/9378 [26:03:32<48:22:35, 28.59s/it]
+
+
+ 35%|████████████████████████████████▉ | 3288/9378 [26:04:01<48:15:12, 28.52s/it]
+
+
+ 35%|████████████████████████████████▉ | 3289/9378 [26:04:31<48:09:45, 28.48s/it]
+
+
+ 35%|████████████████████████████████▉ | 3290/9378 [26:04:59<47:50:03, 28.29s/it]
+
+ 35%|████████████████████████████████▉ | 3291/9378 [26:05:26<48:00:43, 28.40s/it]
+
+
+ 35%|████████████████████████████████▉ | 3292/9378 [26:05:54<47:54:29, 28.34s/it]
+
+
+ 35%|█████████████████████████████████ | 3293/9378 [26:06:22<47:51:47, 28.32s/it]
+
+
+ 35%|█████████████████████████████████ | 3294/9378 [26:06:51<47:53:12, 28.34s/it]
+
+
+
+ 35%|█████████████████████████████████ | 3295/9378 [26:07:21<47:55:16, 28.36s/it]
+
+ 35%|█████████████████████████████████ | 3296/9378 [26:07:49<48:17:00, 28.58s/it]
+
+ 35%|█████████████████████████████████ | 3297/9378 [26:08:16<47:47:27, 28.29s/it]
+
+
+ 35%|█████████████████████████████████ | 3298/9378 [26:08:45<48:13:49, 28.56s/it]
+
+
+ 35%|█████████████████████████████████ | 3299/9378 [26:09:12<47:38:24, 28.21s/it]
+
+
+ 35%|█████████████████████████████████ | 3300/9378 [26:10:00<57:28:20, 34.04s/it]
+
+
+ 35%|█████████████████████████████████ | 3301/9378 [26:10:29<55:10:17, 32.68s/it]
+
+
+
+ 35%|█████████████████████████████████ | 3302/9378 [26:10:59<52:34:40, 31.15s/it]
+
+ 35%|█████████████████████████████████ | 3303/9378 [26:11:27<51:31:27, 30.53s/it]
+
+
+ 35%|█████████████████████████████████ | 3304/9378 [26:11:55<50:01:37, 29.65s/it]
+
+ 35%|█████████████████████████████████▏ | 3305/9378 [26:12:23<49:45:34, 29.50s/it]
+
+
+ 35%|█████████████████████████████████▏ | 3306/9378 [26:12:50<48:50:57, 28.96s/it]
+
+
+ 35%|█████████████████████████████████▏ | 3307/9378 [26:13:19<48:34:25, 28.80s/it]
+
+
+
+ 35%|█████████████████████████████████▏ | 3308/9378 [26:13:49<48:17:39, 28.64s/it]
+
+
+ 35%|█████████████████████████████████▏ | 3309/9378 [26:14:17<48:08:02, 28.55s/it]
+
+ 35%|█████████████████████████████████▏ | 3310/9378 [26:14:46<48:08:58, 28.57s/it]
+
+ 35%|█████████████████████████████████▏ | 3311/9378 [26:15:13<48:11:53, 28.60s/it]
+
+
+ 35%|█████████████████████████████████▏ | 3312/9378 [26:15:41<47:53:38, 28.42s/it]
+
+
+ 35%|█████████████████████████████████▏ | 3313/9378 [26:16:09<47:54:59, 28.44s/it]
+
+
+
+ 35%|█████████████████████████████████▏ | 3314/9378 [26:16:39<47:48:53, 28.39s/it]
+
+
+ 35%|█████████████████████████████████▏ | 3315/9378 [26:17:08<47:52:19, 28.42s/it]
+
+ 35%|█████████████████████████████████▏ | 3316/9378 [26:17:36<48:12:05, 28.63s/it]
+
+ 35%|█████████████████████████████████▏ | 3317/9378 [26:18:03<47:50:09, 28.41s/it]
+
+
+ 35%|█████████████████████████████████▎ | 3318/9378 [26:18:31<47:48:58, 28.41s/it]
+
+
+
+ 35%|█████████████████████████████████▎ | 3319/9378 [26:19:02<47:49:31, 28.42s/it]
+
+
+ 35%|█████████████████████████████████▎ | 3320/9378 [26:19:30<47:49:04, 28.42s/it]
+
+ 35%|█████████████████████████████████▎ | 3321/9378 [26:19:57<47:46:12, 28.39s/it]
+
+
+ 35%|█████████████████████████████████▎ | 3322/9378 [26:20:25<47:45:41, 28.39s/it]
+
+
+ 35%|█████████████████████████████████▎ | 3323/9378 [26:20:54<48:04:30, 28.58s/it]
+
+
+ 35%|█████████████████████████████████▎ | 3324/9378 [26:21:22<47:56:52, 28.51s/it]
+
+
+ 35%|█████████████████████████████████▎ | 3325/9378 [26:21:50<47:28:30, 28.24s/it]
+
+
+ 35%|█████████████████████████████████▎ | 3326/9378 [26:22:20<47:56:13, 28.52s/it]
+
+ 35%|█████████████████████████████████▎ | 3327/9378 [26:22:49<47:29:48, 28.26s/it]
+
+ 35%|█████████████████████████████████▎ | 3328/9378 [26:23:16<47:53:01, 28.49s/it]
+
+
+ 35%|█████████████████████████████████▎ | 3329/9378 [26:23:43<47:24:46, 28.22s/it]
+
+
+ 36%|█████████████████████████████████▍ | 3330/9378 [26:24:12<47:37:06, 28.34s/it]
+
+
+ 36%|█████████████████████████████████▍ | 3331/9378 [26:24:40<47:33:16, 28.31s/it]
+
+
+ 36%|█████████████████████████████████▍ | 3332/9378 [26:25:09<47:32:32, 28.31s/it]
+
+
+
+ 36%|█████████████████████████████████▍ | 3333/9378 [26:25:39<47:31:19, 28.30s/it]
+
+
+ 36%|█████████████████████████████████▍ | 3334/9378 [26:26:07<47:40:14, 28.39s/it]
+
+ 36%|█████████████████████████████████▍ | 3335/9378 [26:26:34<47:42:26, 28.42s/it]
+
+
+ 36%|█████████████████████████████████▍ | 3336/9378 [26:27:02<47:37:49, 28.38s/it]
+
+
+
+ 36%|█████████████████████████████████▍ | 3337/9378 [26:27:33<47:38:41, 28.39s/it]
+
+
+ 36%|█████████████████████████████████▍ | 3338/9378 [26:28:01<47:38:29, 28.40s/it]
+
+
+ 36%|█████████████████████████████████▍ | 3339/9378 [26:28:29<47:45:25, 28.47s/it]
+
+
+ 36%|█████████████████████████████████▍ | 3340/9378 [26:28:57<47:27:28, 28.30s/it]
+
+ 36%|█████████████████████████████████▍ | 3341/9378 [26:29:26<47:31:24, 28.34s/it]
+
+ 36%|█████████████████████████████████▍ | 3342/9378 [26:29:53<47:39:35, 28.43s/it]
+
+
+ 36%|█████████████████████████████████▌ | 3343/9378 [26:30:21<47:42:02, 28.45s/it]
+
+
+ 36%|█████████████████████████████████▌ | 3344/9378 [26:30:51<48:01:36, 28.65s/it]
+{'loss': 1.225, 'grad_norm': 0.6640625, 'learning_rate': 4.996179539063153e-05, 'epoch': 1.05}
+[2024-05-13 13:03:44,790] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-3344/pytorch_model_fsdp.bin
+[2024-05-13 13:03:56,429] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-3344/pytorch_model_fsdp.bin
+[2024-05-13 13:04:21,520] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-3344/optimizer.bin
+ 36%|█████████████████████████████████▌ | 3344/9378 [26:30:51<48:01:36, 28.65s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 36%|█████████████████████████████████▏ | 3345/9378 [26:33:29<113:16:52, 67.60s/it]
+
+
+ 36%|█████████████████████████████████▌ | 3346/9378 [26:33:57<93:31:33, 55.82s/it]
+
+
+
+ 36%|█████████████████████████████████▌ | 3347/9378 [26:34:27<79:42:05, 47.58s/it]
+
+
+ 36%|█████████████████████████████████▌ | 3348/9378 [26:34:56<70:01:37, 41.81s/it]
+
+
+ 36%|█████████████████████████████████▌ | 3349/9378 [26:35:24<63:11:17, 37.73s/it]
+
+
+ 36%|█████████████████████████████████▌ | 3350/9378 [26:36:12<68:14:10, 40.75s/it]
+
+
+ 36%|█████████████████████████████████▌ | 3351/9378 [26:36:40<61:54:26, 36.98s/it]
+
+
+ 36%|█████████████████████████████████▌ | 3352/9378 [26:37:08<57:35:08, 34.40s/it]
+
+ 36%|█████████████████████████████████▌ | 3353/9378 [26:37:35<54:31:50, 32.58s/it]
+
+
+ 36%|█████████████████████████████████▌ | 3354/9378 [26:38:03<52:25:05, 31.33s/it]
+
+
+
+ 36%|█████████████████████████████████▋ | 3355/9378 [26:38:34<51:13:13, 30.61s/it]
+
+
+ 36%|█████████████████████████████████▋ | 3356/9378 [26:39:02<49:54:42, 29.84s/it]
+
+
+ 36%|█████████████████████████████████▋ | 3357/9378 [26:39:30<49:11:32, 29.41s/it]
+
+ 36%|█████████████████████████████████▋ | 3358/9378 [26:39:57<48:37:25, 29.08s/it]
+
+
+ 36%|█████████████████████████████████▋ | 3359/9378 [26:40:25<48:22:21, 28.93s/it]
+
+
+ 36%|█████████████████████████████████▋ | 3360/9378 [26:40:54<48:03:22, 28.75s/it]
+
+
+
+ 36%|█████████████████████████████████▋ | 3361/9378 [26:41:24<48:02:07, 28.74s/it]
+
+
+ 36%|█████████████████████████████████▋ | 3362/9378 [26:41:53<47:49:44, 28.62s/it]
+
+ 36%|█████████████████████████████████▋ | 3363/9378 [26:42:19<47:41:18, 28.54s/it]
+
+
+ 36%|█████████████████████████████████▋ | 3364/9378 [26:42:48<47:40:49, 28.54s/it]
+
+
+
+ 36%|█████████████████████████████████▋ | 3365/9378 [26:43:18<47:40:24, 28.54s/it]
+
+
+ 36%|█████████████████████████████████▋ | 3366/9378 [26:43:46<47:37:42, 28.52s/it]
+
+
+ 36%|█████████████████████████████████▋ | 3367/9378 [26:44:15<47:29:57, 28.45s/it]
+
+ 36%|█████████████████████████████████▊ | 3368/9378 [26:44:41<47:25:55, 28.41s/it]
+
+
+ 36%|█████████████████████████████████▊ | 3369/9378 [26:45:10<47:29:20, 28.45s/it]
+
+
+
+ 36%|█████████████████████████████████▊ | 3370/9378 [26:45:40<47:34:52, 28.51s/it]
+
+
+ 36%|█████████████████████████████████▊ | 3371/9378 [26:46:09<47:31:41, 28.48s/it]
+
+ 36%|█████████████████████████████████▊ | 3372/9378 [26:46:37<47:26:42, 28.44s/it]
+
+ 36%|█████████████████████████████████▊ | 3373/9378 [26:47:04<47:23:26, 28.41s/it]
+
+
+ 36%|█████████████████████████████████▊ | 3374/9378 [26:47:33<47:46:49, 28.65s/it]
+
+
+
+ 36%|█████████████████████████████████▊ | 3375/9378 [26:48:02<47:22:40, 28.41s/it]
+
+
+ 36%|█████████████████████████████████▊ | 3376/9378 [26:48:31<47:24:57, 28.44s/it]
+
+ 36%|█████████████████████████████████▊ | 3377/9378 [26:48:59<47:32:09, 28.52s/it]
+
+ 36%|█████████████████████████████████▊ | 3378/9378 [26:49:27<47:43:41, 28.64s/it]
+
+
+
+ 36%|█████████████████████████████████▊ | 3379/9378 [26:49:56<47:15:18, 28.36s/it]
+
+
+ 36%|█████████████████████████████████▉ | 3380/9378 [26:50:25<47:17:26, 28.38s/it]
+
+ 36%|█████████████████████████████████▉ | 3381/9378 [26:50:53<47:47:57, 28.69s/it]
+
+ 36%|█████████████████████████████████▉ | 3382/9378 [26:51:22<47:11:04, 28.33s/it]
+
+ 36%|█████████████████████████████████▉ | 3383/9378 [26:51:48<47:15:22, 28.38s/it]
+
+
+
+ 36%|█████████████████████████████████▉ | 3384/9378 [26:52:19<47:13:43, 28.37s/it]
+
+
+ 36%|█████████████████████████████████▉ | 3385/9378 [26:52:47<47:16:05, 28.39s/it]
+
+
+ 36%|█████████████████████████████████▉ | 3386/9378 [26:53:15<47:11:31, 28.35s/it]
+
+ 36%|█████████████████████████████████▉ | 3387/9378 [26:53:44<47:47:37, 28.72s/it]
+
+ 36%|█████████████████████████████████▉ | 3388/9378 [26:54:12<47:12:04, 28.37s/it]
+
+
+ 36%|█████████████████████████████████▉ | 3389/9378 [26:54:41<47:17:11, 28.42s/it]
+
+
+ 36%|█████████████████████████████████▉ | 3390/9378 [26:55:09<47:17:33, 28.43s/it]
+
+
+ 36%|█████████████████████████████████▉ | 3391/9378 [26:55:38<47:18:11, 28.44s/it]
+
+ 36%|█████████████████████████████████▉ | 3392/9378 [26:56:06<47:31:55, 28.59s/it]
+
+ 36%|██████████████████████████████████ | 3393/9378 [26:56:34<47:08:19, 28.35s/it]
+
+
+ 36%|██████████████████████████████████ | 3394/9378 [26:57:03<47:08:17, 28.36s/it]
+
+
+ 36%|██████████████████████████████████ | 3395/9378 [26:57:31<47:00:50, 28.29s/it]
+
+
+ 36%|██████████████████████████████████ | 3396/9378 [26:57:59<47:27:11, 28.56s/it]
+
+
+ 36%|██████████████████████████████████ | 3397/9378 [26:58:28<46:54:21, 28.23s/it]
+
+
+ 36%|██████████████████████████████████ | 3398/9378 [26:58:56<47:09:17, 28.39s/it]
+
+ 36%|██████████████████████████████████ | 3399/9378 [26:59:23<47:11:05, 28.41s/it]
+
+
+ 36%|██████████████████████████████████ | 3400/9378 [27:00:13<58:00:21, 34.93s/it]
+
+
+
+ 36%|██████████████████████████████████ | 3401/9378 [27:00:43<54:56:04, 33.09s/it]
+
+
+ 36%|██████████████████████████████████ | 3402/9378 [27:01:12<52:38:05, 31.71s/it]
+
+
+ 36%|██████████████████████████████████ | 3403/9378 [27:01:40<50:36:20, 30.49s/it]
+
+
+ 36%|██████████████████████████████████ | 3404/9378 [27:02:08<49:38:05, 29.91s/it]
+
+
+ 36%|██████████████████████████████████▏ | 3405/9378 [27:02:37<48:40:32, 29.34s/it]
+
+ 36%|██████████████████████████████████▏ | 3406/9378 [27:03:05<48:23:31, 29.17s/it]
+
+
+ 36%|██████████████████████████████████▏ | 3407/9378 [27:03:33<47:53:20, 28.87s/it]
+
+
+ 36%|██████████████████████████████████▏ | 3408/9378 [27:04:02<48:03:57, 28.98s/it]
+
+
+ 36%|██████████████████████████████████▏ | 3409/9378 [27:04:30<47:24:32, 28.59s/it]
+
+
+ 36%|██████████████████████████████████▏ | 3410/9378 [27:04:59<47:17:24, 28.53s/it]
+
+ 36%|██████████████████████████████████▏ | 3411/9378 [27:05:27<47:12:05, 28.48s/it]
+
+ 36%|██████████████████████████████████▏ | 3412/9378 [27:05:54<47:11:00, 28.47s/it]
+
+
+
+ 36%|██████████████████████████████████▏ | 3413/9378 [27:06:24<47:06:31, 28.43s/it]
+
+
+ 36%|██████████████████████████████████▏ | 3414/9378 [27:06:52<47:05:47, 28.43s/it]
+
+
+ 36%|██████████████████████████████████▏ | 3415/9378 [27:07:21<46:55:54, 28.33s/it]
+
+
+ 36%|██████████████████████████████████▏ | 3416/9378 [27:07:49<47:02:00, 28.40s/it]
+
+
+ 36%|██████████████████████████████████▎ | 3417/9378 [27:08:17<46:57:22, 28.36s/it]
+
+ 36%|██████████████████████████████████▎ | 3418/9378 [27:08:44<47:03:06, 28.42s/it]
+
+
+
+ 36%|██████████████████████████████████▎ | 3419/9378 [27:09:14<47:18:01, 28.58s/it]
+
+
+ 36%|██████████████████████████████████▎ | 3420/9378 [27:09:42<46:52:49, 28.33s/it]
+
+ 36%|██████████████████████████████████▎ | 3421/9378 [27:10:11<47:14:57, 28.55s/it]
+
+
+ 36%|██████████████████████████████████▎ | 3422/9378 [27:10:39<46:46:03, 28.27s/it]
+
+
+ 37%|██████████████████████████████████▎ | 3423/9378 [27:11:07<46:46:31, 28.28s/it]
+
+ 37%|██████████████████████████████████▎ | 3424/9378 [27:11:34<46:46:30, 28.28s/it]
+
+
+
+ 37%|██████████████████████████████████▎ | 3425/9378 [27:12:04<47:19:27, 28.62s/it]
+
+
+ 37%|██████████████████████████████████▎ | 3426/9378 [27:12:33<46:52:42, 28.35s/it]
+
+
+ 37%|██████████████████████████████████▎ | 3427/9378 [27:13:01<46:56:02, 28.39s/it]
+
+ 37%|██████████████████████████████████▎ | 3428/9378 [27:13:30<46:57:47, 28.41s/it]
+
+ 37%|██████████████████████████████████▎ | 3429/9378 [27:13:56<46:51:14, 28.35s/it]
+
+
+
+ 37%|██████████████████████████████████▍ | 3430/9378 [27:14:26<46:48:39, 28.33s/it]
+
+
+ 37%|██████████████████████████████████▍ | 3431/9378 [27:14:55<46:52:13, 28.37s/it]
+
+
+ 37%|██████████████████████████████████▍ | 3432/9378 [27:15:23<46:51:37, 28.37s/it]
+
+ 37%|██████████████████████████████████▍ | 3433/9378 [27:15:52<47:14:33, 28.61s/it]
+
+ 37%|██████████████████████████████████▍ | 3434/9378 [27:16:20<46:48:39, 28.35s/it]
+
+ 37%|██████████████████████████████████▍ | 3435/9378 [27:16:47<46:47:11, 28.34s/it]
+
+
+
+ 37%|██████████████████████████████████▍ | 3436/9378 [27:17:17<46:49:22, 28.37s/it]
+
+
+ 37%|██████████████████████████████████▍ | 3437/9378 [27:17:45<46:45:53, 28.34s/it]
+
+
+ 37%|██████████████████████████████████▍ | 3438/9378 [27:18:14<46:48:42, 28.37s/it]
+
+
+ 37%|██████████████████████████████████▍ | 3439/9378 [27:18:42<46:46:07, 28.35s/it]
+
+
+ 37%|██████████████████████████████████▍ | 3440/9378 [27:19:10<46:46:41, 28.36s/it]
+
+
+ 37%|██████████████████████████████████▍ | 3441/9378 [27:19:39<46:48:12, 28.38s/it]
+
+ 37%|██████████████████████████████████▌ | 3442/9378 [27:20:07<46:53:26, 28.44s/it]
+
+ 37%|██████████████████████████████████▌ | 3443/9378 [27:20:33<46:34:08, 28.25s/it]
+
+
+ 37%|██████████████████████████████████▌ | 3444/9378 [27:21:02<46:48:41, 28.40s/it]
+
+
+
+ 37%|██████████████████████████████████▌ | 3445/9378 [27:21:32<46:51:29, 28.43s/it]
+
+
+ 37%|██████████████████████████████████▌ | 3446/9378 [27:22:01<46:50:50, 28.43s/it]
+
+ 37%|██████████████████████████████████▌ | 3447/9378 [27:22:27<46:46:59, 28.40s/it]
+
+
+ 37%|██████████████████████████████████▌ | 3448/9378 [27:22:56<46:56:54, 28.50s/it]
+
+
+
+ 37%|██████████████████████████████████▌ | 3449/9378 [27:23:26<47:01:41, 28.55s/it]
+
+
+ 37%|██████████████████████████████████▌ | 3450/9378 [27:24:16<57:21:25, 34.83s/it]
+
+
+ 37%|██████████████████████████████████▌ | 3451/9378 [27:24:45<54:14:57, 32.95s/it]
+
+ 37%|██████████████████████████████████▌ | 3452/9378 [27:25:13<52:02:14, 31.61s/it]
+
+ 37%|██████████████████████████████████▌ | 3453/9378 [27:25:40<50:47:05, 30.86s/it]
+
+
+ 37%|██████████████████████████████████▌ | 3454/9378 [27:26:09<49:32:36, 30.11s/it]
+
+
+
+ 37%|██████████████████████████████████▋ | 3455/9378 [27:26:38<48:38:05, 29.56s/it]
+
+ 37%|██████████████████████████████████▋ | 3456/9378 [27:27:06<48:02:43, 29.21s/it]
+
+
+ 37%|██████████████████████████████████▋ | 3457/9378 [27:27:35<47:12:11, 28.70s/it]
+
+ 37%|██████████████████████████████████▋ | 3458/9378 [27:28:03<47:30:28, 28.89s/it]
+
+ 37%|██████████████████████████████████▋ | 3459/9378 [27:28:30<46:49:45, 28.48s/it]
+
+
+ 37%|██████████████████████████████████▋ | 3460/9378 [27:28:58<46:46:56, 28.46s/it]
+
+
+
+ 37%|██████████████████████████████████▋ | 3461/9378 [27:29:28<46:55:52, 28.55s/it]
+
+
+ 37%|██████████████████████████████████▋ | 3462/9378 [27:29:57<46:40:05, 28.40s/it]
+
+
+ 37%|██████████████████████████████████▋ | 3463/9378 [27:30:25<46:33:49, 28.34s/it]
+
+
+ 37%|██████████████████████████████████▋ | 3464/9378 [27:30:53<46:28:52, 28.29s/it]
+
+ 37%|██████████████████████████████████▋ | 3465/9378 [27:31:22<46:33:32, 28.35s/it]
+
+ 37%|██████████████████████████████████▋ | 3466/9378 [27:31:48<46:36:37, 28.38s/it]
+
+
+
+ 37%|██████████████████████████████████▊ | 3467/9378 [27:32:19<46:40:40, 28.43s/it]
+
+
+ 37%|██████████████████████████████████▊ | 3468/9378 [27:32:47<46:43:01, 28.46s/it]
+
+
+ 37%|██████████████████████████████████▊ | 3469/9378 [27:33:16<46:39:44, 28.43s/it]
+
+
+ 37%|██████████████████████████████████▊ | 3470/9378 [27:33:44<46:38:45, 28.42s/it]
+
+
+ 37%|██████████████████████████████████▊ | 3471/9378 [27:34:12<46:31:08, 28.35s/it]
+
+ 37%|██████████████████████████████████▊ | 3472/9378 [27:34:40<46:32:01, 28.36s/it]
+
+
+ 37%|██████████████████████████████████▊ | 3473/9378 [27:35:09<46:38:23, 28.43s/it]
+
+
+ 37%|██████████████████████████████████▊ | 3474/9378 [27:35:38<46:37:47, 28.43s/it]
+
+
+ 37%|██████████████████████████████████▊ | 3475/9378 [27:36:06<46:36:54, 28.43s/it]
+
+ 37%|██████████████████████████████████▊ | 3476/9378 [27:36:33<46:39:12, 28.46s/it]
+
+ 37%|██████████████████████████████████▊ | 3476/9378 [27:36:34<46:39:12, 28.46s/it]
+
+
+ 37%|██████████████████████████████████▊ | 3478/9378 [27:37:29<46:32:28, 28.40s/it]
+
+
+ 37%|██████████████████████████████████▊ | 3479/9378 [27:37:59<46:58:26, 28.67s/it]
+
+ 37%|██████████████████████████████████▊ | 3479/9378 [27:38:00<46:58:26, 28.67s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3481/9378 [27:38:55<46:26:13, 28.35s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3482/9378 [27:39:23<46:30:29, 28.40s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3483/9378 [27:39:52<46:27:14, 28.37s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3484/9378 [27:40:20<46:26:38, 28.37s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3485/9378 [27:40:49<46:32:31, 28.43s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3486/9378 [27:41:17<46:22:05, 28.33s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3487/9378 [27:41:45<46:19:30, 28.31s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3488/9378 [27:42:13<46:16:27, 28.28s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3489/9378 [27:42:42<46:30:08, 28.43s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3490/9378 [27:43:10<46:20:21, 28.33s/it]
+
+
+ 37%|██████████████████████████████████▉ | 3491/9378 [27:43:38<46:17:13, 28.31s/it]
+
+
+ 37%|███████████████████████████████████ | 3492/9378 [27:44:07<46:21:58, 28.36s/it]
+
+
+ 37%|███████████████████████████████████ | 3493/9378 [27:44:35<46:26:45, 28.41s/it]
+
+
+ 37%|███████████████████████████████████ | 3494/9378 [27:45:04<46:28:03, 28.43s/it]
+
+
+ 37%|███████████████████████████████████ | 3495/9378 [27:45:32<46:23:14, 28.39s/it]
+
+
+ 37%|███████████████████████████████████ | 3496/9378 [27:46:00<46:20:04, 28.36s/it]
+
+
+ 37%|███████████████████████████████████ | 3497/9378 [27:46:29<46:23:54, 28.40s/it]
+
+
+
+ 37%|███████████████████████████████████ | 3498/9378 [27:46:59<46:33:45, 28.51s/it]
+
+ 37%|███████████████████████████████████ | 3499/9378 [27:47:26<46:34:36, 28.52s/it]
+
+
+ 37%|███████████████████████████████████ | 3500/9378 [27:48:15<55:54:04, 34.24s/it]
+
+ 37%|███████████████████████████████████ | 3501/9378 [27:48:42<53:02:19, 32.49s/it]
+
+
+ 37%|███████████████████████████████████ | 3502/9378 [27:49:10<50:56:40, 31.21s/it]
+
+
+ 37%|███████████████████████████████████ | 3503/9378 [27:49:39<49:37:00, 30.40s/it]
+
+
+ 37%|███████████████████████████████████ | 3504/9378 [27:50:07<48:33:51, 29.76s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3505/9378 [27:50:36<47:50:27, 29.33s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3506/9378 [27:51:04<47:18:41, 29.01s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3507/9378 [27:51:34<47:22:46, 29.05s/it]
+
+ 37%|███████████████████████████████████▏ | 3508/9378 [27:52:00<46:38:21, 28.60s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3509/9378 [27:52:29<46:30:23, 28.53s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3510/9378 [27:52:57<46:23:57, 28.47s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3511/9378 [27:53:25<46:18:20, 28.41s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3512/9378 [27:53:54<46:16:48, 28.40s/it]
+
+ 37%|███████████████████████████████████▏ | 3512/9378 [27:53:56<46:16:48, 28.40s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3514/9378 [27:54:51<46:27:54, 28.53s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3515/9378 [27:55:19<46:23:06, 28.48s/it]
+
+
+ 37%|███████████████████████████████████▏ | 3516/9378 [27:55:48<46:33:27, 28.59s/it]
+
+
+ 38%|███████████████████████████████████▎ | 3517/9378 [27:56:16<46:12:07, 28.38s/it]
+
+
+ 38%|███████████████████████████████████▎ | 3518/9378 [27:56:46<46:15:41, 28.42s/it]
+
+ 38%|███████████████████████████████████▎ | 3519/9378 [27:57:13<46:09:08, 28.36s/it]
+
+
+ 38%|███████████████████████████████████▎ | 3520/9378 [27:57:43<46:08:59, 28.36s/it]
+
+ 38%|███████████████████████████████████▎ | 3521/9378 [27:58:10<46:10:24, 28.38s/it]
+
+
+ 38%|███████████████████████████████████▎ | 3522/9378 [27:58:38<46:12:55, 28.41s/it]
+
+
+ 38%|███████████████████████████████████▎ | 3523/9378 [27:59:07<46:10:36, 28.39s/it]
+
+
+
+ 38%|███████████████████████████████████▎ | 3524/9378 [27:59:37<46:09:59, 28.39s/it]
+
+ 38%|███████████████████████████████████▎ | 3525/9378 [28:00:05<46:11:51, 28.41s/it]
+
+ 38%|███████████████████████████████████▎ | 3526/9378 [28:00:32<46:10:43, 28.41s/it]
+
+
+ 38%|███████████████████████████████████▎ | 3527/9378 [28:01:00<46:14:35, 28.45s/it]
+
+
+
+ 38%|███████████████████████████████████▎ | 3528/9378 [28:01:31<46:20:27, 28.52s/it]
+
+
+ 38%|███████████████████████████████████▎ | 3529/9378 [28:01:59<46:17:10, 28.49s/it]
+
+ 38%|███████████████████████████████████▍ | 3530/9378 [28:02:26<46:15:04, 28.47s/it]
+
+
+ 38%|███████████████████████████████████▍ | 3531/9378 [28:02:55<46:22:36, 28.55s/it]
+
+
+ 38%|███████████████████████████████████▍ | 3532/9378 [28:03:24<46:32:45, 28.66s/it]
+
+
+
+ 38%|███████████████████████████████████▍ | 3533/9378 [28:03:53<46:04:51, 28.38s/it]
+
+ 38%|███████████████████████████████████▍ | 3534/9378 [28:04:21<46:05:34, 28.39s/it]
+
+ 38%|███████████████████████████████████▍ | 3535/9378 [28:04:50<46:02:51, 28.37s/it]
+
+ 38%|███████████████████████████████████▍ | 3536/9378 [28:05:17<46:21:58, 28.57s/it]
+
+
+ 38%|███████████████████████████████████▍ | 3537/9378 [28:05:45<45:59:31, 28.35s/it]
+
+
+
+ 38%|███████████████████████████████████▍ | 3538/9378 [28:06:15<46:03:59, 28.40s/it]
+
+
+ 38%|███████████████████████████████████▍ | 3539/9378 [28:06:44<46:02:11, 28.38s/it]
+
+ 38%|███████████████████████████████████▍ | 3540/9378 [28:07:11<46:26:58, 28.64s/it]
+
+
+ 38%|███████████████████████████████████▍ | 3541/9378 [28:07:38<45:49:48, 28.27s/it]
+
+
+ 38%|███████████████████████████████████▌ | 3542/9378 [28:08:07<45:51:36, 28.29s/it]
+
+
+ 38%|███████████████████████████████████▌ | 3543/9378 [28:08:35<45:49:08, 28.27s/it]
+
+
+ 38%|███████████████████████████████████▌ | 3544/9378 [28:09:03<45:50:34, 28.29s/it]
+
+
+
+ 38%|███████████████████████████████████▌ | 3545/9378 [28:09:33<45:54:06, 28.33s/it]
+
+ 38%|███████████████████████████████████▌ | 3546/9378 [28:10:02<45:58:04, 28.38s/it]
+
+ 38%|███████████████████████████████████▌ | 3547/9378 [28:10:28<45:56:52, 28.37s/it]
+
+
+ 38%|███████████████████████████████████▌ | 3548/9378 [28:10:57<46:02:26, 28.43s/it]
+
+
+ 38%|███████████████████████████████████▌ | 3549/9378 [28:11:25<45:59:38, 28.41s/it]
+
+
+ 38%|███████████████████████████████████▌ | 3550/9378 [28:12:13<55:23:25, 34.22s/it]
+
+
+ 38%|███████████████████████████████████▌ | 3551/9378 [28:12:41<52:28:43, 32.42s/it]
+
+
+ 38%|███████████████████████████████████▌ | 3552/9378 [28:13:10<50:38:50, 31.30s/it]
+
+
+
+ 38%|███████████████████████████████████▌ | 3553/9378 [28:13:40<48:58:36, 30.27s/it]
+{'loss': 1.0623, 'grad_norm': 0.5625, 'learning_rate': 4.707432070775857e-05, 'epoch': 1.12}
+[2024-05-13 14:46:33,101] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-3553/pytorch_model_fsdp.bin
+[2024-05-13 14:46:44,799] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-3553/pytorch_model_fsdp.bin
+[2024-05-13 14:47:06,248] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-3553/optimizer.bin
+ 38%|███████████████████████████████████▌ | 3553/9378 [28:13:40<48:58:36, 30.27s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 38%|███████████████████████████████████▏ | 3554/9378 [28:16:14<109:21:26, 67.60s/it]
+
+ 38%|███████████████████████████████████▋ | 3555/9378 [28:16:41<90:23:09, 55.88s/it]
+
+
+ 38%|███████████████████████████████████▋ | 3556/9378 [28:17:09<76:58:59, 47.60s/it]
+
+
+ 38%|███████████████████████████████████▋ | 3557/9378 [28:17:38<67:43:58, 41.89s/it]
+
+
+
+ 38%|███████████████████████████████████▋ | 3558/9378 [28:18:08<61:08:12, 37.82s/it]
+
+
+ 38%|███████████████████████████████████▋ | 3559/9378 [28:18:37<56:39:49, 35.06s/it]
+
+ 38%|███████████████████████████████████▋ | 3560/9378 [28:19:05<53:37:02, 33.18s/it]
+
+ 38%|███████████████████████████████████▋ | 3561/9378 [28:19:32<51:25:00, 31.82s/it]
+
+
+
+ 38%|███████████████████████████████████▋ | 3562/9378 [28:20:02<49:40:34, 30.75s/it]
+
+ 38%|███████████████████████████████████▋ | 3563/9378 [28:20:31<48:43:04, 30.16s/it]
+
+ 38%|███████████████████████████████████▋ | 3564/9378 [28:20:59<47:44:26, 29.56s/it]
+
+ 38%|███████████████████████████████████▋ | 3565/9378 [28:21:26<47:10:31, 29.22s/it]
+
+
+ 38%|███████████████████████████████████▋ | 3566/9378 [28:21:54<46:47:36, 28.98s/it]
+
+
+
+ 38%|███████████████████████████████████▊ | 3567/9378 [28:22:25<46:29:35, 28.80s/it]
+
+
+ 38%|███████████████████████████████████▊ | 3568/9378 [28:22:53<46:18:17, 28.69s/it]
+
+
+ 38%|███████████████████████████████████▊ | 3569/9378 [28:23:21<46:06:18, 28.57s/it]
+
+ 38%|███████████████████████████████████▊ | 3570/9378 [28:23:49<46:26:17, 28.78s/it]
+
+
+ 38%|███████████████████████████████████▊ | 3571/9378 [28:24:16<45:50:22, 28.42s/it]
+
+
+ 38%|███████████████████████████████████▊ | 3572/9378 [28:24:45<45:48:21, 28.40s/it]
+
+
+ 38%|███████████████████████████████████▊ | 3573/9378 [28:25:15<46:11:48, 28.65s/it]
+
+ 38%|███████████████████████████████████▊ | 3574/9378 [28:25:43<45:48:12, 28.41s/it]
+
+ 38%|███████████████████████████████████▊ | 3575/9378 [28:26:12<45:37:02, 28.30s/it]
+
+ 38%|███████████████████████████████████▊ | 3576/9378 [28:26:38<45:37:19, 28.31s/it]
+
+
+ 38%|███████████████████████████████████▊ | 3577/9378 [28:27:07<45:39:21, 28.33s/it]
+
+
+ 38%|███████████████████████████████████▊ | 3578/9378 [28:27:36<46:02:37, 28.58s/it]
+
+
+ 38%|███████████████████████████████████▊ | 3579/9378 [28:28:04<45:43:09, 28.38s/it]
+
+
+
+ 38%|███████████████████████████████████▉ | 3580/9378 [28:28:33<45:30:09, 28.25s/it]
+
+
+ 38%|███████████████████████████████████▉ | 3581/9378 [28:29:02<45:35:36, 28.31s/it]
+
+ 38%|███████████████████████████████████▉ | 3582/9378 [28:29:29<45:37:13, 28.34s/it]
+
+
+ 38%|███████████████████████████████████▉ | 3583/9378 [28:29:57<45:38:42, 28.36s/it]
+
+
+ 38%|███████████████████████████████████▉ | 3584/9378 [28:30:25<45:41:17, 28.39s/it]
+
+
+
+ 38%|███████████████████████████████████▉ | 3585/9378 [28:30:55<45:36:15, 28.34s/it]
+
+ 38%|███████████████████████████████████▉ | 3586/9378 [28:31:24<45:50:28, 28.49s/it]
+
+ 38%|███████████████████████████████████▉ | 3587/9378 [28:31:51<45:43:53, 28.43s/it]
+
+
+ 38%|███████████████████████████████████▉ | 3588/9378 [28:32:19<45:42:41, 28.42s/it]
+
+
+ 38%|███████████████████████████████████▉ | 3589/9378 [28:32:48<45:44:38, 28.45s/it]
+
+
+
+ 38%|███████████████████████████████████▉ | 3590/9378 [28:33:18<45:35:56, 28.36s/it]
+
+ 38%|███████████████████████████████████▉ | 3591/9378 [28:33:46<45:59:52, 28.61s/it]
+
+
+ 38%|████████████████████████████████████ | 3592/9378 [28:34:14<45:28:05, 28.29s/it]
+
+ 38%|████████████████████████████████████ | 3593/9378 [28:34:41<45:31:07, 28.33s/it]
+
+
+ 38%|████████████████████████████████████ | 3594/9378 [28:35:09<45:35:29, 28.38s/it]
+
+
+ 38%|████████████████████████████████████ | 3595/9378 [28:35:38<45:34:07, 28.37s/it]
+
+
+
+ 38%|████████████████████████████████████ | 3596/9378 [28:36:08<45:35:52, 28.39s/it]
+
+
+ 38%|████████████████████████████████████ | 3597/9378 [28:36:36<45:33:50, 28.37s/it]
+
+ 38%|████████████████████████████████████ | 3598/9378 [28:37:03<45:35:21, 28.39s/it]
+
+
+ 38%|████████████████████████████████████ | 3599/9378 [28:37:31<45:36:11, 28.41s/it]
+
+
+ 38%|████████████████████████████████████ | 3600/9378 [28:38:21<55:59:22, 34.88s/it]
+
+
+ 38%|████████████████████████████████████ | 3601/9378 [28:38:50<52:46:30, 32.89s/it]
+
+
+
+ 38%|████████████████████████████████████ | 3602/9378 [28:39:20<50:47:24, 31.66s/it]
+
+ 38%|████████████████████████████████████ | 3603/9378 [28:39:49<49:21:51, 30.77s/it]
+
+ 38%|████████████████████████████████████ | 3604/9378 [28:40:17<48:02:47, 29.96s/it]
+
+
+ 38%|████████████████████████████████████▏ | 3605/9378 [28:40:45<47:14:42, 29.46s/it]
+
+ 38%|████████████████████████████████████▏ | 3606/9378 [28:41:12<46:39:57, 29.11s/it]
+
+
+ 38%|████████████████████████████████████▏ | 3607/9378 [28:41:41<46:45:45, 29.17s/it]
+
+
+
+ 38%|████████████████████████████████████▏ | 3608/9378 [28:42:10<45:56:08, 28.66s/it]
+
+
+ 38%|████████████████████████████████████▏ | 3609/9378 [28:42:39<45:50:35, 28.61s/it]
+
+
+ 38%|████████████████████████████████████▏ | 3610/9378 [28:43:07<45:49:03, 28.60s/it]
+
+ 39%|████████████████████████████████████▏ | 3611/9378 [28:43:34<45:45:09, 28.56s/it]
+
+
+ 39%|████████████████████████████████████▏ | 3612/9378 [28:44:03<45:40:23, 28.52s/it]
+
+
+
+ 39%|████████████████████████████████████▏ | 3613/9378 [28:44:33<45:35:28, 28.47s/it]
+
+
+ 39%|████████████████████████████████████▏ | 3614/9378 [28:45:01<45:33:11, 28.45s/it]
+
+
+ 39%|████████████████████████████████████▏ | 3615/9378 [28:45:29<45:26:49, 28.39s/it]
+
+
+ 39%|████████████████████████████████████▏ | 3616/9378 [28:45:58<45:23:40, 28.36s/it]
+
+ 39%|████████████████████████████████████▎ | 3617/9378 [28:46:26<45:21:07, 28.34s/it]
+
+ 39%|████████████████████████████████████▎ | 3618/9378 [28:46:53<45:21:52, 28.35s/it]
+
+
+ 39%|████████████████████████████████████▎ | 3619/9378 [28:47:21<45:16:52, 28.31s/it]
+
+
+
+ 39%|████████████████████████████████████▎ | 3620/9378 [28:47:51<45:16:44, 28.31s/it]
+
+
+ 39%|████████████████████████████████████▎ | 3621/9378 [28:48:19<45:16:30, 28.31s/it]
+
+
+ 39%|████████████████████████████████████▎ | 3622/9378 [28:48:48<45:18:51, 28.34s/it]
+
+
+ 39%|████████████████████████████████████▎ | 3623/9378 [28:49:16<45:19:56, 28.36s/it]
+
+ 39%|████████████████████████████████████▎ | 3624/9378 [28:49:44<45:27:29, 28.44s/it]
+
+ 39%|████████████████████████████████████▎ | 3625/9378 [28:50:11<45:20:57, 28.38s/it]
+
+
+
+ 39%|████████████████████████████████████▎ | 3626/9378 [28:50:41<45:20:14, 28.38s/it]
+
+
+ 39%|████████████████████████████████████▎ | 3627/9378 [28:51:10<45:21:26, 28.39s/it]
+
+
+ 39%|████████████████████████████████████▎ | 3628/9378 [28:51:38<45:22:05, 28.40s/it]
+
+ 39%|████████████████████████████████████▍ | 3629/9378 [28:52:06<45:20:25, 28.39s/it]
+
+ 39%|████████████████████████████████████▍ | 3630/9378 [28:52:33<45:18:14, 28.37s/it]
+
+
+
+ 39%|████████████████████████████████████▍ | 3631/9378 [28:53:03<45:25:11, 28.45s/it]
+
+
+ 39%|████████████████████████████████████▍ | 3632/9378 [28:53:32<45:29:41, 28.50s/it]
+
+ 39%|████████████████████████████████████▍ | 3633/9378 [28:54:00<45:37:00, 28.58s/it]
+
+
+ 39%|████████████████████████████████████▍ | 3634/9378 [28:54:28<45:05:42, 28.26s/it]
+
+ 39%|████████████████████████████████████▍ | 3635/9378 [28:54:57<45:32:46, 28.55s/it]
+
+ 39%|████████████████████████████████████▍ | 3636/9378 [28:55:23<44:59:43, 28.21s/it]
+
+
+
+ 39%|████████████████████████████████████▍ | 3637/9378 [28:55:53<45:05:19, 28.27s/it]
+
+
+ 39%|████████████████████████████████████▍ | 3638/9378 [28:56:22<45:07:20, 28.30s/it]
+
+
+ 39%|████████████████████████████████████▍ | 3639/9378 [28:56:50<45:06:33, 28.30s/it]
+
+
+ 39%|████████████████████████████████████▍ | 3640/9378 [28:57:18<45:04:00, 28.27s/it]
+
+
+ 39%|████████████████████████████████████▍ | 3641/9378 [28:57:47<45:09:56, 28.34s/it]
+
+ 39%|████████████████████████████████████▌ | 3642/9378 [28:58:15<45:29:34, 28.55s/it]
+
+ 39%|████████████████████████████████████▌ | 3643/9378 [28:58:42<45:04:13, 28.29s/it]
+
+
+
+ 39%|████████████████████████████████████▌ | 3644/9378 [28:59:12<45:06:04, 28.32s/it]
+
+
+ 39%|████████████████████████████████████▌ | 3645/9378 [28:59:40<45:14:01, 28.40s/it]
+
+ 39%|████████████████████████████████████▌ | 3646/9378 [29:00:09<45:26:20, 28.54s/it]
+
+ 39%|████████████████████████████████████▌ | 3647/9378 [29:00:36<45:20:26, 28.48s/it]
+
+
+ 39%|████████████████████████████████████▌ | 3648/9378 [29:01:04<45:06:59, 28.35s/it]
+
+
+
+ 39%|████████████████████████████████████▌ | 3649/9378 [29:01:34<45:08:40, 28.37s/it]
+
+
+ 39%|████████████████████████████████████▌ | 3650/9378 [29:02:24<55:28:52, 34.87s/it]
+
+
+ 39%|████████████████████████████████████▌ | 3651/9378 [29:02:52<52:23:22, 32.93s/it]
+
+
+ 39%|████████████████████████████████████▌ | 3652/9378 [29:03:21<50:07:08, 31.51s/it]
+
+
+ 39%|████████████████████████████████████▌ | 3653/9378 [29:03:49<48:41:10, 30.61s/it]
+
+
+ 39%|████████████████████████████████████▋ | 3654/9378 [29:04:17<47:28:38, 29.86s/it]
+
+
+ 39%|████████████████████████████████████▋ | 3655/9378 [29:04:46<46:45:21, 29.41s/it]
+
+
+ 39%|████████████████████████████████████▋ | 3656/9378 [29:05:14<46:11:28, 29.06s/it]
+
+ 39%|████████████████████████████████████▋ | 3657/9378 [29:05:40<45:47:57, 28.82s/it]
+
+
+
+ 39%|████████████████████████████████████▋ | 3658/9378 [29:06:11<45:35:23, 28.69s/it]
+
+
+ 39%|████████████████████████████████████▋ | 3659/9378 [29:06:39<45:24:09, 28.58s/it]
+
+
+ 39%|████████████████████████████████████▋ | 3660/9378 [29:07:07<45:19:09, 28.53s/it]
+
+
+ 39%|████████████████████████████████████▋ | 3661/9378 [29:07:36<45:15:29, 28.50s/it]
+
+ 39%|████████████████████████████████████▋ | 3662/9378 [29:08:04<45:29:57, 28.66s/it]
+
+ 39%|████████████████████████████████████▋ | 3663/9378 [29:08:32<45:23:17, 28.59s/it]
+
+
+ 39%|████████████████████████████████████▋ | 3664/9378 [29:09:01<44:50:10, 28.25s/it]
+
+
+ 39%|████████████████████████████████████▋ | 3665/9378 [29:09:29<45:26:32, 28.64s/it]
+
+
+ 39%|████████████████████████████████████▋ | 3666/9378 [29:09:58<44:54:04, 28.30s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3667/9378 [29:10:26<44:59:03, 28.36s/it]
+
+ 39%|████████████████████████████████████▊ | 3668/9378 [29:10:55<44:58:44, 28.36s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3669/9378 [29:11:23<44:59:12, 28.37s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3670/9378 [29:11:51<44:57:57, 28.36s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3671/9378 [29:12:20<44:58:07, 28.37s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3672/9378 [29:12:48<44:56:40, 28.36s/it]
+
+ 39%|████████████████████████████████████▊ | 3673/9378 [29:13:16<45:19:09, 28.60s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3674/9378 [29:13:45<44:45:27, 28.25s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3675/9378 [29:14:13<44:49:36, 28.30s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3676/9378 [29:14:41<44:46:29, 28.27s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3677/9378 [29:15:10<44:50:46, 28.32s/it]
+
+
+ 39%|████████████████████████████████████▊ | 3678/9378 [29:15:38<44:51:21, 28.33s/it]
+
+ 39%|████████████████████████████████████▉ | 3679/9378 [29:16:06<45:09:33, 28.53s/it]
+
+ 39%|████████████████████████████████████▉ | 3680/9378 [29:16:35<44:53:23, 28.36s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3681/9378 [29:17:03<45:14:07, 28.58s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3682/9378 [29:17:32<44:42:44, 28.26s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3683/9378 [29:18:00<44:46:33, 28.30s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3684/9378 [29:18:28<45:02:05, 28.47s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3685/9378 [29:18:56<44:32:25, 28.17s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3686/9378 [29:19:25<44:40:02, 28.25s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3687/9378 [29:19:53<44:38:08, 28.24s/it]
+
+ 39%|████████████████████████████████████▉ | 3688/9378 [29:20:21<44:40:28, 28.27s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3689/9378 [29:20:50<44:43:54, 28.31s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3690/9378 [29:21:18<44:49:08, 28.37s/it]
+
+
+ 39%|████████████████████████████████████▉ | 3691/9378 [29:21:46<44:43:03, 28.31s/it]
+
+ 39%|█████████████████████████████████████ | 3692/9378 [29:22:15<45:05:03, 28.54s/it]
+
+
+ 39%|█████████████████████████████████████ | 3693/9378 [29:22:43<44:46:46, 28.36s/it]
+
+ 39%|█████████████████████████████████████ | 3694/9378 [29:23:10<44:47:46, 28.37s/it]
+
+
+
+ 39%|█████████████████████████████████████ | 3695/9378 [29:23:40<44:49:43, 28.40s/it]
+
+
+ 39%|█████████████████████████████████████ | 3696/9378 [29:24:09<44:47:46, 28.38s/it]
+
+
+ 39%|█████████████████████████████████████ | 3697/9378 [29:24:37<44:52:13, 28.43s/it]
+
+
+ 39%|█████████████████████████████████████ | 3698/9378 [29:25:06<44:48:55, 28.40s/it]
+
+ 39%|█████████████████████████████████████ | 3699/9378 [29:25:32<44:56:02, 28.48s/it]
+
+
+ 39%|█████████████████████████████████████ | 3700/9378 [29:26:22<55:03:16, 34.91s/it]
+
+
+
+ 39%|█████████████████████████████████████ | 3701/9378 [29:26:52<51:58:32, 32.96s/it]
+
+
+ 39%|█████████████████████████████████████ | 3702/9378 [29:27:21<49:49:26, 31.60s/it]
+
+
+ 39%|█████████████████████████████████████ | 3703/9378 [29:27:49<48:16:08, 30.62s/it]
+
+
+ 39%|█████████████████████████████████████▏ | 3704/9378 [29:28:18<47:16:02, 29.99s/it]
+
+ 40%|█████████████████████████████████████▏ | 3705/9378 [29:28:46<46:28:13, 29.49s/it]
+
+
+ 40%|█████████████████████████████████████▏ | 3706/9378 [29:29:15<46:01:27, 29.21s/it]
+
+
+ 40%|█████████████████████████████████████▏ | 3707/9378 [29:29:43<45:42:04, 29.01s/it]
+
+
+ 40%|█████████████████████████████████████▏ | 3708/9378 [29:30:11<45:25:43, 28.84s/it]
+
+
+ 40%|█████████████████████████████████████▏ | 3709/9378 [29:30:40<45:01:16, 28.59s/it]
+
+ 40%|█████████████████████████████████████▏ | 3710/9378 [29:31:06<44:51:10, 28.49s/it]
+
+ 40%|█████████████████████████████████████▏ | 3710/9378 [29:31:08<44:51:10, 28.49s/it]
+
+ 40%|█████████████████████████████████████▏ | 3711/9378 [29:31:36<44:46:28, 28.44s/it]
+
+
+
+ 40%|█████████████████████████████████████▏ | 3713/9378 [29:32:31<44:37:56, 28.36s/it]
+
+
+ 40%|█████████████████████████████████████▏ | 3714/9378 [29:32:59<44:33:12, 28.32s/it]
+
+
+ 40%|█████████████████████████████████████▏ | 3715/9378 [29:33:28<44:34:42, 28.34s/it]
+
+
+ 40%|█████████████████████████████████████▏ | 3716/9378 [29:33:56<44:29:48, 28.29s/it]
+
+
+ 40%|█████████████████████████████████████▎ | 3717/9378 [29:34:24<44:32:41, 28.33s/it]
+
+
+ 40%|█████████████████████████████████████▎ | 3718/9378 [29:34:53<44:30:41, 28.31s/it]
+
+ 40%|█████████████████████████████████████▎ | 3718/9378 [29:34:54<44:30:41, 28.31s/it]
+
+
+ 40%|█████████████████████████████████████▎ | 3720/9378 [29:35:49<44:36:17, 28.38s/it]
+
+
+ 40%|█████████████████████████████████████▎ | 3721/9378 [29:36:18<44:30:56, 28.33s/it]
+
+
+ 40%|█████████████████████████████████████▎ | 3722/9378 [29:36:46<44:32:35, 28.35s/it]
+
+ 40%|█████████████████████████████████████▎ | 3722/9378 [29:36:48<44:32:35, 28.35s/it]
+
+
+ 40%|█████████████████████████████████████▎ | 3724/9378 [29:37:43<44:41:43, 28.46s/it]
+
+
+ 40%|█████████████████████████████████████▎ | 3725/9378 [29:38:14<44:36:40, 28.41s/it]
+
+ 40%|█████████████████████████████████████▎ | 3726/9378 [29:38:40<44:33:29, 28.38s/it]
+
+
+ 40%|█████████████████████████████████████▎ | 3727/9378 [29:39:09<44:53:13, 28.60s/it]
+
+
+ 40%|█████████████████████████████████████▎ | 3728/9378 [29:39:37<44:41:07, 28.47s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3729/9378 [29:40:05<44:30:08, 28.36s/it]
+
+
+
+ 40%|█████████████████████████████████████▍ | 3730/9378 [29:40:36<44:28:09, 28.34s/it]
+
+ 40%|█████████████████████████████████████▍ | 3731/9378 [29:41:03<44:53:42, 28.62s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3732/9378 [29:41:31<44:25:07, 28.32s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3733/9378 [29:41:59<44:25:13, 28.33s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3734/9378 [29:42:28<44:41:13, 28.50s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3735/9378 [29:42:56<44:24:32, 28.33s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3736/9378 [29:43:26<44:46:46, 28.57s/it]
+
+ 40%|█████████████████████████████████████▍ | 3737/9378 [29:43:53<44:22:38, 28.32s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3738/9378 [29:44:21<44:23:31, 28.34s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3739/9378 [29:44:49<44:20:58, 28.31s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3740/9378 [29:45:18<44:27:30, 28.39s/it]
+
+
+ 40%|█████████████████████████████████████▍ | 3741/9378 [29:45:48<44:30:49, 28.43s/it]
+
+ 40%|█████████████████████████████████████▌ | 3742/9378 [29:46:15<44:22:06, 28.34s/it]
+
+
+ 40%|█████████████████████████████████████▌ | 3743/9378 [29:46:44<44:46:42, 28.61s/it]
+
+
+ 40%|█████████████████████████████████████▌ | 3744/9378 [29:47:11<44:17:19, 28.30s/it]
+
+
+ 40%|█████████████████████████████████████▌ | 3745/9378 [29:47:40<44:30:18, 28.44s/it]
+
+
+ 40%|█████████████████████████████████████▌ | 3746/9378 [29:48:10<44:54:36, 28.71s/it]
+
+ 40%|█████████████████████████████████████▌ | 3747/9378 [29:48:38<44:35:55, 28.51s/it]
+
+
+ 40%|█████████████████████████████████████▌ | 3748/9378 [29:49:05<44:19:11, 28.34s/it]
+
+
+ 40%|█████████████████████████████████████▌ | 3749/9378 [29:49:34<44:21:09, 28.37s/it]
+
+
+ 40%|█████████████████████████████████████▌ | 3750/9378 [29:50:22<53:28:52, 34.21s/it]
+
+
+ 40%|█████████████████████████████████████▌ | 3751/9378 [29:50:50<50:44:26, 32.46s/it]
+
+
+
+ 40%|█████████████████████████████████████▌ | 3752/9378 [29:51:20<48:49:57, 31.25s/it]
+
+ 40%|█████████████████████████████████████▌ | 3753/9378 [29:51:47<47:36:47, 30.47s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3754/9378 [29:52:15<46:30:55, 29.78s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3755/9378 [29:52:44<45:54:56, 29.40s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3756/9378 [29:53:12<45:21:31, 29.05s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3757/9378 [29:53:41<45:05:33, 28.88s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3758/9378 [29:54:11<45:07:42, 28.91s/it]
+
+ 40%|█████████████████████████████████████▋ | 3759/9378 [29:54:38<44:53:13, 28.76s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3760/9378 [29:55:06<44:38:15, 28.60s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3761/9378 [29:55:34<44:12:02, 28.33s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3762/9378 [29:56:02<44:18:09, 28.40s/it]
+
+ 40%|█████████████████████████████████████▋ | 3762/9378 [29:56:04<44:18:09, 28.40s/it]
+[2024-05-13 16:28:57,789] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-3762/pytorch_model_fsdp.bin
+[2024-05-13 16:29:09,734] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-3762/pytorch_model_fsdp.bin
+[2024-05-13 16:29:31,090] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-3762/optimizer.bin
+ 40%|█████████████████████████████████████▋ | 3762/9378 [29:56:04<44:18:09, 28.40s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 40%|█████████████████████████████████████▎ | 3763/9378 [29:58:39<104:15:01, 66.84s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3764/9378 [29:59:07<86:12:37, 55.28s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3765/9378 [29:59:35<73:16:24, 47.00s/it]
+
+
+ 40%|█████████████████████████████████████▋ | 3766/9378 [30:00:03<64:35:29, 41.43s/it]
+
+ 40%|█████████████████████████████████████▋ | 3766/9378 [30:00:05<64:35:29, 41.43s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3768/9378 [30:01:00<54:16:11, 34.83s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3769/9378 [30:01:29<51:11:14, 32.85s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3770/9378 [30:01:57<49:11:27, 31.58s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3771/9378 [30:02:25<47:33:26, 30.53s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3772/9378 [30:02:54<46:30:28, 29.87s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3773/9378 [30:03:22<45:48:50, 29.43s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3774/9378 [30:03:51<45:47:35, 29.42s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3775/9378 [30:04:19<44:52:47, 28.84s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3776/9378 [30:04:47<44:42:31, 28.73s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3777/9378 [30:05:16<44:34:48, 28.65s/it]
+
+
+ 40%|█████████████████████████████████████▊ | 3778/9378 [30:05:46<44:25:40, 28.56s/it]
+
+ 40%|█████████████████████████████████████▉ | 3779/9378 [30:06:12<44:16:34, 28.47s/it]
+
+
+ 40%|█████████████████████████████████████▉ | 3780/9378 [30:06:41<44:18:31, 28.49s/it]
+
+
+ 40%|█████████████████████████████████████▉ | 3781/9378 [30:07:10<44:31:05, 28.63s/it]
+
+
+ 40%|█████████████████████████████████████▉ | 3782/9378 [30:07:38<44:13:29, 28.45s/it]
+
+
+ 40%|█████████████████████████████████████▉ | 3783/9378 [30:08:08<44:09:26, 28.41s/it]
+
+ 40%|█████████████████████████████████████▉ | 3784/9378 [30:08:35<44:12:25, 28.45s/it]
+
+
+ 40%|█████████████████████████████████████▉ | 3785/9378 [30:09:03<44:12:06, 28.45s/it]
+
+
+ 40%|█████████████████████████████████████▉ | 3786/9378 [30:09:32<44:14:01, 28.48s/it]
+
+
+ 40%|█████████████████████████████████████▉ | 3787/9378 [30:10:00<44:11:52, 28.46s/it]
+
+
+
+ 40%|█████████████████████████████████████▉ | 3788/9378 [30:10:30<44:06:28, 28.41s/it]
+
+ 40%|█████████████████████████████████████▉ | 3789/9378 [30:10:59<44:10:46, 28.46s/it]
+
+ 40%|█████████████████████████████████████▉ | 3790/9378 [30:11:25<44:02:22, 28.37s/it]
+
+
+ 40%|█████████████████████████████████████▉ | 3791/9378 [30:11:54<44:00:21, 28.36s/it]
+
+
+ 40%|██████████████████████████████████████ | 3792/9378 [30:12:22<44:03:33, 28.39s/it]
+
+
+ 40%|██████████████████████████████████████ | 3793/9378 [30:12:50<44:02:33, 28.39s/it]
+
+
+
+ 40%|██████████████████████████████████████ | 3794/9378 [30:13:20<44:01:54, 28.39s/it]
+
+ 40%|██████████████████████████████████████ | 3795/9378 [30:13:49<44:14:48, 28.53s/it]
+
+ 40%|██████████████████████████████████████ | 3796/9378 [30:14:17<43:51:03, 28.28s/it]
+
+ 40%|██████████████████████████████████████ | 3797/9378 [30:14:44<43:50:55, 28.28s/it]
+
+
+ 40%|██████████████████████████████████████ | 3798/9378 [30:15:13<44:16:21, 28.56s/it]
+
+
+ 41%|██████████████████████████████████████ | 3799/9378 [30:15:41<44:01:42, 28.41s/it]
+
+
+
+ 41%|██████████████████████████████████████ | 3800/9378 [30:16:33<54:01:25, 34.87s/it]
+
+ 41%|██████████████████████████████████████ | 3801/9378 [30:17:01<51:28:33, 33.23s/it]
+
+ 41%|██████████████████████████████████████ | 3802/9378 [30:17:28<48:53:55, 31.57s/it]
+
+
+ 41%|██████████████████████████████████████ | 3803/9378 [30:17:57<47:30:28, 30.68s/it]
+
+
+ 41%|██████████████████████████████████████▏ | 3804/9378 [30:18:25<46:24:30, 29.97s/it]
+
+
+
+ 41%|██████████████████████████████████████▏ | 3805/9378 [30:18:55<45:38:00, 29.48s/it]
+
+
+ 41%|██████████████████████████████████████▏ | 3806/9378 [30:19:23<45:07:22, 29.15s/it]
+
+ 41%|██████████████████████████████████████▏ | 3807/9378 [30:19:52<44:42:29, 28.89s/it]
+
+ 41%|██████████████████████████████████████▏ | 3808/9378 [30:20:18<44:31:10, 28.77s/it]
+
+
+ 41%|██████████████████████████████████████▏ | 3809/9378 [30:20:47<44:22:05, 28.68s/it]
+
+
+
+ 41%|██████████████████████████████████████▏ | 3810/9378 [30:21:17<44:14:46, 28.61s/it]
+
+
+ 41%|██████████████████████████████████████▏ | 3811/9378 [30:21:45<44:00:18, 28.46s/it]
+
+
+ 41%|██████████████████████████████████████▏ | 3812/9378 [30:22:13<43:57:51, 28.44s/it]
+
+ 41%|██████████████████████████████████████▏ | 3813/9378 [30:22:42<43:56:00, 28.42s/it]
+
+ 41%|██████████████████████████████████████▏ | 3814/9378 [30:23:08<43:51:07, 28.37s/it]
+
+
+ 41%|██████████████████████████████████████▏ | 3815/9378 [30:23:37<43:52:43, 28.40s/it]
+
+
+ 41%|██████████████████████████████████████▏ | 3816/9378 [30:24:05<43:51:51, 28.39s/it]
+
+
+
+ 41%|██████████████████████████████████████▎ | 3817/9378 [30:24:35<43:50:29, 28.38s/it]
+
+ 41%|██████████████████████████████████████▎ | 3818/9378 [30:25:04<44:08:48, 28.58s/it]
+
+
+ 41%|██████████████████████████████████████▎ | 3819/9378 [30:25:32<43:37:06, 28.25s/it]
+
+ 41%|██████████████████████████████████████▎ | 3820/9378 [30:26:00<43:41:51, 28.30s/it]
+
+ 41%|██████████████████████████████████████▎ | 3821/9378 [30:26:27<43:43:57, 28.33s/it]
+
+
+ 41%|██████████████████████████████████████▎ | 3822/9378 [30:26:56<44:09:32, 28.61s/it]
+
+
+ 41%|██████████████████████████████████████▎ | 3823/9378 [30:27:26<44:06:25, 28.58s/it]
+
+ 41%|██████████████████████████████████████▎ | 3824/9378 [30:27:53<43:47:32, 28.39s/it]
+
+
+ 41%|██████████████████████████████████████▎ | 3825/9378 [30:28:21<43:46:43, 28.38s/it]
+
+
+ 41%|██████████████████████████████████████▎ | 3826/9378 [30:28:50<44:11:25, 28.65s/it]
+
+
+ 41%|██████████████████████████████████████▎ | 3827/9378 [30:29:18<43:36:06, 28.28s/it]
+
+
+
+ 41%|██████████████████████████████████████▎ | 3828/9378 [30:29:48<43:41:07, 28.34s/it]
+
+ 41%|██████████████████████████████████████▍ | 3829/9378 [30:30:16<44:00:25, 28.55s/it]
+
+ 41%|██████████████████████████████████████▍ | 3830/9378 [30:30:43<43:39:25, 28.33s/it]
+
+
+ 41%|██████████████████████████████████████▍ | 3831/9378 [30:31:11<43:40:41, 28.35s/it]
+
+
+ 41%|██████████████████████████████████████▍ | 3832/9378 [30:31:40<43:37:42, 28.32s/it]
+
+
+ 41%|██████████████████████████████████████▍ | 3833/9378 [30:32:08<43:39:22, 28.34s/it]
+
+
+ 41%|██████████████████████████████████████▍ | 3834/9378 [30:32:36<43:39:26, 28.35s/it]
+
+
+ 41%|██████████████████████████████████████▍ | 3835/9378 [30:33:07<43:58:06, 28.56s/it]
+
+ 41%|██████████████████████████████████████▍ | 3836/9378 [30:33:33<43:34:44, 28.31s/it]
+
+
+ 41%|██████████████████████████████████████▍ | 3837/9378 [30:34:02<43:41:23, 28.39s/it]
+
+
+ 41%|██████████████████████████████████████▍ | 3838/9378 [30:34:30<43:37:13, 28.35s/it]
+
+
+ 41%|██████████████████████████████████████▍ | 3839/9378 [30:34:59<43:48:35, 28.47s/it]
+
+
+
+ 41%|██████████████████████████████████████▍ | 3840/9378 [30:35:29<43:44:22, 28.43s/it]
+
+ 41%|██████████████████████████████████████▌ | 3841/9378 [30:35:57<43:44:52, 28.44s/it]
+
+ 41%|██████████████████████████████████████▌ | 3842/9378 [30:36:24<43:51:02, 28.52s/it]
+
+
+ 41%|██████████████████████████████████████▌ | 3843/9378 [30:36:53<44:05:28, 28.68s/it]
+
+
+
+ 41%|██████████████████████████████████████▌ | 3844/9378 [30:37:23<43:36:12, 28.37s/it]
+
+
+ 41%|██████████████████████████████████████▌ | 3845/9378 [30:37:51<43:45:05, 28.47s/it]
+
+ 41%|██████████████████████████████████████▌ | 3846/9378 [30:38:18<43:42:33, 28.44s/it]
+
+
+ 41%|██████████████████████████████████████▌ | 3847/9378 [30:38:46<43:38:39, 28.41s/it]
+
+
+ 41%|██████████████████████████████████████▌ | 3848/9378 [30:39:15<43:38:36, 28.41s/it]
+
+
+
+ 41%|██████████████████████████████████████▌ | 3849/9378 [30:39:45<43:36:21, 28.39s/it]
+
+ 41%|██████████████████████████████████████▌ | 3850/9378 [30:40:34<53:43:59, 34.99s/it]
+
+
+
+ 41%|██████████████████████████████████████▌ | 3851/9378 [30:41:03<50:29:41, 32.89s/it]
+
+
+ 41%|██████████████████████████████████████▌ | 3852/9378 [30:41:32<48:22:11, 31.51s/it]
+
+ 41%|██████████████████████████████████████▌ | 3853/9378 [30:42:00<46:54:06, 30.56s/it]
+
+ 41%|██████████████████████████████████████▋ | 3854/9378 [30:42:27<45:52:12, 29.89s/it]
+
+
+ 41%|██████████████████████████████████████▋ | 3855/9378 [30:42:55<45:15:27, 29.50s/it]
+
+
+
+ 41%|██████████████████████████████████████▋ | 3856/9378 [30:43:25<44:45:04, 29.18s/it]
+
+
+ 41%|██████████████████████████████████████▋ | 3857/9378 [30:43:54<44:27:31, 28.99s/it]
+
+ 41%|██████████████████████████████████████▋ | 3858/9378 [30:44:21<44:34:38, 29.07s/it]
+
+
+ 41%|██████████████████████████████████████▋ | 3859/9378 [30:44:49<43:51:27, 28.61s/it]
+
+
+ 41%|██████████████████████████████████████▋ | 3860/9378 [30:45:17<43:44:07, 28.53s/it]
+
+
+ 41%|██████████████████████████████████████▋ | 3861/9378 [30:45:45<43:34:45, 28.44s/it]
+
+
+
+ 41%|██████████████████████████████████████▋ | 3862/9378 [30:46:16<43:34:59, 28.44s/it]
+
+
+ 41%|██████████████████████████████████████▋ | 3863/9378 [30:46:44<43:25:48, 28.35s/it]
+
+ 41%|██████████████████████████████████████▋ | 3864/9378 [30:47:12<43:29:49, 28.40s/it]
+
+ 41%|██████████████████████████████████████▋ | 3865/9378 [30:47:39<43:29:04, 28.40s/it]
+
+
+ 41%|██████████████████████████████████████▊ | 3866/9378 [30:48:07<43:30:23, 28.41s/it]
+
+
+
+ 41%|██████████████████████████████████████▊ | 3867/9378 [30:48:38<43:29:43, 28.41s/it]
+
+
+ 41%|██████████████████████████████████████▊ | 3868/9378 [30:49:06<43:33:49, 28.46s/it]
+
+ 41%|██████████████████████████████████████▊ | 3869/9378 [30:49:35<43:35:13, 28.48s/it]
+
+ 41%|██████████████████████████████████████▊ | 3870/9378 [30:50:01<43:30:17, 28.43s/it]
+
+
+ 41%|██████████████████████████████████████▊ | 3871/9378 [30:50:30<43:25:57, 28.39s/it]
+
+
+
+ 41%|██████████████████████████████████████▊ | 3872/9378 [30:51:00<43:50:32, 28.67s/it]
+
+ 41%|██████████████████████████████████████▊ | 3873/9378 [30:51:29<43:35:04, 28.50s/it]
+
+ 41%|██████████████████████████████████████▊ | 3874/9378 [30:51:55<43:30:24, 28.46s/it]
+
+
+ 41%|██████████████████████████████████████▊ | 3875/9378 [30:52:24<43:30:18, 28.46s/it]
+
+
+
+ 41%|██████████████████████████████████████▊ | 3876/9378 [30:52:54<43:25:09, 28.41s/it]
+
+
+ 41%|██████████████████████████████████████▊ | 3877/9378 [30:53:22<43:22:17, 28.38s/it]
+
+
+ 41%|██████████████████████████████████████▊ | 3878/9378 [30:53:51<43:19:36, 28.36s/it]
+
+ 41%|██████████████████████████████████████▉ | 3879/9378 [30:54:19<43:40:27, 28.59s/it]
+
+ 41%|██████████████████████████████████████▉ | 3880/9378 [30:54:47<43:09:29, 28.26s/it]
+
+ 41%|██████████████████████████████████████▉ | 3881/9378 [30:55:14<43:12:45, 28.30s/it]
+
+
+ 41%|██████████████████████████████████████▉ | 3882/9378 [30:55:42<43:12:21, 28.30s/it]
+
+
+ 41%|██████████████████████████████████████▉ | 3883/9378 [30:56:10<43:11:49, 28.30s/it]
+
+
+
+ 41%|██████████████████████████████████████▉ | 3884/9378 [30:56:40<43:20:32, 28.40s/it]
+
+ 41%|██████████████████████████████████████▉ | 3885/9378 [30:57:09<43:25:58, 28.47s/it]
+
+ 41%|██████████████████████████████████████▉ | 3886/9378 [30:57:37<43:21:52, 28.43s/it]
+
+ 41%|██████████████████████████████████████▉ | 3887/9378 [30:58:04<43:06:03, 28.26s/it]
+
+
+ 41%|██████████████████████████████████████▉ | 3888/9378 [30:58:33<43:23:22, 28.45s/it]
+
+
+ 41%|██████████████████████████████████████▉ | 3889/9378 [30:59:00<42:55:58, 28.16s/it]
+
+
+
+ 41%|██████████████████████████████████████▉ | 3890/9378 [30:59:31<43:09:17, 28.31s/it]
+
+
+ 41%|███████████████████████████████████████ | 3891/9378 [30:59:59<43:10:54, 28.33s/it]
+
+
+ 42%|███████████████████████████████████████ | 3892/9378 [31:00:27<43:08:34, 28.31s/it]
+
+
+ 42%|███████████████████████████████████████ | 3893/9378 [31:00:56<43:10:38, 28.34s/it]
+
+ 42%|███████████████████████████████████████ | 3894/9378 [31:01:24<43:15:55, 28.40s/it]
+
+ 42%|███████████████████████████████████████ | 3895/9378 [31:01:51<43:16:41, 28.42s/it]
+
+
+
+ 42%|███████████████████████████████████████ | 3896/9378 [31:02:21<43:11:55, 28.37s/it]
+
+
+ 42%|███████████████████████████████████████ | 3897/9378 [31:02:49<43:10:28, 28.36s/it]
+
+
+ 42%|███████████████████████████████████████ | 3898/9378 [31:03:17<43:06:23, 28.32s/it]
+
+
+ 42%|███████████████████████████████████████ | 3899/9378 [31:03:46<43:05:31, 28.31s/it]
+
+
+ 42%|███████████████████████████████████████ | 3900/9378 [31:04:36<53:13:58, 34.98s/it]
+
+
+ 42%|███████████████████████████████████████ | 3901/9378 [31:05:04<50:00:03, 32.87s/it]
+
+ 42%|███████████████████████████████████████ | 3902/9378 [31:05:32<47:56:55, 31.52s/it]
+
+
+ 42%|███████████████████████████████████████ | 3903/9378 [31:06:01<46:27:54, 30.55s/it]
+
+
+ 42%|███████████████████████████████████████▏ | 3904/9378 [31:06:29<45:32:04, 29.95s/it]
+
+
+ 42%|███████████████████████████████████████▏ | 3905/9378 [31:06:58<44:44:58, 29.44s/it]
+
+
+ 42%|███████████████████████████████████████▏ | 3906/9378 [31:07:26<44:28:33, 29.26s/it]
+
+ 42%|███████████████████████████████████████▏ | 3907/9378 [31:07:54<44:11:55, 29.08s/it]
+
+ 42%|███████████████████████████████████████▏ | 3908/9378 [31:08:21<43:33:42, 28.67s/it]
+
+
+ 42%|███████████████████████████████████████▏ | 3909/9378 [31:08:49<43:24:26, 28.57s/it]
+
+
+ 42%|███████████████████████████████████████▏ | 3910/9378 [31:09:18<43:19:25, 28.52s/it]
+
+
+
+ 42%|███████████████████████████████████████▏ | 3911/9378 [31:09:48<43:17:11, 28.50s/it]
+
+
+ 42%|███████████████████████████████████████▏ | 3912/9378 [31:10:16<43:13:06, 28.46s/it]
+
+
+ 42%|███████████████████████████████████████▏ | 3913/9378 [31:10:45<43:05:57, 28.39s/it]
+
+ 42%|███████████████████████████████████████▏ | 3914/9378 [31:11:11<43:05:12, 28.39s/it]
+
+
+ 42%|███████████████████████████████████████▏ | 3915/9378 [31:11:40<43:05:05, 28.39s/it]
+
+
+ 42%|███████████████████████████████████████▎ | 3916/9378 [31:12:08<42:59:26, 28.34s/it]
+
+
+
+ 42%|███████████████████████████████████████▎ | 3917/9378 [31:12:38<42:59:00, 28.34s/it]
+
+
+ 42%|███████████████████████████████████████▎ | 3918/9378 [31:13:06<42:53:55, 28.28s/it]
+
+
+ 42%|███████████████████████████████████████▎ | 3919/9378 [31:13:35<42:55:03, 28.30s/it]
+
+
+ 42%|███████████████████████████████████████▎ | 3920/9378 [31:14:03<43:02:05, 28.39s/it]
+
+ 42%|███████████████████████████████████████▎ | 3921/9378 [31:14:30<43:00:11, 28.37s/it]
+
+
+ 42%|███████████████████████████████████████▎ | 3922/9378 [31:14:58<43:04:33, 28.42s/it]
+
+
+
+ 42%|███████████████████████████████████████▎ | 3923/9378 [31:15:28<43:03:36, 28.42s/it]
+
+
+ 42%|███████████████████████████████████████▎ | 3924/9378 [31:15:57<43:03:00, 28.42s/it]
+
+ 42%|███████████████████████████████████████▎ | 3925/9378 [31:16:25<43:07:58, 28.48s/it]
+
+ 42%|███████████████████████████████████████▎ | 3926/9378 [31:16:52<43:16:56, 28.58s/it]
+
+
+
+ 42%|███████████████████████████████████████▎ | 3927/9378 [31:17:22<43:21:06, 28.63s/it]
+
+
+ 42%|███████████████████████████████████████▎ | 3928/9378 [31:17:51<42:59:47, 28.40s/it]
+
+
+ 42%|███████████████████████████████████████▍ | 3929/9378 [31:18:19<42:56:32, 28.37s/it]
+
+
+ 42%|███████████████████████████████████████▍ | 3930/9378 [31:18:47<42:51:25, 28.32s/it]
+
+ 42%|███████████████████████████████████████▍ | 3931/9378 [31:19:14<42:54:20, 28.36s/it]
+
+
+ 42%|███████████████████████████████████████▍ | 3932/9378 [31:19:42<42:52:58, 28.35s/it]
+
+
+
+ 42%|███████████████████████████████████████▍ | 3933/9378 [31:20:12<42:54:08, 28.37s/it]
+
+
+ 42%|███████████████████████████████████████▍ | 3934/9378 [31:20:41<42:50:56, 28.34s/it]
+
+
+ 42%|███████████████████████████████████████▍ | 3935/9378 [31:21:09<42:50:16, 28.33s/it]
+
+
+ 42%|███████████████████████████████████████▍ | 3936/9378 [31:21:37<42:51:29, 28.35s/it]
+
+
+ 42%|███████████████████████████████████████▍ | 3937/9378 [31:22:06<42:50:13, 28.34s/it]
+
+ 42%|███████████████████████████████████████▍ | 3938/9378 [31:22:34<43:19:48, 28.67s/it]
+
+
+
+ 42%|███████████████████████████████████████▍ | 3939/9378 [31:23:03<43:10:44, 28.58s/it]
+
+
+ 42%|███████████████████████████████████████▍ | 3940/9378 [31:23:31<42:44:13, 28.29s/it]
+
+
+ 42%|███████████████████████████████████████▌ | 3941/9378 [31:24:00<42:45:43, 28.31s/it]
+
+
+ 42%|███████████████████████████████████████▌ | 3942/9378 [31:24:28<42:43:59, 28.30s/it]
+
+ 42%|███████████████████████████████████████▌ | 3943/9378 [31:24:56<43:06:21, 28.55s/it]
+
+ 42%|███████████████████████████████████████▌ | 3944/9378 [31:25:23<42:41:10, 28.28s/it]
+
+
+
+ 42%|███████████████████████████████████████▌ | 3945/9378 [31:25:53<43:05:26, 28.55s/it]
+
+
+ 42%|███████████████████████████████████████▌ | 3946/9378 [31:26:22<42:38:47, 28.26s/it]
+
+
+ 42%|███████████████████████████████████████▌ | 3947/9378 [31:26:50<42:48:02, 28.37s/it]
+
+ 42%|███████████████████████████████████████▌ | 3948/9378 [31:27:18<42:43:49, 28.33s/it]
+
+ 42%|███████████████████████████████████████▌ | 3949/9378 [31:27:45<42:43:52, 28.34s/it]
+
+
+ 42%|███████████████████████████████████████▌ | 3950/9378 [31:28:35<52:33:59, 34.86s/it]
+
+
+ 42%|███████████████████████████████████████▌ | 3951/9378 [31:29:03<49:36:46, 32.91s/it]
+
+
+
+ 42%|███████████████████████████████████████▌ | 3952/9378 [31:29:33<47:30:05, 31.52s/it]
+
+
+ 42%|███████████████████████████████████████▌ | 3953/9378 [31:30:02<46:06:54, 30.60s/it]
+
+ 42%|███████████████████████████████████████▋ | 3954/9378 [31:30:30<45:23:17, 30.12s/it]
+
+
+ 42%|███████████████████████████████████████▋ | 3955/9378 [31:30:59<44:17:08, 29.40s/it]
+
+ 42%|███████████████████████████████████████▋ | 3956/9378 [31:31:25<43:48:26, 29.09s/it]
+
+
+
+ 42%|███████████████████████████████████████▋ | 3957/9378 [31:31:55<43:30:50, 28.90s/it]
+
+
+ 42%|███████████████████████████████████████▋ | 3958/9378 [31:32:24<43:34:47, 28.95s/it]
+
+
+ 42%|███████████████████████████████████████▋ | 3959/9378 [31:32:52<42:55:54, 28.52s/it]
+
+
+ 42%|███████████████████████████████████████▋ | 3960/9378 [31:33:20<42:55:00, 28.52s/it]
+
+
+ 42%|███████████████████████████████████████▋ | 3961/9378 [31:33:49<42:48:40, 28.45s/it]
+
+ 42%|███████████████████████████████████████▋ | 3962/9378 [31:34:16<43:15:00, 28.75s/it]
+
+
+
+ 42%|███████████████████████████████████████▋ | 3963/9378 [31:34:46<42:41:53, 28.39s/it]
+
+
+ 42%|███████████████████████████████████████▋ | 3964/9378 [31:35:14<42:40:55, 28.38s/it]
+
+
+ 42%|███████████████████████████████████████▋ | 3965/9378 [31:35:43<42:41:10, 28.39s/it]
+
+ 42%|███████████████████████████████████████▊ | 3966/9378 [31:36:11<43:00:20, 28.61s/it]
+
+ 42%|███████████████████████████████████████▊ | 3967/9378 [31:36:38<42:59:44, 28.61s/it]
+
+
+
+ 42%|███████████████████████████████████████▊ | 3968/9378 [31:37:08<42:33:39, 28.32s/it]
+
+
+ 42%|███████████████████████████████████████▊ | 3969/9378 [31:37:36<42:32:12, 28.31s/it]
+
+
+ 42%|███████████████████████████████████████▊ | 3970/9378 [31:38:05<42:40:23, 28.41s/it]
+
+
+ 42%|███████████████████████████████████████▊ | 3971/9378 [31:38:33<42:36:05, 28.36s/it]
+{'loss': 1.0856, 'grad_norm': 0.73828125, 'learning_rate': 3.0345017509794794e-05, 'epoch': 1.25}
+[2024-05-13 18:11:26,374] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-3971/pytorch_model_fsdp.bin
+[2024-05-13 18:11:38,118] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-3971/pytorch_model_fsdp.bin
+[2024-05-13 18:11:59,918] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-3971/optimizer.bin
+ 42%|███████████████████████████████████████▊ | 3971/9378 [31:38:33<42:36:05, 28.36s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+
+ 42%|███████████████████████████████████████▍ | 3972/9378 [31:41:10<100:19:37, 66.81s/it]
+
+
+ 42%|███████████████████████████████████████▊ | 3973/9378 [31:41:38<82:56:35, 55.24s/it]
+
+
+ 42%|███████████████████████████████████████▊ | 3974/9378 [31:42:06<70:42:41, 47.11s/it]
+
+ 42%|███████████████████████████████████████▊ | 3975/9378 [31:42:34<62:24:52, 41.59s/it]
+
+
+ 42%|███████████████████████████████████████▊ | 3976/9378 [31:43:03<56:13:09, 37.47s/it]
+
+
+ 42%|███████████████████████████████████████▊ | 3977/9378 [31:43:31<52:08:52, 34.76s/it]
+
+
+ 42%|███████████████████████████████████████▊ | 3978/9378 [31:43:59<49:09:02, 32.77s/it]
+
+ 42%|███████████████████████████████████████▉ | 3979/9378 [31:44:28<47:37:34, 31.76s/it]
+
+ 42%|███████████████████████████████████████▉ | 3980/9378 [31:44:56<46:09:28, 30.78s/it]
+
+
+ 42%|███████████████████████████████████████▉ | 3981/9378 [31:45:25<44:45:25, 29.85s/it]
+
+
+ 42%|███████████████████████████████████████▉ | 3982/9378 [31:45:53<44:27:51, 29.66s/it]
+
+
+ 42%|███████████████████████████████████████▉ | 3983/9378 [31:46:22<43:44:36, 29.19s/it]
+
+
+ 42%|███████████████████████████████████████▉ | 3984/9378 [31:46:50<43:05:56, 28.76s/it]
+
+
+ 42%|███████████████████████████████████████▉ | 3985/9378 [31:47:18<42:54:16, 28.64s/it]
+
+
+ 43%|███████████████████████████████████████▉ | 3986/9378 [31:47:46<42:47:25, 28.57s/it]
+
+ 43%|███████████████████████████████████████▉ | 3987/9378 [31:48:15<42:38:15, 28.47s/it]
+
+
+ 43%|███████████████████████████████████████▉ | 3988/9378 [31:48:43<42:37:33, 28.47s/it]
+
+
+ 43%|███████████████████████████████████████▉ | 3989/9378 [31:49:12<42:33:29, 28.43s/it]
+
+
+ 43%|███████████████████████████████████████▉ | 3990/9378 [31:49:40<42:30:42, 28.40s/it]
+
+ 43%|████████████████████████████████████████ | 3991/9378 [31:50:08<42:46:23, 28.58s/it]
+
+
+ 43%|████████████████████████████████████████ | 3992/9378 [31:50:37<42:16:34, 28.26s/it]
+
+ 43%|████████████████████████████████████████ | 3993/9378 [31:51:05<42:43:08, 28.56s/it]
+
+ 43%|████████████████████████████████████████ | 3994/9378 [31:51:32<42:17:19, 28.28s/it]
+
+
+
+ 43%|████████████████████████████████████████ | 3995/9378 [31:52:02<42:16:59, 28.28s/it]
+
+ 43%|████████████████████████████████████████ | 3996/9378 [31:52:28<42:15:39, 28.27s/it]
+
+
+ 43%|████████████████████████████████████████ | 3997/9378 [31:52:56<42:14:38, 28.26s/it]
+
+
+ 43%|████████████████████████████████████████ | 3998/9378 [31:53:25<42:19:48, 28.33s/it]
+
+
+ 43%|████████████████████████████████████████ | 3999/9378 [31:53:53<42:19:46, 28.33s/it]
+
+
+ 43%|████████████████████████████████████████ | 4000/9378 [31:54:41<51:01:39, 34.16s/it]
+
+
+ 43%|████████████████████████████████████████ | 4001/9378 [31:55:09<48:22:16, 32.39s/it]
+
+
+ 43%|████████████████████████████████████████ | 4002/9378 [31:55:37<46:31:17, 31.15s/it]
+
+ 43%|████████████████████████████████████████ | 4002/9378 [31:55:39<46:31:17, 31.15s/it]
+
+
+
+ 43%|████████████████████████████████████████▏ | 4004/9378 [31:56:35<44:49:19, 30.03s/it]
+
+
+ 43%|████████████████████████████████████████▏ | 4005/9378 [31:57:03<43:44:18, 29.31s/it]
+
+
+ 43%|████████████████████████████████████████▏ | 4006/9378 [31:57:31<43:18:23, 29.02s/it]
+
+ 43%|████████████████████████████████████████▏ | 4006/9378 [31:57:33<43:18:23, 29.02s/it]
+
+
+ 43%|████████████████████████████████████████▏ | 4008/9378 [31:58:28<42:42:15, 28.63s/it]
+
+ 43%|████████████████████████████████████████▏ | 4008/9378 [31:58:30<42:42:15, 28.63s/it]
+[2024-05-13 18:31:00,114] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: be50e76b-8d8e-4d21-b6c9-7239244c72ed)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/97a72440d880f503e7284a1df4316d48c698700bfbadf37f2f1ceb978f479244?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T181258Z&X-Amz-Expires=86400&X-Amz-Signature=6aa00794b742b549aa29c3b2f8debc8096e0bf5195eb3e0386c11b12896feb9c&X-Amz-SignedHeaders=host&partNumber=1&uploadId=MUm8Bo8woGwNyIZvcnHaD2i6lGhdeqCrLjcB0YolwyLP5TSvDQMZf3..7FWBnzaM8iURmqXB5nIzlFlrODfVdlwQUtYXhx9UdVScZ_T8GFPI1a3sj7hdFW0ULerrwbKr&x-id=UploadPart
+ 43%|████████████████████████████████████████▏ | 4008/9378 [31:58:30<42:42:15, 28.63s/it]'(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: be50e76b-8d8e-4d21-b6c9-7239244c72ed)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/97a72440d880f503e7284a1df4316d48c698700bfbadf37f2f1ceb978f479244?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T181258Z&X-Amz-Expires=86400&X-Amz-Signature=6aa00794b742b549aa29c3b2f8debc8096e0bf5195eb3e0386c11b12896feb9c&X-Amz-SignedHeaders=host&partNumber=1&uploadId=MUm8Bo8woGwNyIZvcnHaD2i6lGhdeqCrLjcB0YolwyLP5TSvDQMZf3..7FWBnzaM8iURmqXB5nIzlFlrODfVdlwQUtYXhx9UdVScZ_T8GFPI1a3sj7hdFW0ULerrwbKr&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+{'loss': 1.0528, 'grad_norm': 0.52734375, 'learning_rate': 2.840621289169701e-05, 'epoch': 1.27}
+
+ 43%|████████████████████████████████████████▏ | 4010/9378 [31:59:25<42:43:06, 28.65s/it]
+
+
+ 43%|████████████████████████████████████████▏ | 4011/9378 [31:59:53<42:14:52, 28.34s/it]
+
+
+ 43%|████████████████████████████████████████▏ | 4012/9378 [32:00:21<42:13:47, 28.33s/it]
+
+
+ 43%|████████████████████████████████████████▏ | 4013/9378 [32:00:50<42:24:20, 28.45s/it]
+
+
+ 43%|████████████████████████████████████████▏ | 4014/9378 [32:01:18<42:15:48, 28.36s/it]
+
+
+ 43%|████████████████████████████████████████▏ | 4015/9378 [32:01:48<42:42:29, 28.67s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4016/9378 [32:02:16<42:40:21, 28.65s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4017/9378 [32:02:44<42:21:12, 28.44s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4018/9378 [32:03:12<42:03:11, 28.24s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4019/9378 [32:03:40<42:07:24, 28.30s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4020/9378 [32:04:11<42:18:55, 28.43s/it]
+
+ 43%|████████████████████████████████████████▎ | 4021/9378 [32:04:37<42:07:06, 28.30s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4022/9378 [32:05:06<42:11:25, 28.36s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4023/9378 [32:05:34<42:07:31, 28.32s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4024/9378 [32:06:02<42:10:42, 28.36s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4025/9378 [32:06:31<42:15:07, 28.42s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4026/9378 [32:06:59<42:12:55, 28.40s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4027/9378 [32:07:29<42:39:55, 28.70s/it]
+
+
+ 43%|████████████████████████████████████████▎ | 4028/9378 [32:07:56<42:09:49, 28.37s/it]
+
+ 43%|████████████████████████████████████████▎ | 4028/9378 [32:07:58<42:09:49, 28.37s/it]
+
+ 43%|████████████████████████████████████████▍ | 4029/9378 [32:08:26<42:32:50, 28.64s/it]
+
+
+ 43%|████████████████████████████████████████▍ | 4031/9378 [32:09:21<42:00:27, 28.28s/it]
+
+
+ 43%|████████████████████████████████████████▍ | 4032/9378 [32:09:50<42:13:04, 28.43s/it]
+
+
+ 43%|████████████████████████████████████████▍ | 4033/9378 [32:10:19<42:14:42, 28.45s/it]
+
+
+ 43%|████████████████████████████████████████▍ | 4034/9378 [32:10:47<42:15:53, 28.47s/it]
+
+
+ 43%|████████████████████████████████████████▍ | 4035/9378 [32:11:17<42:09:39, 28.41s/it]
+
+ 43%|████████████████████████████████████████▍ | 4036/9378 [32:11:45<42:30:38, 28.65s/it]
+
+
+ 43%|████████████████████████████████████████▍ | 4037/9378 [32:12:13<42:07:40, 28.40s/it]
+
+
+ 43%|████████████████████████████████████████▍ | 4038/9378 [32:12:41<42:07:24, 28.40s/it]
+
+
+
+ 43%|████████████████████████████████████████▍ | 4039/9378 [32:13:11<42:08:47, 28.42s/it]
+
+ 43%|████████████████████████████████████████▍ | 4040/9378 [32:13:38<42:07:46, 28.41s/it]
+
+
+ 43%|████████████████████████████████████████▌ | 4041/9378 [32:14:06<42:10:55, 28.45s/it]
+
+ 43%|████████████████████████████████████████▌ | 4041/9378 [32:14:08<42:10:55, 28.45s/it]
+
+
+ 43%|████████████████████████████████████████▌ | 4043/9378 [32:15:05<42:12:53, 28.49s/it]
+
+ 43%|████████████████████████████████████████▌ | 4044/9378 [32:15:32<42:09:06, 28.45s/it]
+
+
+ 43%|████████████████████████████████████████▌ | 4045/9378 [32:16:01<42:16:29, 28.54s/it]
+
+
+ 43%|████████████████████████████████████████▌ | 4046/9378 [32:16:29<42:01:08, 28.37s/it]
+
+
+ 43%|████████████████████████████████████████▌ | 4047/9378 [32:16:57<42:00:57, 28.37s/it]
+
+
+ 43%|████████████████████████████████████████▌ | 4048/9378 [32:17:26<41:59:26, 28.36s/it]
+
+
+ 43%|████████████████████████████████████████▌ | 4049/9378 [32:17:56<42:18:20, 28.58s/it]
+
+ 43%|████████████████████████████████████████▌ | 4050/9378 [32:18:46<51:38:19, 34.89s/it]
+
+
+ 43%|████████████████████████████████████████▌ | 4051/9378 [32:19:14<48:28:43, 32.76s/it]
+
+ 43%|████████████████████████████████████████▌ | 4052/9378 [32:19:41<46:35:52, 31.50s/it]
+
+
+ 43%|████████████████████████████████████████▋ | 4053/9378 [32:20:09<45:09:30, 30.53s/it]
+
+
+ 43%|████████████████████████████████████████▋ | 4054/9378 [32:20:37<44:12:25, 29.89s/it]
+
+ 43%|████████████████████████████████████████▋ | 4054/9378 [32:20:39<44:12:25, 29.89s/it]
+
+
+ 43%|████████████████████████████████████████▋ | 4056/9378 [32:21:34<42:56:43, 29.05s/it]
+
+
+
+ 43%|████████████████████████████████████████▋ | 4057/9378 [32:22:04<42:35:39, 28.82s/it]
+
+ 43%|████████████████████████████████████████▋ | 4058/9378 [32:22:31<42:24:04, 28.69s/it]
+
+
+ 43%|████████████████████████████████████████▋ | 4059/9378 [32:22:59<42:11:22, 28.55s/it]
+
+
+ 43%|████████████████████████████████████████▋ | 4060/9378 [32:23:28<42:25:19, 28.72s/it]
+
+
+ 43%|████████████████████████████████████████▋ | 4061/9378 [32:23:56<42:03:07, 28.47s/it]
+
+
+
+ 43%|████████████████████████████████████████▋ | 4062/9378 [32:24:26<41:59:00, 28.43s/it]
+
+ 43%|████████████████████████████████████████▋ | 4063/9378 [32:24:54<42:08:12, 28.54s/it]
+
+ 43%|████████████████████████████████████████▋ | 4064/9378 [32:25:23<41:56:33, 28.41s/it]
+
+ 43%|████████████████████████████████████████▋ | 4065/9378 [32:25:49<41:40:07, 28.23s/it]
+
+
+ 43%|████████████████████████████████████████▊ | 4066/9378 [32:26:17<41:39:08, 28.23s/it]
+
+
+ 43%|████████████████████████████████████████▊ | 4067/9378 [32:26:46<41:43:32, 28.28s/it]
+
+
+ 43%|████████████████████████████████████████▊ | 4068/9378 [32:27:15<42:02:15, 28.50s/it]
+
+
+ 43%|████████████████████████████████████████▊ | 4069/9378 [32:27:43<41:55:50, 28.43s/it]
+
+
+ 43%|████████████████████████████████████████▊ | 4070/9378 [32:28:10<41:29:06, 28.14s/it]
+
+
+ 43%|████████████████████████████████████████▊ | 4071/9378 [32:28:41<41:47:20, 28.35s/it]
+
+ 43%|████████████████████████████████████████▊ | 4072/9378 [32:29:09<41:55:12, 28.44s/it]
+
+ 43%|████████████████████████████████████████▊ | 4073/9378 [32:29:36<41:50:02, 28.39s/it]
+
+
+ 43%|████████████████████████████████████████▊ | 4074/9378 [32:30:04<41:43:37, 28.32s/it]
+
+
+ 43%|████████████████████████████████████████▊ | 4075/9378 [32:30:34<41:44:27, 28.34s/it]
+
+ 43%|████████████████████████████████████████▊ | 4076/9378 [32:31:01<41:45:33, 28.35s/it]
+
+
+ 43%|████████████████████████████████████████▊ | 4077/9378 [32:31:29<41:31:06, 28.20s/it]
+
+
+ 43%|████████████████████████████████████████▉ | 4078/9378 [32:31:57<41:22:02, 28.10s/it]
+
+
+
+ 43%|████████████████████████████████████████▉ | 4079/9378 [32:32:27<41:31:48, 28.21s/it]
+
+
+ 44%|████████████████████████████████████████▉ | 4080/9378 [32:32:55<41:35:57, 28.27s/it]
+
+
+ 44%|████████████████████████████████████████▉ | 4081/9378 [32:33:24<41:36:30, 28.28s/it]
+
+ 44%|████████████████████████████████████████▉ | 4082/9378 [32:33:51<41:46:45, 28.40s/it]
+
+
+ 44%|████████████████████████████████████████▉ | 4083/9378 [32:34:19<41:43:33, 28.37s/it]
+
+
+ 44%|████████████████████████████████████████▉ | 4084/9378 [32:34:48<41:50:49, 28.46s/it]
+
+
+
+ 44%|████████████████████████████████████████▉ | 4085/9378 [32:35:17<41:37:18, 28.31s/it]
+
+
+ 44%|████████████████████████████████████████▉ | 4086/9378 [32:35:46<41:38:11, 28.32s/it]
+
+
+ 44%|████████████████████████████████████████▉ | 4087/9378 [32:36:14<41:36:02, 28.31s/it]
+
+ 44%|████████████████████████████████████████▉ | 4088/9378 [32:36:40<41:35:03, 28.30s/it]
+
+
+ 44%|████████████████████████████████████████▉ | 4089/9378 [32:37:09<41:44:04, 28.41s/it]
+
+
+ 44%|████████████████████████████████████████▉ | 4090/9378 [32:37:37<41:40:10, 28.37s/it]
+
+
+ 44%|█████████████████████████████████████████ | 4091/9378 [32:38:07<41:58:36, 28.58s/it]
+
+
+ 44%|█████████████████████████████████████████ | 4092/9378 [32:38:36<41:32:09, 28.29s/it]
+
+ 44%|█████████████████████████████████████████ | 4093/9378 [32:39:04<41:48:43, 28.48s/it]
+
+ 44%|█████████████████████████████████████████ | 4094/9378 [32:39:31<41:30:15, 28.28s/it]
+
+
+ 44%|█████████████████████████████████████████ | 4095/9378 [32:39:59<41:40:30, 28.40s/it]
+
+
+ 44%|█████████████████████████████████████████ | 4096/9378 [32:40:28<41:33:43, 28.33s/it]
+
+
+
+ 44%|█████████████████████████████████████████ | 4097/9378 [32:40:58<41:44:33, 28.46s/it]
+
+
+ 44%|█████████████████████████████████████████ | 4098/9378 [32:41:26<41:33:22, 28.33s/it]
+
+ 44%|█████████████████████████████████████████ | 4099/9378 [32:41:54<41:52:02, 28.55s/it]
+
+ 44%|█████████████████████████████████████████ | 4100/9378 [32:42:43<51:03:29, 34.83s/it]
+
+
+ 44%|█████████████████████████████████████████ | 4101/9378 [32:43:15<49:06:33, 33.50s/it]
+
+
+ 44%|█████████████████████████████████████████ | 4102/9378 [32:43:45<47:07:29, 32.15s/it]
+
+ 44%|█████████████████████████████████████████▏ | 4103/9378 [32:44:13<46:29:54, 31.73s/it]
+
+
+ 44%|█████████████████████████████████████████▏ | 4104/9378 [32:44:45<45:55:11, 31.34s/it]
+
+ 44%|█████████████████████████████████████████▏ | 4105/9378 [32:45:12<44:30:53, 30.39s/it]
+
+
+ 44%|█████████████████████████████████████████▏ | 4106/9378 [32:45:41<44:00:10, 30.05s/it]
+
+
+
+ 44%|█████████████████████████████████████████▏ | 4107/9378 [32:46:10<42:54:14, 29.30s/it]
+
+
+ 44%|█████████████████████████████████████████▏ | 4108/9378 [32:46:39<42:27:30, 29.00s/it]
+
+ 44%|█████████████████████████████████████████▏ | 4109/9378 [32:47:07<42:13:51, 28.85s/it]
+
+ 44%|█████████████████████████████████████████▏ | 4110/9378 [32:47:35<42:37:36, 29.13s/it]
+
+
+
+ 44%|█████████████████████████████████████████▏ | 4111/9378 [32:48:04<41:55:11, 28.65s/it]
+
+
+ 44%|█████████████████████████████████████████▏ | 4112/9378 [32:48:33<41:47:29, 28.57s/it]
+
+
+ 44%|█████████████████████████████████████████▏ | 4113/9378 [32:49:01<41:35:36, 28.44s/it]
+
+ 44%|█████████████████████████████████████████▏ | 4114/9378 [32:49:28<41:48:20, 28.59s/it]
+
+
+ 44%|█████████████████████████████████████████▏ | 4115/9378 [32:49:56<41:28:54, 28.37s/it]
+
+
+ 44%|█████████████████████████████████████████▎ | 4116/9378 [32:50:25<41:48:24, 28.60s/it]
+
+
+ 44%|█████████████████████████████████████████▎ | 4117/9378 [32:50:53<41:18:02, 28.26s/it]
+
+
+
+ 44%|█████████████████████████████████████████▎ | 4118/9378 [32:51:23<41:18:59, 28.28s/it]
+
+
+ 44%|█████████████████████████████████████████▎ | 4119/9378 [32:51:51<41:14:19, 28.23s/it]
+
+ 44%|█████████████████████████████████████████▎ | 4120/9378 [32:52:19<41:25:41, 28.36s/it]
+
+ 44%|█████████████████████████████████████████▎ | 4121/9378 [32:52:47<41:32:43, 28.45s/it]
+
+ 44%|█████████████████████████████████████████▎ | 4122/9378 [32:53:15<41:31:10, 28.44s/it]
+
+
+ 44%|█████████████████████████████████████████▎ | 4123/9378 [32:53:42<41:07:28, 28.17s/it]
+
+
+ 44%|█████████████████████████████████████████▎ | 4124/9378 [32:54:11<41:15:09, 28.27s/it]
+
+
+
+ 44%|█████████████████████████████████████████▎ | 4125/9378 [32:54:41<41:13:37, 28.25s/it]
+
+
+ 44%|█████████████████████████████████████████▎ | 4126/9378 [32:55:09<41:12:05, 28.24s/it]
+
+
+ 44%|█████████████████████████████████████████▎ | 4127/9378 [32:55:37<41:15:44, 28.29s/it]
+
+ 44%|█████████████████████████████████████████▍ | 4128/9378 [32:56:06<41:40:56, 28.58s/it]
+
+ 44%|█████████████████████████████████████████▍ | 4129/9378 [32:56:33<41:16:18, 28.31s/it]
+
+
+ 44%|█████████████████████████████████████████▍ | 4130/9378 [32:57:02<41:41:05, 28.59s/it]
+
+
+
+ 44%|█████████████████████████████████████████▍ | 4131/9378 [32:57:31<41:14:08, 28.29s/it]
+
+
+ 44%|███████████████████████▊ | 4132/9378 [32:58:00<41:15:55, 28.32s/it]
+
+ 44%|███████████████████████▊ | 4133/9378 [32:58:28<41:20:22, 28.37s/it]
+
+ 44%|███████████████████████▊ | 4134/9378 [32:58:55<41:20:16, 28.38s/it]
+
+
+ 44%|███████████████████████▊ | 4135/9378 [32:59:24<41:41:32, 28.63s/it]
+
+
+ 44%|███████████████████████▊ | 4136/9378 [32:59:51<41:12:29, 28.30s/it]
+
+ 44%|███████████████████████▊ | 4136/9378 [32:59:53<41:12:29, 28.30s/it]
+
+
+ 44%|███████████████████████▊ | 4138/9378 [33:00:50<41:30:19, 28.52s/it]
+
+ 44%|███████████████████████▊ | 4139/9378 [33:01:17<41:14:46, 28.34s/it]
+
+
+ 44%|███████████████████████▊ | 4140/9378 [33:01:48<41:50:33, 28.76s/it]
+
+ 44%|███████████████████████▊ | 4141/9378 [33:02:16<41:49:41, 28.75s/it]
+
+ 44%|███████████████████████▊ | 4142/9378 [33:02:43<41:30:10, 28.54s/it]
+
+
+ 44%|███████████████████████▊ | 4143/9378 [33:03:11<41:11:13, 28.32s/it]
+
+
+ 44%|███████████████████████▊ | 4144/9378 [33:03:40<41:20:17, 28.43s/it]
+
+
+ 44%|███████████████████████▊ | 4145/9378 [33:04:08<41:15:36, 28.38s/it]
+
+
+
+ 44%|███████████████████████▊ | 4146/9378 [33:04:38<41:14:05, 28.37s/it]
+
+ 44%|███████████████████████▉ | 4147/9378 [33:05:07<41:10:15, 28.33s/it]
+
+ 44%|███████████████████████▉ | 4148/9378 [33:05:33<41:09:58, 28.34s/it]
+
+
+ 44%|███████████████████████▉ | 4149/9378 [33:06:02<41:11:01, 28.35s/it]
+
+
+ 44%|███████████████████████▉ | 4150/9378 [33:06:49<49:39:35, 34.20s/it]
+
+
+ 44%|███████████████████████▉ | 4151/9378 [33:07:18<47:07:27, 32.46s/it]
+
+
+ 44%|███████████████████████▉ | 4152/9378 [33:07:46<45:20:53, 31.24s/it]
+
+ 44%|███████████████████████▉ | 4152/9378 [33:07:48<45:20:53, 31.24s/it]
+
+
+ 44%|███████████████████████▉ | 4154/9378 [33:08:45<43:11:31, 29.76s/it]
+
+ 44%|███████████████████████▉ | 4155/9378 [33:09:13<43:03:08, 29.67s/it]
+
+ 44%|███████████████████████▉ | 4156/9378 [33:09:40<42:07:07, 29.04s/it]
+
+
+ 44%|███████████████████████▉ | 4157/9378 [33:10:09<41:56:52, 28.92s/it]
+
+
+
+ 44%|███████████████████████▉ | 4158/9378 [33:10:39<41:39:33, 28.73s/it]
+
+ 44%|███████████████████████▉ | 4159/9378 [33:11:07<41:26:05, 28.58s/it]
+
+ 44%|███████████████████████▉ | 4160/9378 [33:11:34<41:19:18, 28.51s/it]
+
+
+ 44%|███████████████████████▉ | 4161/9378 [33:12:03<41:31:57, 28.66s/it]
+
+
+ 44%|███████████████████████▉ | 4162/9378 [33:12:30<41:02:56, 28.33s/it]
+
+
+ 44%|███████████████████████▉ | 4163/9378 [33:12:59<41:24:13, 28.58s/it]
+
+
+ 44%|███████████████████████▉ | 4164/9378 [33:13:27<40:58:52, 28.30s/it]
+
+
+ 44%|███████████████████████▉ | 4165/9378 [33:13:57<41:18:55, 28.53s/it]
+
+
+ 44%|███████████████████████▉ | 4166/9378 [33:14:25<40:59:31, 28.31s/it]
+
+ 44%|███████████████████████▉ | 4167/9378 [33:14:54<40:57:08, 28.29s/it]
+
+ 44%|████████████████████████ | 4168/9378 [33:15:21<41:01:35, 28.35s/it]
+
+
+ 44%|████████████████████████ | 4169/9378 [33:15:49<41:04:27, 28.39s/it]
+
+
+
+ 44%|████████████████████████ | 4170/9378 [33:16:19<41:09:27, 28.45s/it]
+
+
+ 44%|████████████████████████ | 4171/9378 [33:16:47<41:04:56, 28.40s/it]
+
+
+ 44%|████████████████████████ | 4172/9378 [33:17:16<40:53:21, 28.28s/it]
+
+
+ 44%|████████████████████████ | 4173/9378 [33:17:44<40:51:49, 28.26s/it]
+
+ 45%|████████████████████████ | 4174/9378 [33:18:11<41:18:47, 28.58s/it]
+
+
+ 45%|████████████████████████ | 4175/9378 [33:18:39<40:48:29, 28.24s/it]
+
+
+ 45%|████████████████████████ | 4176/9378 [33:19:07<40:49:58, 28.26s/it]
+
+
+
+ 45%|████████████████████████ | 4177/9378 [33:19:37<40:52:33, 28.29s/it]
+
+
+ 45%|████████████████████████ | 4178/9378 [33:20:06<40:55:47, 28.34s/it]
+
+
+ 45%|████████████████████████ | 4179/9378 [33:20:34<40:52:12, 28.30s/it]
+
+ 45%|████████████████████████ | 4180/9378 [33:21:02<40:56:07, 28.35s/it]
+{'loss': 1.1942, 'grad_norm': 0.5625, 'learning_rate': 1.9595921478845917e-05, 'epoch': 1.32}
+[2024-05-13 19:53:55,667] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-4180/pytorch_model_fsdp.bin
+[2024-05-13 19:54:07,363] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-4180/pytorch_model_fsdp.bin
+[2024-05-13 19:54:29,136] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-4180/optimizer.bin
+ 45%|████████████████████████ | 4180/9378 [33:21:02<40:56:07, 28.35s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 45%|████████████████████████ | 4181/9378 [33:23:42<98:01:30, 67.90s/it]
+
+
+ 45%|████████████████████████ | 4182/9378 [33:24:10<80:40:23, 55.89s/it]
+
+ 45%|████████████████████████ | 4183/9378 [33:24:37<68:42:37, 47.61s/it]
+
+
+ 45%|████████████████████████ | 4184/9378 [33:25:06<60:31:27, 41.95s/it]
+
+
+ 45%|████████████████████████ | 4185/9378 [33:25:34<54:23:40, 37.71s/it]
+
+
+
+ 45%|████████████████████████ | 4186/9378 [33:26:04<50:23:25, 34.94s/it]
+
+
+ 45%|████████████████████████ | 4187/9378 [33:26:32<47:34:46, 33.00s/it]
+
+
+ 45%|████████████████████████ | 4188/9378 [33:27:01<45:32:33, 31.59s/it]
+
+ 45%|████████████████████████ | 4189/9378 [33:27:29<44:08:41, 30.63s/it]
+
+ 45%|████████████████████████▏ | 4190/9378 [33:27:55<43:06:45, 29.92s/it]
+
+
+ 45%|████████████████████████▏ | 4191/9378 [33:28:24<42:27:16, 29.47s/it]
+
+
+ 45%|████████████████████████▏ | 4192/9378 [33:28:52<41:57:12, 29.12s/it]
+
+
+
+ 45%|████████████████████████▏ | 4193/9378 [33:29:22<41:35:44, 28.88s/it]
+
+
+ 45%|████████████████████████▏ | 4194/9378 [33:29:51<41:24:15, 28.75s/it]
+
+ 45%|████████████████████████▏ | 4195/9378 [33:30:19<41:11:32, 28.61s/it]
+
+ 45%|████████████████████████▏ | 4196/9378 [33:30:45<41:01:16, 28.50s/it]
+
+
+ 45%|████████████████████████▏ | 4197/9378 [33:31:14<40:56:34, 28.45s/it]
+
+
+ 45%|████████████████████████▏ | 4198/9378 [33:31:42<40:53:14, 28.42s/it]
+
+
+
+ 45%|████████████████████████▏ | 4199/9378 [33:32:12<40:56:32, 28.46s/it]
+
+ 45%|████████████████████████▏ | 4200/9378 [33:33:00<50:07:25, 34.85s/it]
+
+
+
+ 45%|████████████████████████▏ | 4201/9378 [33:33:31<47:19:34, 32.91s/it]
+
+
+ 45%|████████████████████████▏ | 4202/9378 [33:33:59<45:17:21, 31.50s/it]
+
+
+ 45%|████████████████████████▏ | 4203/9378 [33:34:27<43:54:01, 30.54s/it]
+
+ 45%|████████████████████████▏ | 4204/9378 [33:34:56<43:18:41, 30.14s/it]
+
+ 45%|████████████████████████▏ | 4205/9378 [33:35:23<42:33:29, 29.62s/it]
+
+
+ 45%|████████████████████████▏ | 4206/9378 [33:35:50<41:35:03, 28.94s/it]
+
+
+ 45%|████████████████████████▏ | 4207/9378 [33:36:19<41:15:34, 28.72s/it]
+
+
+
+ 45%|████████████████████████▏ | 4208/9378 [33:36:49<41:16:59, 28.75s/it]
+
+
+ 45%|████████████████████████▏ | 4209/9378 [33:37:17<40:55:00, 28.50s/it]
+
+
+ 45%|████████████████████████▏ | 4210/9378 [33:37:45<40:49:58, 28.44s/it]
+
+
+ 45%|████████████████████████▏ | 4211/9378 [33:38:14<40:46:36, 28.41s/it]
+
+ 45%|████████████████████████▎ | 4212/9378 [33:38:42<40:41:33, 28.36s/it]
+
+ 45%|████████████████████████▎ | 4213/9378 [33:39:08<40:38:54, 28.33s/it]
+
+
+ 45%|████████████████████████▎ | 4214/9378 [33:39:38<40:59:51, 28.58s/it]
+
+
+ 45%|████████████████████████▎ | 4215/9378 [33:40:06<40:42:44, 28.39s/it]
+
+
+
+ 45%|████████████████████████▎ | 4216/9378 [33:40:36<40:36:46, 28.32s/it]
+
+
+ 45%|████████████████████████▎ | 4217/9378 [33:41:04<40:34:08, 28.30s/it]
+
+ 45%|████████████████████████▎ | 4218/9378 [33:41:32<40:34:38, 28.31s/it]
+
+ 45%|████████████████████████▎ | 4219/9378 [33:41:59<40:36:40, 28.34s/it]
+
+
+ 45%|████████████████████████▎ | 4220/9378 [33:42:27<40:37:51, 28.36s/it]
+
+
+ 45%|████████████████████████▎ | 4221/9378 [33:42:56<40:45:00, 28.45s/it]
+
+
+
+ 45%|████████████████████████▎ | 4222/9378 [33:43:26<40:37:31, 28.37s/it]
+
+
+ 45%|████████████████████████▎ | 4223/9378 [33:43:54<40:43:55, 28.45s/it]
+
+
+ 45%|████████████████████████▎ | 4224/9378 [33:44:22<40:29:19, 28.28s/it]
+
+
+ 45%|████████████████████████▎ | 4225/9378 [33:44:50<40:29:15, 28.29s/it]
+
+ 45%|████████████████████████▎ | 4226/9378 [33:45:19<40:32:45, 28.33s/it]
+
+ 45%|████████████████████████▎ | 4227/9378 [33:45:46<40:32:33, 28.33s/it]
+
+
+ 45%|████████████████████████▎ | 4228/9378 [33:46:14<40:33:41, 28.35s/it]
+
+
+
+ 45%|████████████████████████▎ | 4229/9378 [33:46:44<40:53:00, 28.58s/it]
+
+ 45%|████████████████████████▎ | 4230/9378 [33:47:12<40:47:37, 28.53s/it]
+
+
+ 45%|████████████████████████▎ | 4231/9378 [33:47:41<40:24:33, 28.26s/it]
+
+ 45%|████████████████████████▎ | 4232/9378 [33:48:08<40:28:25, 28.31s/it]
+
+ 45%|████████████████████████▎ | 4232/9378 [33:48:09<40:28:25, 28.31s/it]
+[2024-05-13 20:20:28,050] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: ff880d40-8eab-4852-8939-89e219910dc3)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/4deecbd887e9d6b716c18757219a4b21fd328bb694c6c5726424388a15e2654e?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T200238Z&X-Amz-Expires=86400&X-Amz-Signature=b583a515ffab8133cb9f01fd3dd79c2acede4ae516840c59c44b9305b7c598ba&X-Amz-SignedHeaders=host&partNumber=1&uploadId=M3k40K3ZtO4_F9Sk8oJ4BhnEuDHBbNZVdoTAX5X1hdb38.JT7XXeFaHxtQgOaS_37sNKw5IGt7n.KGTa9PztFfBdZqFGNzRzMVpBvl4kmWWh31mSonvpeBNgqxybrP4A&x-id=UploadPart
+ 45%|████████████████████████▎ | 4232/9378 [33:48:09<40:28:25, 28.31s/it]'(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: ff880d40-8eab-4852-8939-89e219910dc3)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/4deecbd887e9d6b716c18757219a4b21fd328bb694c6c5726424388a15e2654e?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T200238Z&X-Amz-Expires=86400&X-Amz-Signature=b583a515ffab8133cb9f01fd3dd79c2acede4ae516840c59c44b9305b7c598ba&X-Amz-SignedHeaders=host&partNumber=1&uploadId=M3k40K3ZtO4_F9Sk8oJ4BhnEuDHBbNZVdoTAX5X1hdb38.JT7XXeFaHxtQgOaS_37sNKw5IGt7n.KGTa9PztFfBdZqFGNzRzMVpBvl4kmWWh31mSonvpeBNgqxybrP4A&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+ 45%|████████████████████████▎ | 4233/9378 [33:48:36<40:31:24, 28.35s/it]
+
+ 45%|████████████████████████▎ | 4233/9378 [33:48:38<40:31:24, 28.35s/it]'(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: a5c113f8-c6fc-4fdf-97fb-d728f950609d)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/abb290cc4c3293eb4b16bf1ee57ec991e525716d36f9c675ba88410f289e4689?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T200238Z&X-Amz-Expires=86400&X-Amz-Signature=467a3cc48fd194e97f53a3c21d68b7de44546b0216bd354fac121bdd13c54c2a&X-Amz-SignedHeaders=host&partNumber=1&uploadId=W6inFbkDKg5m_YH2LsHmZCoWa_haiW6ZKkE64RJBXWeLfd680AvRSxqnDNW8PuynyQYwCUTevaH_ENKW6sS3xyJcIRl_hrGKOu_Y3lWjA_DL_bLR9RnyP8A03wLNM_SQ&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+[2024-05-13 20:20:50,578] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:44754] '(ReadTimeoutError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Read timed out. (read timeout=None)"), '(Request ID: a5c113f8-c6fc-4fdf-97fb-d728f950609d)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/e2/17/e2171a760c7937a703ae36fa42969b69c96e452a57acf632d191fb68df5fdc23/abb290cc4c3293eb4b16bf1ee57ec991e525716d36f9c675ba88410f289e4689?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240513%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240513T200238Z&X-Amz-Expires=86400&X-Amz-Signature=467a3cc48fd194e97f53a3c21d68b7de44546b0216bd354fac121bdd13c54c2a&X-Amz-SignedHeaders=host&partNumber=1&uploadId=W6inFbkDKg5m_YH2LsHmZCoWa_haiW6ZKkE64RJBXWeLfd680AvRSxqnDNW8PuynyQYwCUTevaH_ENKW6sS3xyJcIRl_hrGKOu_Y3lWjA_DL_bLR9RnyP8A03wLNM_SQ&x-id=UploadPart
+[2024-05-13 20:20:50,579] [WARNING] [huggingface_hub.utils._http.http_backoff:302] [PID:44754] Retrying in 1s [Retry 1/5].
+
+ 45%|████████████████████████▍ | 4234/9378 [33:49:06<40:33:10, 28.38s/it]
+
+
+ 45%|████████████████████████▍ | 4235/9378 [33:49:34<40:27:19, 28.32s/it]
+
+
+ 45%|████████████████████████▍ | 4236/9378 [33:50:03<40:26:26, 28.31s/it]
+
+
+ 45%|████████████████████████▍ | 4237/9378 [33:50:31<40:27:22, 28.33s/it]
+
+ 45%|████████████████████████▍ | 4238/9378 [33:50:59<40:32:51, 28.40s/it]
+
+ 45%|████████████████████████▍ | 4239/9378 [33:51:26<40:34:03, 28.42s/it]
+
+
+
+ 45%|████████████████████████▍ | 4240/9378 [33:51:56<40:50:10, 28.61s/it]
+
+
+ 45%|████████████████████████▍ | 4241/9378 [33:52:25<40:21:12, 28.28s/it]
+
+
+ 45%|████████████████████████▍ | 4242/9378 [33:52:53<40:19:16, 28.26s/it]
+
+ 45%|████████████████████████▍ | 4243/9378 [33:53:22<40:30:16, 28.40s/it]
+
+ 45%|████████████████████████▍ | 4244/9378 [33:53:48<40:30:50, 28.41s/it]
+
+
+ 45%|████████████████████████▍ | 4245/9378 [33:54:17<40:27:51, 28.38s/it]
+
+
+
+ 45%|████████████████████████▍ | 4246/9378 [33:54:47<40:27:32, 28.38s/it]
+
+ 45%|████████████████████████▍ | 4247/9378 [33:55:15<40:48:42, 28.63s/it]
+
+
+ 45%|████████████████████████▍ | 4248/9378 [33:55:44<40:23:08, 28.34s/it]
+
+ 45%|████████████████████████▍ | 4249/9378 [33:56:12<40:20:00, 28.31s/it]
+
+
+ 45%|████████████████████████▍ | 4250/9378 [33:57:00<48:40:01, 34.17s/it]
+
+ 45%|████████████████████████▍ | 4251/9378 [33:57:28<46:24:28, 32.59s/it]
+
+ 45%|████████████████████████▍ | 4252/9378 [33:57:54<44:16:33, 31.10s/it]
+
+
+ 45%|████████████████████████▍ | 4253/9378 [33:58:23<43:03:45, 30.25s/it]
+
+
+ 45%|████████████████████████▍ | 4254/9378 [33:58:51<42:12:50, 29.66s/it]
+
+
+
+ 45%|████████████████████████▌ | 4255/9378 [33:59:21<41:40:12, 29.28s/it]
+
+
+ 45%|████████████████████████▌ | 4256/9378 [33:59:49<41:14:04, 28.98s/it]
+
+
+ 45%|████████████████████████▌ | 4257/9378 [34:00:18<41:02:37, 28.85s/it]
+
+
+ 45%|████████████████████████▌ | 4258/9378 [34:00:46<40:44:56, 28.65s/it]
+
+ 45%|████████████████████████▌ | 4259/9378 [34:01:13<40:34:14, 28.53s/it]
+
+
+ 45%|████████████████████████▌ | 4260/9378 [34:01:42<40:46:17, 28.68s/it]
+
+
+ 45%|████████████████████████▌ | 4261/9378 [34:02:09<40:15:09, 28.32s/it]
+
+
+
+ 45%|████████████████████████▌ | 4262/9378 [34:02:39<40:31:08, 28.51s/it]
+
+
+ 45%|████████████████████████▌ | 4263/9378 [34:03:08<40:27:51, 28.48s/it]
+
+
+ 45%|████████████████████████▌ | 4264/9378 [34:03:36<40:05:28, 28.22s/it]
+
+
+ 45%|████████████████████████▌ | 4265/9378 [34:04:04<40:08:27, 28.26s/it]
+
+ 45%|████████████████████████▌ | 4266/9378 [34:04:33<40:11:37, 28.31s/it]
+
+
+ 46%|████████████████████████▌ | 4267/9378 [34:05:01<40:19:39, 28.41s/it]
+
+
+ 46%|████████████████████████▌ | 4268/9378 [34:05:30<40:22:41, 28.45s/it]
+
+
+ 46%|████████████████████████▌ | 4269/9378 [34:05:58<40:19:58, 28.42s/it]
+
+ 46%|████████████████████████▌ | 4270/9378 [34:06:27<40:33:39, 28.59s/it]
+
+ 46%|████████████████████████▌ | 4271/9378 [34:06:53<40:16:49, 28.39s/it]
+
+
+
+ 46%|████████████████████████▌ | 4272/9378 [34:07:23<40:33:36, 28.60s/it]
+
+
+ 46%|████████████████████████▌ | 4273/9378 [34:07:52<40:06:43, 28.29s/it]
+
+ 46%|████████████████████████▌ | 4274/9378 [34:08:20<40:27:32, 28.54s/it]
+
+
+ 46%|████████████████████████▌ | 4275/9378 [34:08:49<40:06:09, 28.29s/it]
+
+ 46%|████████████████████████▌ | 4276/9378 [34:09:17<40:28:47, 28.56s/it]
+
+ 46%|████████████████████████▋ | 4277/9378 [34:09:44<40:02:37, 28.26s/it]
+
+
+ 46%|████████████████████████▋ | 4278/9378 [34:10:12<40:12:19, 28.38s/it]
+
+
+
+ 46%|████████████████████████▋ | 4279/9378 [34:10:42<40:16:44, 28.44s/it]
+
+
+ 46%|████████████████████████▋ | 4280/9378 [34:11:11<40:12:43, 28.40s/it]
+
+
+ 46%|████████████████████████▋ | 4281/9378 [34:11:39<40:07:39, 28.34s/it]
+
+ 46%|████████████████████████▋ | 4282/9378 [34:12:06<40:09:04, 28.36s/it]
+
+
+ 46%|████████████████████████▋ | 4283/9378 [34:12:35<40:29:04, 28.61s/it]
+
+
+ 46%|████████████████████████▋ | 4284/9378 [34:13:03<40:20:05, 28.51s/it]
+
+
+
+ 46%|████████████████████████▋ | 4285/9378 [34:13:32<40:13:45, 28.44s/it]
+
+ 46%|████████████████████████▋ | 4286/9378 [34:14:01<40:17:16, 28.48s/it]
+
+
+ 46%|████████████████████████▋ | 4287/9378 [34:14:29<39:49:34, 28.16s/it]
+
+ 46%|████████████████████████▋ | 4288/9378 [34:14:56<39:55:46, 28.24s/it]
+
+
+ 46%|████████████████████████▋ | 4289/9378 [34:15:24<39:56:54, 28.26s/it]
+
+
+ 46%|████████████████████████▋ | 4290/9378 [34:15:52<39:57:01, 28.27s/it]
+
+
+ 46%|████████████████████████▋ | 4291/9378 [34:16:21<39:56:26, 28.27s/it]
+
+
+
+ 46%|████████████████████████▋ | 4292/9378 [34:16:51<40:16:12, 28.50s/it]
+
+
+ 46%|████████████████████████▋ | 4293/9378 [34:17:19<39:52:35, 28.23s/it]
+
+
+ 46%|████████████████████████▋ | 4294/9378 [34:17:47<39:55:04, 28.27s/it]
+
+ 46%|████████████████████████▋ | 4295/9378 [34:18:16<39:56:13, 28.29s/it]
+
+ 46%|████████████████████████▋ | 4296/9378 [34:18:42<39:58:31, 28.32s/it]
+
+
+ 46%|████████████████████████▋ | 4297/9378 [34:19:11<39:53:21, 28.26s/it]
+
+
+ 46%|████████████████████████▋ | 4298/9378 [34:19:39<39:56:50, 28.31s/it]
+
+
+
+ 46%|████████████████████████▊ | 4299/9378 [34:20:09<39:53:15, 28.27s/it]
+
+ 46%|████████████████████████▊ | 4300/9378 [34:20:55<48:05:47, 34.10s/it]
+
+
+ 46%|████████████████████████▊ | 4301/9378 [34:21:23<45:36:08, 32.34s/it]
+
+
+
+ 46%|████████████████████████▊ | 4302/9378 [34:21:53<43:53:41, 31.13s/it]
+
+
+ 46%|████████████████████████▊ | 4303/9378 [34:22:21<42:41:33, 30.28s/it]
+
+ 46%|████████████████████████▊ | 4304/9378 [34:22:50<42:10:01, 29.92s/it]
+
+ 46%|████████████████████████▊ | 4305/9378 [34:23:18<41:12:26, 29.24s/it]
+
+ 46%|████████████████████████▊ | 4306/9378 [34:23:45<40:48:17, 28.96s/it]
+
+
+ 46%|████████████████████████▊ | 4307/9378 [34:24:13<40:27:37, 28.72s/it]
+
+
+ 46%|████████████████████████▊ | 4308/9378 [34:24:41<40:17:22, 28.61s/it]
+
+
+
+ 46%|████████████████████████▊ | 4309/9378 [34:25:11<40:06:51, 28.49s/it]
+
+ 46%|████████████████████████▊ | 4310/9378 [34:25:40<40:23:08, 28.69s/it]
+
+
+ 46%|████████████████████████▊ | 4311/9378 [34:26:08<39:55:34, 28.37s/it]
+
+ 46%|████████████████████████▊ | 4312/9378 [34:26:37<40:00:26, 28.43s/it]
+
+ 46%|████████████████████████▊ | 4313/9378 [34:27:03<39:56:31, 28.39s/it]
+
+
+ 46%|████████████████████████▊ | 4314/9378 [34:27:31<39:53:15, 28.36s/it]
+
+
+
+ 46%|████████████████████████▊ | 4315/9378 [34:28:02<40:01:41, 28.46s/it]
+
+
+ 46%|████████████████████████▊ | 4316/9378 [34:28:30<39:56:44, 28.41s/it]
+
+
+ 46%|████████████████████████▊ | 4317/9378 [34:28:59<39:55:49, 28.40s/it]
+
+ 46%|████████████████████████▊ | 4318/9378 [34:29:25<40:02:52, 28.49s/it]
+
+
+ 46%|████████████████████████▊ | 4319/9378 [34:29:53<39:45:43, 28.29s/it]
+
+
+ 46%|████████████████████████▉ | 4320/9378 [34:30:22<39:47:47, 28.32s/it]
+
+
+
+ 46%|████████████████████████▉ | 4321/9378 [34:30:52<39:44:16, 28.29s/it]
+
+
+ 46%|████████████████████████▉ | 4322/9378 [34:31:20<39:46:07, 28.32s/it]
+
+ 46%|████████████████████████▉ | 4323/9378 [34:31:48<40:03:52, 28.53s/it]
+
+
+ 46%|████████████████████████▉ | 4324/9378 [34:32:17<39:39:16, 28.25s/it]
+
+ 46%|████████████████████████▉ | 4325/9378 [34:32:45<39:43:08, 28.30s/it]
+
+ 46%|████████████████████████▉ | 4326/9378 [34:33:12<39:54:12, 28.43s/it]
+
+
+ 46%|████████████████████████▉ | 4327/9378 [34:33:41<39:59:16, 28.50s/it]
+
+
+ 46%|████████████████████████▉ | 4328/9378 [34:34:08<39:33:45, 28.20s/it]
+
+
+
+ 46%|████████████████████████▉ | 4329/9378 [34:34:38<39:40:00, 28.28s/it]
+
+
+ 46%|████████████████████████▉ | 4330/9378 [34:35:07<39:45:42, 28.36s/it]
+
+ 46%|████████████████████████▉ | 4331/9378 [34:35:35<39:43:15, 28.33s/it]
+
+ 46%|████████████████████████▉ | 4332/9378 [34:36:02<39:40:42, 28.31s/it]
+
+
+ 46%|████████████████████████▉ | 4333/9378 [34:36:30<39:44:06, 28.35s/it]
+
+
+ 46%|████████████████████████▉ | 4334/9378 [34:36:59<39:45:37, 28.38s/it]
+
+
+
+ 46%|████████████████████████▉ | 4335/9378 [34:37:29<39:47:56, 28.41s/it]
+
+ 46%|████████████████████████▉ | 4336/9378 [34:37:57<40:05:16, 28.62s/it]
+
+ 46%|████████████████████████▉ | 4337/9378 [34:38:24<39:39:02, 28.32s/it]
+
+
+ 46%|████████████████████████▉ | 4338/9378 [34:38:52<39:42:06, 28.36s/it]
+
+
+ 46%|████████████████████████▉ | 4339/9378 [34:39:21<39:39:04, 28.33s/it]
+
+
+
+ 46%|████████████████████████▉ | 4340/9378 [34:39:51<39:43:00, 28.38s/it]
+
+
+ 46%|████████████████████████▉ | 4341/9378 [34:40:19<39:43:33, 28.39s/it]
+
+ 46%|█████████████████████████ | 4342/9378 [34:40:48<39:46:21, 28.43s/it]
+
+ 46%|█████████████████████████ | 4343/9378 [34:41:14<39:43:29, 28.40s/it]
+
+
+ 46%|█████████████████████████ | 4344/9378 [34:41:43<39:39:22, 28.36s/it]
+
+
+ 46%|█████████████████████████ | 4345/9378 [34:42:11<39:35:22, 28.32s/it]
+
+
+ 46%|█████████████████████████ | 4346/9378 [34:42:40<39:54:58, 28.56s/it]
+
+
+
+ 46%|█████████████████████████ | 4347/9378 [34:43:09<39:29:19, 28.26s/it]
+
+ 46%|█████████████████████████ | 4348/9378 [34:43:38<39:52:27, 28.54s/it]
+
+
+ 46%|█████████████████████████ | 4349/9378 [34:44:06<39:28:54, 28.26s/it]
+
+
+ 46%|█████████████████████████ | 4350/9378 [34:44:54<47:38:13, 34.11s/it]
+
+
+ 46%|█████████████████████████ | 4351/9378 [34:45:22<45:19:17, 32.46s/it]
+
+ 46%|█████████████████████████ | 4352/9378 [34:45:50<43:38:39, 31.26s/it]
+
+ 46%|█████████████████████████ | 4353/9378 [34:46:19<42:27:01, 30.41s/it]
+
+ 46%|█████████████████████████ | 4354/9378 [34:46:45<41:23:05, 29.65s/it]
+
+
+ 46%|█████████████████████████ | 4355/9378 [34:47:14<40:49:24, 29.26s/it]
+
+
+
+ 46%|█████████████████████████ | 4356/9378 [34:47:43<40:16:27, 28.87s/it]
+
+
+ 46%|█████████████████████████ | 4357/9378 [34:48:12<40:22:39, 28.95s/it]
+
+
+ 46%|█████████████████████████ | 4358/9378 [34:48:40<39:40:32, 28.45s/it]
+
+
+ 46%|█████████████████████████ | 4359/9378 [34:49:08<39:36:21, 28.41s/it]
+
+
+ 46%|█████████████████████████ | 4360/9378 [34:49:37<39:34:03, 28.39s/it]
+
+ 47%|█████████████████████████ | 4361/9378 [34:50:04<39:53:37, 28.63s/it]
+
+
+ 47%|█████████████████████████ | 4362/9378 [34:50:32<39:30:07, 28.35s/it]
+
+
+
+ 47%|█████████████████████████ | 4363/9378 [34:51:02<39:23:57, 28.28s/it]
+
+
+ 47%|█████████████████████████▏ | 4364/9378 [34:51:30<39:20:16, 28.24s/it]
+
+
+ 47%|█████████████████████████▏ | 4365/9378 [34:51:58<39:21:34, 28.27s/it]
+
+
+ 47%|█████████████████████████▏ | 4366/9378 [34:52:26<39:19:20, 28.24s/it]
+
+ 47%|█████████████████████████▏ | 4367/9378 [34:52:55<39:42:09, 28.52s/it]
+
+ 47%|█████████████████████████▏ | 4368/9378 [34:53:23<39:34:40, 28.44s/it]
+
+ 47%|█████████████████████████▏ | 4369/9378 [34:53:51<39:13:26, 28.19s/it]
+
+ 47%|█████████████████████████▏ | 4370/9378 [34:54:18<39:22:32, 28.31s/it]
+
+
+
+ 47%|█████████████████████████▏ | 4371/9378 [34:54:48<39:19:19, 28.27s/it]
+
+
+ 47%|█████████████████████████▏ | 4372/9378 [34:55:16<39:24:02, 28.33s/it]
+
+
+ 47%|█████████████████████████▏ | 4373/9378 [34:55:45<39:24:08, 28.34s/it]
+
+ 47%|█████████████████████████▏ | 4374/9378 [34:56:13<39:44:43, 28.59s/it]
+
+ 47%|█████████████████████████▏ | 4375/9378 [34:56:41<39:14:26, 28.24s/it]
+
+ 47%|█████████████████████████▏ | 4376/9378 [34:57:09<39:39:28, 28.54s/it]
+
+
+
+ 47%|█████████████████████████▏ | 4377/9378 [34:57:38<39:15:08, 28.26s/it]
+
+
+ 47%|█████████████████████████▏ | 4378/9378 [34:58:07<39:35:45, 28.51s/it]
+
+ 47%|█████████████████████████▏ | 4379/9378 [34:58:35<39:28:46, 28.43s/it]
+
+
+ 47%|█████████████████████████▏ | 4380/9378 [34:59:03<39:10:34, 28.22s/it]
+
+ 47%|█████████████████████████▏ | 4381/9378 [34:59:32<39:10:52, 28.23s/it]
+
+ 47%|█████████████████████████▏ | 4382/9378 [34:59:58<39:11:40, 28.24s/it]
+
+
+ 47%|█████████████████████████▏ | 4383/9378 [35:00:27<39:21:26, 28.37s/it]
+
+
+
+ 47%|█████████████████████████▏ | 4384/9378 [35:00:57<39:14:47, 28.29s/it]
+
+
+ 47%|█████████████████████████▏ | 4385/9378 [35:01:25<39:16:17, 28.32s/it]
+
+
+ 47%|█████████████████████████▎ | 4386/9378 [35:01:54<39:21:57, 28.39s/it]
+
+ 47%|█████████████████████████▎ | 4387/9378 [35:02:22<39:22:18, 28.40s/it]
+
+ 47%|█████████████████████████▎ | 4388/9378 [35:02:49<39:21:58, 28.40s/it]
+
+
+
+ 47%|█████████████████████████▎ | 4389/9378 [35:03:19<39:19:16, 28.37s/it]
+{'loss': 1.023, 'grad_norm': 0.466796875, 'learning_rate': 9.84587298690402e-06, 'epoch': 1.39}
+[2024-05-13 21:36:14,753] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-4389/pytorch_model_fsdp.bin
+[2024-05-13 21:36:26,156] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-4389/pytorch_model_fsdp.bin
+[2024-05-13 21:36:50,305] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-4389/optimizer.bin
+ 47%|█████████████████████████▎ | 4389/9378 [35:03:19<39:19:16, 28.37s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 47%|█████████████████████████▎ | 4390/9378 [35:05:57<93:58:41, 67.83s/it]
+
+
+ 47%|█████████████████████████▎ | 4391/9378 [35:06:25<77:35:54, 56.02s/it]
+
+
+
+ 47%|█████████████████████████▎ | 4392/9378 [35:06:55<66:23:17, 47.93s/it]
+
+
+ 47%|█████████████████████████▎ | 4393/9378 [35:07:24<57:58:25, 41.87s/it]
+
+ 47%|█████████████████████████▎ | 4394/9378 [35:07:52<52:29:46, 37.92s/it]
+
+ 47%|█████████████████████████▎ | 4395/9378 [35:08:20<48:21:48, 34.94s/it]
+
+ 47%|█████████████████████████▎ | 4396/9378 [35:08:47<45:38:10, 32.98s/it]
+
+
+ 47%|█████████████████████████▎ | 4397/9378 [35:09:16<43:57:49, 31.77s/it]
+
+
+
+ 47%|█████████████████████████▎ | 4398/9378 [35:09:46<42:18:02, 30.58s/it]
+
+
+ 47%|█████████████████████████▎ | 4399/9378 [35:10:14<41:25:25, 29.95s/it]
+
+
+ 47%|█████████████████████████▎ | 4400/9378 [35:11:04<49:36:22, 35.87s/it]
+
+
+ 47%|█████████████████████████▎ | 4401/9378 [35:11:32<46:28:56, 33.62s/it]
+
+ 47%|█████████████████████████▎ | 4402/9378 [35:12:01<44:22:31, 32.10s/it]
+
+ 47%|█████████████████████████▎ | 4403/9378 [35:12:27<42:49:24, 30.99s/it]
+
+
+ 47%|█████████████████████████▎ | 4404/9378 [35:12:56<41:44:27, 30.21s/it]
+
+
+
+ 47%|█████████████████████████▎ | 4405/9378 [35:13:26<41:00:22, 29.68s/it]
+
+
+ 47%|█████████████████████████▎ | 4406/9378 [35:13:54<40:27:00, 29.29s/it]
+
+ 47%|█████████████████████████▍ | 4407/9378 [35:14:23<40:10:34, 29.10s/it]
+
+ 47%|█████████████████████████▍ | 4408/9378 [35:14:50<39:50:53, 28.86s/it]
+
+
+ 47%|█████████████████████████▍ | 4409/9378 [35:15:19<39:57:08, 28.95s/it]
+
+
+ 47%|█████████████████████████▍ | 4410/9378 [35:15:47<39:39:00, 28.73s/it]
+
+
+
+ 47%|█████████████████████████▍ | 4411/9378 [35:16:16<39:08:07, 28.36s/it]
+
+
+ 47%|█████████████████████████▍ | 4412/9378 [35:16:45<39:09:24, 28.39s/it]
+
+ 47%|█████████████████████████▍ | 4413/9378 [35:17:13<39:08:53, 28.39s/it]
+
+ 47%|█████████████████████████▍ | 4414/9378 [35:17:40<39:06:49, 28.37s/it]
+
+
+ 47%|█████████████████████████▍ | 4415/9378 [35:18:08<39:07:07, 28.38s/it]
+
+
+ 47%|█████████████████████████▍ | 4416/9378 [35:18:36<39:02:34, 28.33s/it]
+
+
+
+ 47%|█████████████████████████▍ | 4417/9378 [35:19:06<39:02:11, 28.33s/it]
+
+
+ 47%|█████████████████████████▍ | 4418/9378 [35:19:35<39:02:04, 28.33s/it]
+
+ 47%|█████████████████████████▍ | 4419/9378 [35:20:03<39:10:50, 28.44s/it]
+
+ 47%|█████████████████████████▍ | 4420/9378 [35:20:30<39:03:34, 28.36s/it]
+
+
+ 47%|█████████████████████████▍ | 4421/9378 [35:20:58<39:07:38, 28.42s/it]
+
+
+
+ 47%|█████████████████████████▍ | 4422/9378 [35:21:29<39:08:41, 28.43s/it]
+
+ 47%|█████████████████████████▍ | 4423/9378 [35:21:57<39:25:24, 28.64s/it]
+
+ 47%|█████████████████████████▍ | 4424/9378 [35:22:25<39:19:03, 28.57s/it]
+
+ 47%|█████████████████████████▍ | 4425/9378 [35:22:54<39:13:57, 28.52s/it]
+
+ 47%|█████████████████████████▍ | 4426/9378 [35:23:21<38:57:56, 28.33s/it]
+
+
+ 47%|█████████████████████████▍ | 4427/9378 [35:23:49<39:07:15, 28.45s/it]
+
+
+
+ 47%|█████████████████████████▍ | 4428/9378 [35:24:19<39:04:08, 28.41s/it]
+
+
+ 47%|█████████████████████████▌ | 4429/9378 [35:24:47<38:41:34, 28.15s/it]
+
+ 47%|█████████████████████████▌ | 4430/9378 [35:25:15<39:02:00, 28.40s/it]
+
+
+ 47%|█████████████████████████▌ | 4431/9378 [35:25:44<38:43:45, 28.18s/it]
+
+ 47%|█████████████████████████▌ | 4432/9378 [35:26:12<38:49:40, 28.26s/it]
+
+ 47%|█████████████████████████▌ | 4433/9378 [35:26:39<38:51:07, 28.28s/it]
+
+
+ 47%|█████████████████████████▌ | 4434/9378 [35:27:09<38:50:36, 28.28s/it]
+
+
+ 47%|█████████████████████████▌ | 4435/9378 [35:27:37<39:05:58, 28.48s/it]
+
+
+ 47%|█████████████████████████▌ | 4436/9378 [35:28:05<38:39:38, 28.16s/it]
+
+
+ 47%|█████████████████████████▌ | 4437/9378 [35:28:33<39:03:27, 28.46s/it]
+
+
+ 47%|█████████████████████████▌ | 4438/9378 [35:29:02<38:39:42, 28.17s/it]
+
+
+ 47%|█████████████████████████▌ | 4439/9378 [35:29:30<38:42:01, 28.21s/it]
+
+
+ 47%|█████████████████████████▌ | 4440/9378 [35:29:59<38:51:12, 28.33s/it]
+
+ 47%|█████████████████████████▌ | 4441/9378 [35:30:27<39:07:59, 28.54s/it]
+
+ 47%|█████████████████████████▌ | 4442/9378 [35:30:54<39:00:39, 28.45s/it]
+
+
+
+ 47%|█████████████████████████▌ | 4443/9378 [35:31:23<38:35:24, 28.15s/it]
+
+
+ 47%|█████████████████████████▌ | 4444/9378 [35:31:52<38:41:17, 28.23s/it]
+
+
+ 47%|█████████████████████████▌ | 4445/9378 [35:32:20<38:43:04, 28.26s/it]
+
+
+ 47%|█████████████████████████▌ | 4446/9378 [35:32:48<38:41:44, 28.25s/it]
+
+
+ 47%|█████████████████████████▌ | 4447/9378 [35:33:17<38:47:49, 28.32s/it]
+
+ 47%|█████████████████████████▌ | 4448/9378 [35:33:44<38:51:50, 28.38s/it]
+
+
+
+ 47%|█████████████████████████▌ | 4449/9378 [35:34:13<38:44:28, 28.30s/it]
+
+ 47%|█████████████████████████▌ | 4450/9378 [35:35:00<46:46:00, 34.16s/it]
+
+
+
+ 47%|█████████████████████████▋ | 4451/9378 [35:35:30<44:22:45, 32.43s/it]
+
+
+ 47%|█████████████████████████▋ | 4452/9378 [35:35:58<42:41:35, 31.20s/it]
+
+
+ 47%|█████████████████████████▋ | 4453/9378 [35:36:26<41:29:34, 30.33s/it]
+
+ 47%|█████████████████████████▋ | 4454/9378 [35:36:55<41:00:14, 29.98s/it]
+
+ 48%|█████████████████████████▋ | 4455/9378 [35:37:23<40:18:45, 29.48s/it]
+
+ 48%|█████████████████████████▋ | 4456/9378 [35:37:51<39:41:13, 29.03s/it]
+
+ 48%|█████████████████████████▋ | 4457/9378 [35:38:19<39:34:43, 28.95s/it]
+
+
+
+ 48%|█████████████████████████▋ | 4458/9378 [35:38:48<39:12:12, 28.69s/it]
+
+
+ 48%|█████████████████████████▋ | 4459/9378 [35:39:17<38:54:12, 28.47s/it]
+
+ 48%|█████████████████████████▋ | 4460/9378 [35:39:45<39:14:07, 28.72s/it]
+
+ 48%|█████████████████████████▋ | 4461/9378 [35:40:13<38:52:38, 28.46s/it]
+
+ 48%|█████████████████████████▋ | 4462/9378 [35:40:42<38:38:48, 28.30s/it]
+
+ 48%|█████████████████████████▋ | 4463/9378 [35:41:10<38:34:25, 28.25s/it]
+
+
+ 48%|█████████████████████████▋ | 4464/9378 [35:41:38<38:38:11, 28.31s/it]
+
+
+ 48%|█████████████████████████▋ | 4465/9378 [35:42:07<38:56:51, 28.54s/it]
+
+
+ 48%|█████████████████████████▋ | 4466/9378 [35:42:35<38:38:47, 28.32s/it]
+
+
+ 48%|█████████████████████████▋ | 4467/9378 [35:43:03<38:39:18, 28.34s/it]
+
+
+ 48%|█████████████████████████▋ | 4468/9378 [35:43:32<38:36:14, 28.30s/it]
+
+
+ 48%|█████████████████████████▋ | 4469/9378 [35:44:00<39:01:27, 28.62s/it]
+
+
+ 48%|█████████████████████████▋ | 4470/9378 [35:44:28<38:30:39, 28.25s/it]
+
+
+ 48%|█████████████████████████▋ | 4471/9378 [35:44:57<38:35:31, 28.31s/it]
+
+
+ 48%|█████████████████████████▊ | 4472/9378 [35:45:25<38:29:24, 28.24s/it]
+
+
+ 48%|█████████████████████████▊ | 4473/9378 [35:45:53<38:34:07, 28.31s/it]
+
+
+ 48%|█████████████████████████▊ | 4474/9378 [35:46:22<38:34:36, 28.32s/it]
+
+
+ 48%|█████████████████████████▊ | 4475/9378 [35:46:51<38:48:38, 28.50s/it]
+
+
+ 48%|█████████████████████████▊ | 4476/9378 [35:47:19<38:41:34, 28.42s/it]
+
+
+ 48%|█████████████████████████▊ | 4477/9378 [35:47:47<38:40:37, 28.41s/it]
+
+
+ 48%|█████████████████████████▊ | 4478/9378 [35:48:16<38:42:08, 28.43s/it]
+
+
+ 48%|█████████████████████████▊ | 4479/9378 [35:48:44<38:35:44, 28.36s/it]
+
+ 48%|█████████████████████████▊ | 4480/9378 [35:49:12<38:39:21, 28.41s/it]
+
+
+ 48%|█████████████████████████▊ | 4481/9378 [35:49:41<38:53:42, 28.59s/it]
+
+
+ 48%|█████████████████████████▊ | 4482/9378 [35:50:10<38:37:26, 28.40s/it]
+
+
+ 48%|█████████████████████████▊ | 4483/9378 [35:50:38<38:35:26, 28.38s/it]
+
+
+ 48%|█████████████████████████▊ | 4484/9378 [35:51:06<38:36:04, 28.39s/it]
+
+ 48%|█████████████████████████▊ | 4485/9378 [35:51:33<38:40:30, 28.46s/it]
+
+
+
+ 48%|█████████████████████████▊ | 4486/9378 [35:52:03<38:31:31, 28.35s/it]
+
+
+ 48%|█████████████████████████▊ | 4487/9378 [35:52:31<38:35:17, 28.40s/it]
+
+
+ 48%|█████████████████████████▊ | 4488/9378 [35:53:00<38:31:25, 28.36s/it]
+
+
+ 48%|█████████████████████████▊ | 4489/9378 [35:53:28<38:34:32, 28.41s/it]
+
+
+ 48%|█████████████████████████▊ | 4490/9378 [35:53:56<38:26:44, 28.32s/it]
+
+ 48%|█████████████████████████▊ | 4491/9378 [35:54:25<38:37:22, 28.45s/it]
+
+ 48%|█████████████████████████▊ | 4492/9378 [35:54:51<38:25:32, 28.31s/it]
+
+
+ 48%|█████████████████████████▊ | 4493/9378 [35:55:20<38:22:34, 28.28s/it]
+
+
+
+ 48%|█████████████████████████▉ | 4494/9378 [35:55:50<38:25:26, 28.32s/it]
+
+
+ 48%|█████████████████████████▉ | 4495/9378 [35:56:18<38:23:30, 28.30s/it]
+
+
+ 48%|█████████████████████████▉ | 4496/9378 [35:56:46<38:24:49, 28.33s/it]
+
+
+ 48%|█████████████████████████▉ | 4497/9378 [35:57:15<38:26:18, 28.35s/it]
+
+ 48%|█████████████████████████▉ | 4498/9378 [35:57:42<38:28:07, 28.38s/it]
+
+
+ 48%|█████████████████████████▉ | 4499/9378 [35:58:10<38:28:10, 28.39s/it]
+
+
+ 48%|█████████████████████████▉ | 4500/9378 [35:59:00<47:18:56, 34.92s/it]
+
+
+
+ 48%|█████████████████████████▉ | 4501/9378 [35:59:30<44:55:14, 33.16s/it]
+
+
+ 48%|█████████████████████████▉ | 4502/9378 [35:59:58<42:40:22, 31.51s/it]
+
+ 48%|█████████████████████████▉ | 4503/9378 [36:00:27<41:38:41, 30.75s/it]
+
+ 48%|█████████████████████████▉ | 4504/9378 [36:00:55<40:37:43, 30.01s/it]
+
+ 48%|█████████████████████████▉ | 4505/9378 [36:01:23<39:57:30, 29.52s/it]
+
+ 48%|█████████████████████████▉ | 4506/9378 [36:01:51<39:27:35, 29.16s/it]
+
+
+
+ 48%|█████████████████████████▉ | 4507/9378 [36:02:20<39:09:25, 28.94s/it]
+
+
+ 48%|█████████████████████████▉ | 4508/9378 [36:02:48<38:35:48, 28.53s/it]
+
+
+ 48%|█████████████████████████▉ | 4509/9378 [36:03:17<38:27:04, 28.43s/it]
+
+
+ 48%|█████████████████████████▉ | 4510/9378 [36:03:45<38:24:20, 28.40s/it]
+
+
+ 48%|█████████████████████████▉ | 4511/9378 [36:04:13<38:21:24, 28.37s/it]
+
+ 48%|█████████████████████████▉ | 4512/9378 [36:04:42<38:38:20, 28.59s/it]
+
+ 48%|█████████████████████████▉ | 4513/9378 [36:05:08<38:12:14, 28.27s/it]
+
+
+ 48%|█████████████████████████▉ | 4514/9378 [36:05:37<38:32:10, 28.52s/it]
+
+
+
+ 48%|█████████████████████████▉ | 4515/9378 [36:06:07<38:07:54, 28.23s/it]
+
+
+ 48%|██████████████████████████ | 4516/9378 [36:06:35<38:11:57, 28.28s/it]
+
+
+ 48%|██████████████████████████ | 4517/9378 [36:07:03<38:07:38, 28.24s/it]
+
+
+ 48%|██████████████████████████ | 4518/9378 [36:07:32<38:10:03, 28.27s/it]
+
+
+ 48%|██████████████████████████ | 4519/9378 [36:08:00<38:10:38, 28.29s/it]
+
+ 48%|██████████████████████████ | 4520/9378 [36:08:28<38:27:54, 28.50s/it]
+
+ 48%|██████████████████████████ | 4521/9378 [36:08:55<38:06:51, 28.25s/it]
+
+
+
+ 48%|██████████████████████████ | 4522/9378 [36:09:25<38:14:49, 28.35s/it]
+
+
+ 48%|██████████████████████████ | 4523/9378 [36:09:53<38:13:54, 28.35s/it]
+
+
+ 48%|██████████████████████████ | 4524/9378 [36:10:22<38:13:47, 28.35s/it]
+
+ 48%|██████████████████████████ | 4525/9378 [36:10:50<38:29:56, 28.56s/it]
+
+ 48%|██████████████████████████ | 4526/9378 [36:11:17<38:05:52, 28.27s/it]
+
+
+ 48%|██████████████████████████ | 4527/9378 [36:11:46<38:24:45, 28.51s/it]
+
+
+
+ 48%|██████████████████████████ | 4528/9378 [36:12:15<38:00:33, 28.21s/it]
+
+
+ 48%|██████████████████████████ | 4529/9378 [36:12:43<38:03:40, 28.26s/it]
+
+
+ 48%|██████████████████████████ | 4530/9378 [36:13:12<38:07:43, 28.31s/it]
+
+
+ 48%|██████████████████████████ | 4531/9378 [36:13:40<38:04:38, 28.28s/it]
+
+
+ 48%|██████████████████████████ | 4532/9378 [36:14:08<38:06:26, 28.31s/it]
+
+ 48%|██████████████████████████ | 4533/9378 [36:14:37<38:25:19, 28.55s/it]
+
+ 48%|██████████████████████████ | 4534/9378 [36:15:05<38:02:26, 28.27s/it]
+
+
+ 48%|██████████████████████████ | 4535/9378 [36:15:34<38:30:51, 28.63s/it]
+
+
+ 48%|██████████████████████████ | 4536/9378 [36:16:02<38:04:05, 28.30s/it]
+
+
+ 48%|██████████████████████████ | 4537/9378 [36:16:30<38:07:20, 28.35s/it]
+
+ 48%|██████████████████████████▏ | 4538/9378 [36:16:59<38:11:43, 28.41s/it]
+
+
+ 48%|██████████████████████████▏ | 4539/9378 [36:17:27<38:01:20, 28.29s/it]
+
+
+ 48%|██████████████████████████▏ | 4540/9378 [36:17:56<38:04:44, 28.33s/it]
+
+
+ 48%|██████████████████████████▏ | 4541/9378 [36:18:24<38:12:50, 28.44s/it]
+
+
+ 48%|██████████████████████████▏ | 4542/9378 [36:18:52<38:05:09, 28.35s/it]
+
+
+ 48%|██████████████████████████▏ | 4543/9378 [36:19:21<38:05:03, 28.36s/it]
+
+
+ 48%|██████████████████████████▏ | 4544/9378 [36:19:49<38:08:26, 28.40s/it]
+
+ 48%|██████████████████████████▏ | 4545/9378 [36:20:17<38:04:26, 28.36s/it]
+
+
+ 48%|██████████████████████████▏ | 4546/9378 [36:20:46<38:21:23, 28.58s/it]
+
+
+ 48%|██████████████████████████▏ | 4547/9378 [36:21:14<37:57:32, 28.29s/it]
+
+
+ 48%|██████████████████████████▏ | 4548/9378 [36:21:43<38:15:00, 28.51s/it]
+
+
+ 49%|██████████████████████████▏ | 4549/9378 [36:22:11<37:54:56, 28.27s/it]
+
+
+ 49%|██████████████████████████▏ | 4550/9378 [36:23:01<46:41:25, 34.81s/it]
+
+
+ 49%|██████████████████████████▏ | 4551/9378 [36:23:29<44:02:46, 32.85s/it]
+
+ 49%|██████████████████████████▏ | 4552/9378 [36:23:58<42:12:17, 31.48s/it]
+
+ 49%|██████████████████████████▏ | 4553/9378 [36:24:26<40:54:15, 30.52s/it]
+
+
+ 49%|██████████████████████████▏ | 4554/9378 [36:24:54<40:07:11, 29.94s/it]
+
+
+ 49%|██████████████████████████▏ | 4555/9378 [36:25:22<39:15:24, 29.30s/it]
+
+
+ 49%|██████████████████████████▏ | 4556/9378 [36:25:51<38:58:54, 29.10s/it]
+
+
+ 49%|██████████████████████████▏ | 4557/9378 [36:26:19<38:36:36, 28.83s/it]
+
+
+ 49%|██████████████████████████▏ | 4558/9378 [36:26:47<38:22:53, 28.67s/it]
+
+ 49%|██████████████████████████▎ | 4559/9378 [36:27:16<38:17:06, 28.60s/it]
+
+
+ 49%|██████████████████████████▎ | 4560/9378 [36:27:44<38:07:40, 28.49s/it]
+
+
+ 49%|██████████████████████████▎ | 4561/9378 [36:28:13<38:08:16, 28.50s/it]
+
+
+ 49%|██████████████████████████▎ | 4562/9378 [36:28:41<38:23:22, 28.70s/it]
+
+
+ 49%|██████████████████████████▎ | 4563/9378 [36:29:09<37:56:18, 28.37s/it]
+
+
+ 49%|██████████████████████████▎ | 4564/9378 [36:29:38<37:56:33, 28.37s/it]
+
+ 49%|██████████████████████████▎ | 4565/9378 [36:30:06<37:56:19, 28.38s/it]
+
+
+ 49%|██████████████████████████▎ | 4566/9378 [36:30:34<37:50:44, 28.31s/it]
+
+
+ 49%|██████████████████████████▎ | 4567/9378 [36:31:03<37:51:31, 28.33s/it]
+
+
+ 49%|██████████████████████████▎ | 4568/9378 [36:31:31<37:52:37, 28.35s/it]
+
+
+ 49%|██████████████████████████▎ | 4569/9378 [36:31:59<37:51:54, 28.35s/it]
+
+
+ 49%|██████████████████████████▎ | 4570/9378 [36:32:28<37:51:02, 28.34s/it]
+
+ 49%|██████████████████████████▎ | 4571/9378 [36:32:56<37:58:34, 28.44s/it]
+
+ 49%|██████████████████████████▎ | 4572/9378 [36:33:24<37:47:33, 28.31s/it]
+
+ 49%|██████████████████████████▎ | 4573/9378 [36:33:51<37:53:34, 28.39s/it]
+
+
+
+ 49%|██████████████████████████▎ | 4574/9378 [36:34:21<37:53:23, 28.39s/it]
+
+
+ 49%|██████████████████████████▎ | 4575/9378 [36:34:50<37:54:41, 28.42s/it]
+
+
+ 49%|██████████████████████████▎ | 4576/9378 [36:35:18<37:54:26, 28.42s/it]
+
+ 49%|██████████████████████████▎ | 4577/9378 [36:35:47<38:09:24, 28.61s/it]
+
+ 49%|██████████████████████████▎ | 4578/9378 [36:36:15<37:40:48, 28.26s/it]
+
+ 49%|██████████████████████████▎ | 4579/9378 [36:36:41<37:40:24, 28.26s/it]
+
+
+
+ 49%|██████████████████████████▎ | 4580/9378 [36:37:12<37:48:32, 28.37s/it]
+
+
+ 49%|██████████████████████████▍ | 4581/9378 [36:37:40<38:00:46, 28.53s/it]
+
+
+ 49%|██████████████████████████▍ | 4582/9378 [36:38:08<37:36:39, 28.23s/it]
+
+
+ 49%|██████████████████████████▍ | 4583/9378 [36:38:37<37:44:18, 28.33s/it]
+
+ 49%|██████████████████████████▍ | 4584/9378 [36:39:05<37:40:01, 28.29s/it]
+
+
+ 49%|██████████████████████████▍ | 4585/9378 [36:39:34<37:53:34, 28.46s/it]
+
+
+ 49%|██████████████████████████▍ | 4586/9378 [36:40:02<37:49:17, 28.41s/it]
+
+
+ 49%|██████████████████████████▍ | 4587/9378 [36:40:30<37:43:17, 28.34s/it]
+
+
+ 49%|██████████████████████████▍ | 4588/9378 [36:40:58<37:39:46, 28.31s/it]
+
+ 49%|██████████████████████████▍ | 4589/9378 [36:41:27<37:49:01, 28.43s/it]
+
+ 49%|██████████████████████████▍ | 4590/9378 [36:41:55<37:52:33, 28.48s/it]
+
+ 49%|██████████████████████████▍ | 4591/9378 [36:42:23<37:46:13, 28.40s/it]
+
+ 49%|██████████████████████████▍ | 4592/9378 [36:42:52<37:27:52, 28.18s/it]
+
+
+ 49%|██████████████████████████▍ | 4593/9378 [36:43:20<37:28:53, 28.20s/it]
+
+
+ 49%|██████████████████████████▍ | 4594/9378 [36:43:48<37:35:13, 28.28s/it]
+
+
+ 49%|██████████████████████████▍ | 4595/9378 [36:44:17<37:35:06, 28.29s/it]
+
+
+ 49%|██████████████████████████▍ | 4596/9378 [36:44:45<37:37:43, 28.33s/it]
+
+ 49%|██████████████████████████▍ | 4597/9378 [36:45:13<37:52:35, 28.52s/it]
+
+
+ 49%|██████████████████████████▍ | 4598/9378 [36:45:42<37:30:01, 28.24s/it]
+{'loss': 0.9168, 'grad_norm': 0.63671875, 'learning_rate': 2.897355456988052e-06, 'epoch': 1.45}
+[2024-05-13 23:18:36,592] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving model to ./buzz-3b-small-v0.6/checkpoint-4598/pytorch_model_fsdp.bin
+[2024-05-13 23:18:48,098] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Model saved to ./buzz-3b-small-v0.6/checkpoint-4598/pytorch_model_fsdp.bin
+[2024-05-13 23:19:09,615] [INFO] [accelerate.utils.fsdp_utils.log:61] [PID:44754] Saving Optimizer state to ./buzz-3b-small-v0.6/checkpoint-4598/optimizer.bin
+ 49%|██████████████████████████▍ | 4598/9378 [36:45:42<37:30:01, 28.24s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+
+ 49%|██████████████████████████▍ | 4599/9378 [36:48:21<89:36:35, 67.50s/it]
+
+ 49%|██████████████████████████▍ | 4600/9378 [36:49:10<82:24:32, 62.09s/it]
+
+
+ 49%|██████████████████████████▍ | 4601/9378 [36:49:39<69:01:13, 52.01s/it]
+
+
+ 49%|██████████████████████████▍ | 4602/9378 [36:50:08<59:44:39, 45.03s/it]
+
+ 49%|██████████████████████████▌ | 4603/9378 [36:50:36<53:24:45, 40.27s/it]
+
+
+ 49%|██████████████████████████▌ | 4604/9378 [36:51:05<48:34:23, 36.63s/it]
+
+
+ 49%|██████████████████████████▌ | 4605/9378 [36:51:33<45:24:45, 34.25s/it]
+
+
+ 49%|██████████████████████████▌ | 4606/9378 [36:52:02<43:08:33, 32.55s/it]
+
+
+ 49%|██████████████████████████▌ | 4607/9378 [36:52:31<41:52:01, 31.59s/it]
+
+
+ 49%|██████████████████████████▌ | 4608/9378 [36:52:59<40:32:00, 30.59s/it]
+
+
+ 49%|██████████████████████████▌ | 4609/9378 [36:53:28<39:27:10, 29.78s/it]
+
+
+ 49%|██████████████████████████▌ | 4610/9378 [36:53:56<38:53:53, 29.37s/it]
+
+
+ 49%|██████████████████████████▌ | 4611/9378 [36:54:24<38:31:43, 29.10s/it]
+
+
+ 49%|██████████████████████████▌ | 4612/9378 [36:54:53<38:18:16, 28.93s/it]
+
+
+ 49%|██████████████████████████▌ | 4613/9378 [36:55:21<38:08:19, 28.81s/it]
+
+
+ 49%|██████████████████████████▌ | 4614/9378 [36:55:50<38:01:28, 28.73s/it]
+
+
+ 49%|██████████████████████████▌ | 4615/9378 [36:56:18<37:54:50, 28.66s/it]
+
+ 49%|██████████████████████████▌ | 4616/9378 [36:56:47<37:49:43, 28.60s/it]
+
+
+ 49%|██████████████████████████▌ | 4617/9378 [36:57:15<37:43:46, 28.53s/it]
+
+
+ 49%|██████████████████████████▌ | 4618/9378 [36:57:44<38:02:45, 28.77s/it]
+
+
+ 49%|██████████████████████████▌ | 4619/9378 [36:58:12<37:36:21, 28.45s/it]
+
+ 49%|██████████████████████████▌ | 4620/9378 [36:58:41<37:41:55, 28.52s/it]
+
+
+ 49%|██████████████████████████▌ | 4621/9378 [36:59:09<37:53:04, 28.67s/it]
+
+
+ 49%|██████████████████████████▌ | 4622/9378 [36:59:38<37:26:58, 28.35s/it]
+
+
+ 49%|██████████████████████████▌ | 4623/9378 [37:00:06<37:32:21, 28.42s/it]
+
+
+ 49%|██████████████████████████▋ | 4624/9378 [37:00:35<37:33:34, 28.44s/it]
+
+
+ 49%|██████████████████████████▋ | 4625/9378 [37:01:03<37:29:26, 28.40s/it]
+
+
+ 49%|██████████████████████████▋ | 4626/9378 [37:01:32<37:30:24, 28.41s/it]
+
+
+ 49%|██████████████████████████▋ | 4627/9378 [37:02:00<37:54:13, 28.72s/it]
+
+ 49%|██████████████████████████▋ | 4628/9378 [37:02:29<37:49:54, 28.67s/it]
+
+ 49%|██████████████████████████▋ | 4629/9378 [37:02:57<37:40:12, 28.56s/it]
+
+ 49%|██████████████████████████▋ | 4630/9378 [37:03:25<37:38:01, 28.53s/it]
+
+
+
+ 49%|██████████████████████████▋ | 4631/9378 [37:03:54<37:36:17, 28.52s/it]
+
+
+ 49%|██████████████████████████▋ | 4632/9378 [37:04:22<37:30:14, 28.45s/it]
+
+
+ 49%|██████████████████████████▋ | 4633/9378 [37:04:51<37:09:14, 28.19s/it]
+
+ 49%|██████████████████████████▋ | 4634/9378 [37:05:19<37:36:54, 28.54s/it]
+
+ 49%|██████████████████████████▋ | 4635/9378 [37:05:48<37:17:52, 28.31s/it]
+
+ 49%|██████████████████████████▋ | 4636/9378 [37:06:14<37:16:36, 28.30s/it]
+
+
+
+ 49%|██████████████████████████▋ | 4637/9378 [37:06:44<37:17:25, 28.32s/it]
+
+
+ 49%|██████████████████████████▋ | 4638/9378 [37:07:13<37:19:18, 28.35s/it]
+
+ 49%|██████████████████████████▋ | 4639/9378 [37:07:41<37:38:21, 28.59s/it]
+
+ 49%|██████████████████████████▋ | 4640/9378 [37:08:10<37:32:55, 28.53s/it]
+
+ 49%|██████████████████████████▋ | 4641/9378 [37:08:38<37:08:17, 28.22s/it]
+
+ 49%|██████████████████████████▋ | 4642/9378 [37:09:04<37:06:32, 28.21s/it]
+
+
+
+ 50%|██████████████████████████▋ | 4643/9378 [37:09:35<37:18:35, 28.37s/it]
+
+
+ 50%|██████████████████████████▋ | 4644/9378 [37:10:03<37:09:18, 28.25s/it]
+
+
+ 50%|██████████████████████████▋ | 4645/9378 [37:10:31<37:03:53, 28.19s/it]
+
+
+ 50%|██████████████████████████▊ | 4646/9378 [37:10:58<36:54:16, 28.08s/it]
+
+ 50%|██████████████████████████▊ | 4647/9378 [37:11:25<36:49:00, 28.02s/it]
+
+
+ 50%|██████████████████████████▊ | 4648/9378 [37:11:53<36:47:56, 28.01s/it]
+
+
+ 50%|██████████████████████████▊ | 4649/9378 [37:12:22<36:41:39, 27.93s/it]
+
+
+ 50%|██████████████████████████▊ | 4650/9378 [37:13:12<45:14:52, 34.45s/it]
+
+
+ 50%|██████████████████████████▊ | 4651/9378 [37:13:40<42:41:14, 32.51s/it]
+
+
+ 50%|██████████████████████████▊ | 4652/9378 [37:14:07<40:59:47, 31.23s/it]
+
+
+ 50%|██████████████████████████▊ | 4653/9378 [37:14:35<39:28:53, 30.08s/it]
+
+ 50%|██████████████████████████▊ | 4654/9378 [37:15:02<38:35:30, 29.41s/it]
+
+
+ 50%|██████████████████████████▊ | 4655/9378 [37:15:29<37:54:52, 28.90s/it]
+
+
+ 50%|██████████████████████████▊ | 4656/9378 [37:15:58<37:43:54, 28.77s/it]
+
+
+ 50%|██████████████████████████▊ | 4657/9378 [37:16:26<37:18:36, 28.45s/it]
+
+
+ 50%|██████████████████████████▊ | 4658/9378 [37:16:53<37:04:40, 28.28s/it]
+
+
+ 50%|██████████████████████████▊ | 4659/9378 [37:17:22<37:08:00, 28.33s/it]
+
+
+ 50%|██████████████████████████▊ | 4660/9378 [37:17:50<37:07:56, 28.33s/it]
+
+
+ 50%|██████████████████████████▊ | 4661/9378 [37:18:18<36:59:06, 28.23s/it]
+
+ 50%|██████████████████████████▊ | 4661/9378 [37:18:19<36:59:06, 28.23s/it]
+
+ 50%|██████████████████████████▊ | 4662/9378 [37:18:47<36:29:23, 27.85s/it]
+
+
+ 50%|██████████████████████████▊ | 4664/9378 [37:19:41<36:23:00, 27.79s/it]
+
+
+ 50%|██████████████████████████▊ | 4665/9378 [37:20:09<36:28:09, 27.86s/it]
+
+
+ 50%|██████████████████████████▊ | 4666/9378 [37:20:37<36:34:39, 27.95s/it]
+
+ 50%|██████████████████████████▊ | 4666/9378 [37:20:39<36:34:39, 27.95s/it]
+
+
+ 50%|██████████████████████████▉ | 4668/9378 [37:21:33<36:32:40, 27.93s/it]
+
+
+ 50%|██████████████████████████▉ | 4669/9378 [37:22:00<36:23:50, 27.83s/it]
+
+
+ 50%|██████████████████████████▉ | 4670/9378 [37:22:29<36:33:08, 27.95s/it]
+
+
+ 50%|██████████████████████████▉ | 4671/9378 [37:22:57<36:39:21, 28.04s/it]
+
+
+ 50%|██████████████████████████▉ | 4672/9378 [37:23:24<36:19:21, 27.79s/it]
+
+
+ 50%|██████████████████████████▉ | 4673/9378 [37:23:52<36:25:18, 27.87s/it]
+
+
+ 50%|██████████████████████████▉ | 4674/9378 [37:24:20<36:25:13, 27.87s/it]
+
+
+ 50%|██████████████████████████▉ | 4675/9378 [37:24:48<36:25:35, 27.88s/it]
+
+
+ 50%|██████████████████████████▉ | 4676/9378 [37:25:16<36:29:47, 27.94s/it]
+
+ 50%|██████████████████████████▉ | 4676/9378 [37:25:18<36:29:47, 27.94s/it]
+
+ 50%|██████████████████████████▉ | 4677/9378 [37:25:45<36:21:31, 27.84s/it]
+
+ 50%|██████████████████████████▉ | 4678/9378 [37:26:13<36:25:20, 27.90s/it]
+
+
+ 50%|██████████████████████████▉ | 4680/9378 [37:27:07<36:20:42, 27.85s/it]
+
+
+ 50%|██████████████████████████▉ | 4681/9378 [37:27:35<36:23:59, 27.90s/it]
+
+
+ 50%|██████████████████████████▉ | 4682/9378 [37:28:03<36:23:52, 27.90s/it]
+
+
+ 50%|██████████████████████████▉ | 4683/9378 [37:28:31<36:23:33, 27.90s/it]
+
+
+ 50%|██████████████████████████▉ | 4684/9378 [37:28:59<36:25:46, 27.94s/it]
+
+
+ 50%|██████████████████████████▉ | 4685/9378 [37:29:27<36:32:02, 28.03s/it]
+
+ 50%|██████████████████████████▉ | 4685/9378 [37:29:29<36:32:02, 28.03s/it]
+
+
+ 50%|██████████████████████████▉ | 4687/9378 [37:30:23<36:19:28, 27.88s/it]
+
+
+ 50%|██████████████████████████▉ | 4688/9378 [37:30:51<36:15:51, 27.84s/it]
+
+
+ 50%|███████████████████████████ | 4689/9378 [37:31:19<36:16:35, 27.85s/it]
+
+
+ 50%|███████████████████████████ | 4690/9378 [37:31:47<36:33:45, 28.08s/it]
+
+
+ 50%|███████████████████████████ | 4691/9378 [37:32:15<36:28:25, 28.01s/it]
+
+
+ 50%|███████████████████████████ | 4692/9378 [37:32:43<36:10:48, 27.80s/it]
+
+ 50%|███████████████████████████ | 4692/9378 [37:32:44<36:10:48, 27.80s/it]
+
+ 50%|███████████████████████████ | 4693/9378 [37:33:12<36:09:10, 27.78s/it]
+
+
+ 50%|███████████████████████████ | 4695/9378 [37:34:06<36:04:54, 27.74s/it]
+
+
+ 50%|███████████████████████████ | 4696/9378 [37:34:34<36:10:36, 27.82s/it]
+
+
+ 50%|███████████████████████████ | 4697/9378 [37:35:02<36:14:49, 27.88s/it]
+
+
+ 50%|███████████████████████████ | 4698/9378 [37:35:30<36:17:58, 27.92s/it]
+
+
+ 50%|███████████████████████████ | 4699/9378 [37:35:58<36:11:03, 27.84s/it]
+
diff --git a/wandb/run-20240512_103207-vfqtzg1a/files/requirements.txt b/wandb/run-20240512_103207-vfqtzg1a/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..54e63626ab0d75f9dec791c2e6499b990c4392b8
--- /dev/null
+++ b/wandb/run-20240512_103207-vfqtzg1a/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash-attn==2.5.5
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240512_103207-vfqtzg1a/files/wandb-metadata.json b/wandb/run-20240512_103207-vfqtzg1a/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..d3c04e991374206fac452e5944dbbea23790ffb8
--- /dev/null
+++ b/wandb/run-20240512_103207-vfqtzg1a/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-12T10:32:08.657622",
+ "startedAt": "2024-05-12T10:32:07.556919",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "2147cf6837e2b90a2ea7045262083cbb0da03858"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "a9815c79a99f",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1406.157375,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1700.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3443.938,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1700.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1700.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.002,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2043.9,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 149.25497436523438
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240512_103207-vfqtzg1a/files/wandb-summary.json b/wandb/run-20240512_103207-vfqtzg1a/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..e16f9474390154f2300fe37d6f9645910a3f5f05
--- /dev/null
+++ b/wandb/run-20240512_103207-vfqtzg1a/files/wandb-summary.json
@@ -0,0 +1 @@
+{"train/loss": 1.0396, "train/grad_norm": 0.470703125, "train/learning_rate": 9.516488704429893e-07, "train/epoch": 1.49, "train/global_step": 4699, "_timestamp": 1715645289.4069505, "_runtime": 135361.8394665718, "_step": 4698, "parameters/_orig_mod._fsdp_wrapped_module.model.embed_tokens.weight": {"_type": "histogram", "values": [2.0, 1.0, 6.0, 8.0, 9.0, 14.0, 25.0, 47.0, 88.0, 314.0, 1316.0, 5746.0, 20883.0, 84104.0, 277676.0, 782856.0, 2006050.0, 4315809.0, 8093267.0, 12518982.0, 16665844.0, 16777216.0, 16176138.0, 11827339.0, 7497016.0, 3872541.0, 1775839.0, 681293.0, 236928.0, 72507.0, 20733.0, 6064.0, 2135.0, 891.0, 401.0, 206.0, 82.0, 64.0, 42.0, 30.0, 13.0, 9.0, 8.0, 4.0, 5.0, 3.0, 3.0, 4.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.08203125, -0.0782012939453125, -0.074371337890625, -0.0705413818359375, -0.06671142578125, -0.0628814697265625, -0.059051513671875, -0.0552215576171875, -0.0513916015625, -0.0475616455078125, -0.043731689453125, -0.0399017333984375, -0.03607177734375, -0.0322418212890625, -0.028411865234375, -0.0245819091796875, -0.020751953125, -0.0169219970703125, -0.013092041015625, -0.0092620849609375, -0.00543212890625, -0.0016021728515625, 0.002227783203125, 0.0060577392578125, 0.0098876953125, 0.0137176513671875, 0.017547607421875, 0.0213775634765625, 0.02520751953125, 0.0290374755859375, 0.032867431640625, 0.0366973876953125, 0.04052734375, 0.0443572998046875, 0.048187255859375, 0.0520172119140625, 0.05584716796875, 0.0596771240234375, 0.063507080078125, 0.0673370361328125, 0.0711669921875, 0.0749969482421875, 0.078826904296875, 0.0826568603515625, 0.08648681640625, 0.0903167724609375, 0.094146728515625, 0.0979766845703125, 0.101806640625, 0.1056365966796875, 0.109466552734375, 0.1132965087890625, 0.11712646484375, 0.1209564208984375, 0.124786376953125, 0.1286163330078125, 0.1324462890625, 0.1362762451171875, 0.140106201171875, 0.1439361572265625, 0.14776611328125, 0.1515960693359375, 0.155426025390625, 0.1592559814453125, 0.1630859375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.0._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 2.0, 1.0, 0.0, 3.0, 5.0, 3.0, 3.0, 13.0, 13.0, 19.0, 25.0, 33.0, 34.0, 58.0, 88.0, 141.0, 276.0, 464.0, 868.0, 1442.0, 2926.0, 6519.0, 16346.0, 49035.0, 188027.0, 897749.0, 9387130.0, 5155089.0, 820347.0, 178426.0, 44064.0, 15983.0, 5973.0, 2729.0, 1481.0, 797.0, 424.0, 220.0, 148.0, 92.0, 71.0, 50.0, 24.0, 18.0, 14.0, 14.0, 7.0, 2.0, 3.0, 5.0, 2.0, 0.0, 2.0, 1.0, 1.0, 0.0, 2.0], "bins": [-0.7734375, -0.75067138671875, -0.7279052734375, -0.70513916015625, -0.682373046875, -0.65960693359375, -0.6368408203125, -0.61407470703125, -0.59130859375, -0.56854248046875, -0.5457763671875, -0.52301025390625, -0.500244140625, -0.47747802734375, -0.4547119140625, -0.43194580078125, -0.4091796875, -0.38641357421875, -0.3636474609375, -0.34088134765625, -0.318115234375, -0.29534912109375, -0.2725830078125, -0.24981689453125, -0.22705078125, -0.20428466796875, -0.1815185546875, -0.15875244140625, -0.135986328125, -0.11322021484375, -0.0904541015625, -0.06768798828125, -0.044921875, -0.02215576171875, 0.0006103515625, 0.02337646484375, 0.046142578125, 0.06890869140625, 0.0916748046875, 0.11444091796875, 0.13720703125, 0.15997314453125, 0.1827392578125, 0.20550537109375, 0.228271484375, 0.25103759765625, 0.2738037109375, 0.29656982421875, 0.3193359375, 0.34210205078125, 0.3648681640625, 0.38763427734375, 0.410400390625, 0.43316650390625, 0.4559326171875, 0.47869873046875, 0.50146484375, 0.52423095703125, 0.5469970703125, 0.56976318359375, 0.592529296875, 0.61529541015625, 0.6380615234375, 0.66082763671875, 0.68359375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.0._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [3.0, 5.0, 1.0, 2.0, 1.0, 5.0, 3.0, 4.0, 9.0, 21.0, 22.0, 18.0, 76.0, 69.0, 93.0, 134.0, 277.0, 435.0, 736.0, 1234.0, 2154.0, 3716.0, 7577.0, 16016.0, 38378.0, 110340.0, 344922.0, 2235734.0, 990752.0, 286577.0, 92032.0, 33250.0, 14808.0, 6744.0, 3376.0, 1898.0, 1090.0, 706.0, 392.0, 246.0, 138.0, 104.0, 61.0, 50.0, 32.0, 20.0, 11.0, 12.0, 5.0, 0.0, 2.0, 3.0, 1.0, 1.0, 4.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.6015625, -0.58001708984375, -0.5584716796875, -0.53692626953125, -0.515380859375, -0.49383544921875, -0.4722900390625, -0.45074462890625, -0.42919921875, -0.40765380859375, -0.3861083984375, -0.36456298828125, -0.343017578125, -0.32147216796875, -0.2999267578125, -0.27838134765625, -0.2568359375, -0.23529052734375, -0.2137451171875, -0.19219970703125, -0.170654296875, -0.14910888671875, -0.1275634765625, -0.10601806640625, -0.08447265625, -0.06292724609375, -0.0413818359375, -0.01983642578125, 0.001708984375, 0.02325439453125, 0.0447998046875, 0.06634521484375, 0.087890625, 0.10943603515625, 0.1309814453125, 0.15252685546875, 0.174072265625, 0.19561767578125, 0.2171630859375, 0.23870849609375, 0.26025390625, 0.28179931640625, 0.3033447265625, 0.32489013671875, 0.346435546875, 0.36798095703125, 0.3895263671875, 0.41107177734375, 0.4326171875, 0.45416259765625, 0.4757080078125, 0.49725341796875, 0.518798828125, 0.54034423828125, 0.5618896484375, 0.58343505859375, 0.60498046875, 0.62652587890625, 0.6480712890625, 0.66961669921875, 0.691162109375, 0.71270751953125, 0.7342529296875, 0.75579833984375, 0.77734375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.0._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 1.0, 2.0, 3.0, 4.0, 5.0, 10.0, 12.0, 25.0, 42.0, 55.0, 79.0, 132.0, 210.0, 306.0, 474.0, 752.0, 1033.0, 1458.0, 2060.0, 3362.0, 4498.0, 6775.0, 8867.0, 12790.0, 16612.0, 23344.0, 31347.0, 44330.0, 57107.0, 86352.0, 147074.0, 123636.0, 73820.0, 51885.0, 38183.0, 28383.0, 20614.0, 15682.0, 11018.0, 8538.0, 5835.0, 4079.0, 2774.0, 2063.0, 1315.0, 912.0, 620.0, 448.0, 257.0, 187.0, 111.0, 85.0, 36.0, 31.0, 17.0, 15.0, 7.0, 0.0, 2.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.043701171875, -0.042327880859375, -0.04095458984375, -0.039581298828125, -0.0382080078125, -0.036834716796875, -0.03546142578125, -0.034088134765625, -0.03271484375, -0.031341552734375, -0.02996826171875, -0.028594970703125, -0.0272216796875, -0.025848388671875, -0.02447509765625, -0.023101806640625, -0.021728515625, -0.020355224609375, -0.01898193359375, -0.017608642578125, -0.0162353515625, -0.014862060546875, -0.01348876953125, -0.012115478515625, -0.0107421875, -0.009368896484375, -0.00799560546875, -0.006622314453125, -0.0052490234375, -0.003875732421875, -0.00250244140625, -0.001129150390625, 0.000244140625, 0.001617431640625, 0.00299072265625, 0.004364013671875, 0.0057373046875, 0.007110595703125, 0.00848388671875, 0.009857177734375, 0.01123046875, 0.012603759765625, 0.01397705078125, 0.015350341796875, 0.0167236328125, 0.018096923828125, 0.01947021484375, 0.020843505859375, 0.022216796875, 0.023590087890625, 0.02496337890625, 0.026336669921875, 0.0277099609375, 0.029083251953125, 0.03045654296875, 0.031829833984375, 0.033203125, 0.034576416015625, 0.03594970703125, 0.037322998046875, 0.0386962890625, 0.040069580078125, 0.04144287109375, 0.042816162109375, 0.044189453125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.1._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 2.0, 1.0, 1.0, 4.0, 6.0, 14.0, 19.0, 30.0, 51.0, 104.0, 186.0, 285.0, 686.0, 1703.0, 4899.0, 16039.0, 55044.0, 201305.0, 660293.0, 2166512.0, 6703643.0, 4818113.0, 1500672.0, 458901.0, 134978.0, 37016.0, 11041.0, 3301.0, 1282.0, 489.0, 261.0, 134.0, 84.0, 44.0, 28.0, 10.0, 10.0, 3.0, 4.0, 3.0, 2.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.5390625, -0.523590087890625, -0.50811767578125, -0.492645263671875, -0.4771728515625, -0.461700439453125, -0.44622802734375, -0.430755615234375, -0.415283203125, -0.399810791015625, -0.38433837890625, -0.368865966796875, -0.3533935546875, -0.337921142578125, -0.32244873046875, -0.306976318359375, -0.29150390625, -0.276031494140625, -0.26055908203125, -0.245086669921875, -0.2296142578125, -0.214141845703125, -0.19866943359375, -0.183197021484375, -0.167724609375, -0.152252197265625, -0.13677978515625, -0.121307373046875, -0.1058349609375, -0.090362548828125, -0.07489013671875, -0.059417724609375, -0.0439453125, -0.028472900390625, -0.01300048828125, 0.002471923828125, 0.0179443359375, 0.033416748046875, 0.04888916015625, 0.064361572265625, 0.079833984375, 0.095306396484375, 0.11077880859375, 0.126251220703125, 0.1417236328125, 0.157196044921875, 0.17266845703125, 0.188140869140625, 0.20361328125, 0.219085693359375, 0.23455810546875, 0.250030517578125, 0.2655029296875, 0.280975341796875, 0.29644775390625, 0.311920166015625, 0.327392578125, 0.342864990234375, 0.35833740234375, 0.373809814453125, 0.3892822265625, 0.404754638671875, 0.42022705078125, 0.435699462890625, 0.451171875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.1._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 3.0, 3.0, 6.0, 2.0, 4.0, 12.0, 13.0, 15.0, 42.0, 51.0, 82.0, 131.0, 172.0, 285.0, 437.0, 717.0, 979.0, 1743.0, 2961.0, 4933.0, 8410.0, 15376.0, 24949.0, 48232.0, 86473.0, 147258.0, 262479.0, 439742.0, 819771.0, 1040789.0, 541249.0, 320264.0, 182165.0, 106352.0, 61465.0, 32835.0, 18084.0, 10604.0, 6004.0, 3614.0, 2005.0, 1260.0, 826.0, 523.0, 311.0, 238.0, 148.0, 90.0, 59.0, 44.0, 32.0, 22.0, 19.0, 6.0, 3.0, 4.0, 4.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.337890625, -0.327392578125, -0.31689453125, -0.306396484375, -0.2958984375, -0.285400390625, -0.27490234375, -0.264404296875, -0.25390625, -0.243408203125, -0.23291015625, -0.222412109375, -0.2119140625, -0.201416015625, -0.19091796875, -0.180419921875, -0.169921875, -0.159423828125, -0.14892578125, -0.138427734375, -0.1279296875, -0.117431640625, -0.10693359375, -0.096435546875, -0.0859375, -0.075439453125, -0.06494140625, -0.054443359375, -0.0439453125, -0.033447265625, -0.02294921875, -0.012451171875, -0.001953125, 0.008544921875, 0.01904296875, 0.029541015625, 0.0400390625, 0.050537109375, 0.06103515625, 0.071533203125, 0.08203125, 0.092529296875, 0.10302734375, 0.113525390625, 0.1240234375, 0.134521484375, 0.14501953125, 0.155517578125, 0.166015625, 0.176513671875, 0.18701171875, 0.197509765625, 0.2080078125, 0.218505859375, 0.22900390625, 0.239501953125, 0.25, 0.260498046875, 0.27099609375, 0.281494140625, 0.2919921875, 0.302490234375, 0.31298828125, 0.323486328125, 0.333984375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.1._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [4.0, 1.0, 0.0, 5.0, 11.0, 11.0, 23.0, 63.0, 60.0, 135.0, 209.0, 369.0, 528.0, 840.0, 1230.0, 1982.0, 2855.0, 3957.0, 5639.0, 7168.0, 10132.0, 13753.0, 16819.0, 21521.0, 26269.0, 30539.0, 37706.0, 42161.0, 47703.0, 53823.0, 58220.0, 65528.0, 59506.0, 54338.0, 49076.0, 42947.0, 37663.0, 32270.0, 25673.0, 21805.0, 17582.0, 13313.0, 10433.0, 7976.0, 5066.0, 4224.0, 2852.0, 1902.0, 1364.0, 946.0, 589.0, 348.0, 227.0, 141.0, 76.0, 38.0, 30.0, 9.0, 9.0, 5.0, 3.0, 2.0, 1.0, 2.0], "bins": [-0.041748046875, -0.040424346923828125, -0.03910064697265625, -0.037776947021484375, -0.0364532470703125, -0.035129547119140625, -0.03380584716796875, -0.032482147216796875, -0.031158447265625, -0.029834747314453125, -0.02851104736328125, -0.027187347412109375, -0.0258636474609375, -0.024539947509765625, -0.02321624755859375, -0.021892547607421875, -0.02056884765625, -0.019245147705078125, -0.01792144775390625, -0.016597747802734375, -0.0152740478515625, -0.013950347900390625, -0.01262664794921875, -0.011302947998046875, -0.009979248046875, -0.008655548095703125, -0.00733184814453125, -0.006008148193359375, -0.0046844482421875, -0.003360748291015625, -0.00203704833984375, -0.000713348388671875, 0.0006103515625, 0.001934051513671875, 0.00325775146484375, 0.004581451416015625, 0.0059051513671875, 0.007228851318359375, 0.00855255126953125, 0.009876251220703125, 0.011199951171875, 0.012523651123046875, 0.01384735107421875, 0.015171051025390625, 0.0164947509765625, 0.017818450927734375, 0.01914215087890625, 0.020465850830078125, 0.02178955078125, 0.023113250732421875, 0.02443695068359375, 0.025760650634765625, 0.0270843505859375, 0.028408050537109375, 0.02973175048828125, 0.031055450439453125, 0.032379150390625, 0.033702850341796875, 0.03502655029296875, 0.036350250244140625, 0.0376739501953125, 0.038997650146484375, 0.04032135009765625, 0.041645050048828125, 0.04296875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.2._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [4.0, 0.0, 3.0, 0.0, 3.0, 5.0, 15.0, 14.0, 31.0, 24.0, 51.0, 60.0, 78.0, 136.0, 147.0, 205.0, 374.0, 479.0, 781.0, 1283.0, 2573.0, 6083.0, 12280.0, 27728.0, 66217.0, 131030.0, 265158.0, 490235.0, 930083.0, 1621927.0, 2686406.0, 3573757.0, 2929350.0, 1826449.0, 1040983.0, 576085.0, 292895.0, 156192.0, 74592.0, 34179.0, 15059.0, 6822.0, 3285.0, 1547.0, 909.0, 532.0, 359.0, 230.0, 159.0, 92.0, 112.0, 56.0, 54.0, 33.0, 24.0, 17.0, 8.0, 8.0, 8.0, 0.0, 2.0, 0.0, 2.0, 3.0], "bins": [-0.234375, -0.226959228515625, -0.21954345703125, -0.212127685546875, -0.2047119140625, -0.197296142578125, -0.18988037109375, -0.182464599609375, -0.175048828125, -0.167633056640625, -0.16021728515625, -0.152801513671875, -0.1453857421875, -0.137969970703125, -0.13055419921875, -0.123138427734375, -0.11572265625, -0.108306884765625, -0.10089111328125, -0.093475341796875, -0.0860595703125, -0.078643798828125, -0.07122802734375, -0.063812255859375, -0.056396484375, -0.048980712890625, -0.04156494140625, -0.034149169921875, -0.0267333984375, -0.019317626953125, -0.01190185546875, -0.004486083984375, 0.0029296875, 0.010345458984375, 0.01776123046875, 0.025177001953125, 0.0325927734375, 0.040008544921875, 0.04742431640625, 0.054840087890625, 0.062255859375, 0.069671630859375, 0.07708740234375, 0.084503173828125, 0.0919189453125, 0.099334716796875, 0.10675048828125, 0.114166259765625, 0.12158203125, 0.128997802734375, 0.13641357421875, 0.143829345703125, 0.1512451171875, 0.158660888671875, 0.16607666015625, 0.173492431640625, 0.180908203125, 0.188323974609375, 0.19573974609375, 0.203155517578125, 0.2105712890625, 0.217987060546875, 0.22540283203125, 0.232818603515625, 0.240234375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.2._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 2.0, 1.0, 4.0, 7.0, 12.0, 17.0, 18.0, 36.0, 40.0, 52.0, 52.0, 79.0, 145.0, 236.0, 351.0, 624.0, 1177.0, 2383.0, 4216.0, 8753.0, 15093.0, 27493.0, 49545.0, 86583.0, 141129.0, 236840.0, 401150.0, 659110.0, 855344.0, 682527.0, 419783.0, 250450.0, 146020.0, 89562.0, 51306.0, 29159.0, 15819.0, 9147.0, 4710.0, 2279.0, 1326.0, 583.0, 401.0, 224.0, 159.0, 105.0, 69.0, 63.0, 27.0, 33.0, 18.0, 17.0, 6.0, 5.0, 5.0, 2.0, 3.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.306640625, -0.296600341796875, -0.28656005859375, -0.276519775390625, -0.2664794921875, -0.256439208984375, -0.24639892578125, -0.236358642578125, -0.226318359375, -0.216278076171875, -0.20623779296875, -0.196197509765625, -0.1861572265625, -0.176116943359375, -0.16607666015625, -0.156036376953125, -0.14599609375, -0.135955810546875, -0.12591552734375, -0.115875244140625, -0.1058349609375, -0.095794677734375, -0.08575439453125, -0.075714111328125, -0.065673828125, -0.055633544921875, -0.04559326171875, -0.035552978515625, -0.0255126953125, -0.015472412109375, -0.00543212890625, 0.004608154296875, 0.0146484375, 0.024688720703125, 0.03472900390625, 0.044769287109375, 0.0548095703125, 0.064849853515625, 0.07489013671875, 0.084930419921875, 0.094970703125, 0.105010986328125, 0.11505126953125, 0.125091552734375, 0.1351318359375, 0.145172119140625, 0.15521240234375, 0.165252685546875, 0.17529296875, 0.185333251953125, 0.19537353515625, 0.205413818359375, 0.2154541015625, 0.225494384765625, 0.23553466796875, 0.245574951171875, 0.255615234375, 0.265655517578125, 0.27569580078125, 0.285736083984375, 0.2957763671875, 0.305816650390625, 0.31585693359375, 0.325897216796875, 0.3359375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.2._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 2.0, 7.0, 10.0, 20.0, 34.0, 59.0, 92.0, 166.0, 256.0, 512.0, 771.0, 1056.0, 1788.0, 2350.0, 3972.0, 4981.0, 7571.0, 10091.0, 13504.0, 17637.0, 21754.0, 26869.0, 32953.0, 38874.0, 44362.0, 50774.0, 56658.0, 58476.0, 61567.0, 57681.0, 55673.0, 49008.0, 43290.0, 38068.0, 31783.0, 25790.0, 20855.0, 16705.0, 12694.0, 9375.0, 6796.0, 5006.0, 3520.0, 2253.0, 1527.0, 880.0, 666.0, 381.0, 253.0, 125.0, 75.0, 43.0, 29.0, 11.0, 12.0, 4.0, 4.0, 3.0], "bins": [-0.044189453125, -0.042903900146484375, -0.04161834716796875, -0.040332794189453125, -0.0390472412109375, -0.037761688232421875, -0.03647613525390625, -0.035190582275390625, -0.033905029296875, -0.032619476318359375, -0.03133392333984375, -0.030048370361328125, -0.0287628173828125, -0.027477264404296875, -0.02619171142578125, -0.024906158447265625, -0.02362060546875, -0.022335052490234375, -0.02104949951171875, -0.019763946533203125, -0.0184783935546875, -0.017192840576171875, -0.01590728759765625, -0.014621734619140625, -0.013336181640625, -0.012050628662109375, -0.01076507568359375, -0.009479522705078125, -0.0081939697265625, -0.006908416748046875, -0.00562286376953125, -0.004337310791015625, -0.0030517578125, -0.001766204833984375, -0.00048065185546875, 0.000804901123046875, 0.0020904541015625, 0.003376007080078125, 0.00466156005859375, 0.005947113037109375, 0.007232666015625, 0.008518218994140625, 0.00980377197265625, 0.011089324951171875, 0.0123748779296875, 0.013660430908203125, 0.01494598388671875, 0.016231536865234375, 0.01751708984375, 0.018802642822265625, 0.02008819580078125, 0.021373748779296875, 0.0226593017578125, 0.023944854736328125, 0.02523040771484375, 0.026515960693359375, 0.027801513671875, 0.029087066650390625, 0.03037261962890625, 0.031658172607421875, 0.0329437255859375, 0.034229278564453125, 0.03551483154296875, 0.036800384521484375, 0.0380859375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.3._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 2.0, 1.0, 1.0, 5.0, 2.0, 3.0, 6.0, 10.0, 10.0, 14.0, 27.0, 51.0, 60.0, 142.0, 261.0, 495.0, 897.0, 1613.0, 3582.0, 7069.0, 14740.0, 29770.0, 63718.0, 119196.0, 218878.0, 398372.0, 723198.0, 1133177.0, 1810865.0, 2529629.0, 2895361.0, 2458573.0, 1762328.0, 1092722.0, 694704.0, 379492.0, 208556.0, 113770.0, 60737.0, 28032.0, 13871.0, 6530.0, 3273.0, 1526.0, 875.0, 461.0, 278.0, 124.0, 79.0, 46.0, 25.0, 20.0, 14.0, 7.0, 5.0, 2.0, 2.0, 1.0, 2.0, 0.0, 0.0, 2.0], "bins": [-0.2080078125, -0.20159912109375, -0.1951904296875, -0.18878173828125, -0.182373046875, -0.17596435546875, -0.1695556640625, -0.16314697265625, -0.15673828125, -0.15032958984375, -0.1439208984375, -0.13751220703125, -0.131103515625, -0.12469482421875, -0.1182861328125, -0.11187744140625, -0.10546875, -0.09906005859375, -0.0926513671875, -0.08624267578125, -0.079833984375, -0.07342529296875, -0.0670166015625, -0.06060791015625, -0.05419921875, -0.04779052734375, -0.0413818359375, -0.03497314453125, -0.028564453125, -0.02215576171875, -0.0157470703125, -0.00933837890625, -0.0029296875, 0.00347900390625, 0.0098876953125, 0.01629638671875, 0.022705078125, 0.02911376953125, 0.0355224609375, 0.04193115234375, 0.04833984375, 0.05474853515625, 0.0611572265625, 0.06756591796875, 0.073974609375, 0.08038330078125, 0.0867919921875, 0.09320068359375, 0.099609375, 0.10601806640625, 0.1124267578125, 0.11883544921875, 0.125244140625, 0.13165283203125, 0.1380615234375, 0.14447021484375, 0.15087890625, 0.15728759765625, 0.1636962890625, 0.17010498046875, 0.176513671875, 0.18292236328125, 0.1893310546875, 0.19573974609375, 0.2021484375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.3._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 6.0, 16.0, 22.0, 34.0, 89.0, 133.0, 222.0, 404.0, 732.0, 1430.0, 2497.0, 4308.0, 7906.0, 12968.0, 21478.0, 38487.0, 65480.0, 106307.0, 160357.0, 265095.0, 376650.0, 532918.0, 653385.0, 607938.0, 459586.0, 323662.0, 207980.0, 135860.0, 86293.0, 51458.0, 28358.0, 18325.0, 10083.0, 6287.0, 3255.0, 1906.0, 1086.0, 601.0, 280.0, 187.0, 111.0, 63.0, 21.0, 15.0, 8.0, 3.0, 0.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.26953125, -0.26104736328125, -0.2525634765625, -0.24407958984375, -0.235595703125, -0.22711181640625, -0.2186279296875, -0.21014404296875, -0.20166015625, -0.19317626953125, -0.1846923828125, -0.17620849609375, -0.167724609375, -0.15924072265625, -0.1507568359375, -0.14227294921875, -0.1337890625, -0.12530517578125, -0.1168212890625, -0.10833740234375, -0.099853515625, -0.09136962890625, -0.0828857421875, -0.07440185546875, -0.06591796875, -0.05743408203125, -0.0489501953125, -0.04046630859375, -0.031982421875, -0.02349853515625, -0.0150146484375, -0.00653076171875, 0.001953125, 0.01043701171875, 0.0189208984375, 0.02740478515625, 0.035888671875, 0.04437255859375, 0.0528564453125, 0.06134033203125, 0.06982421875, 0.07830810546875, 0.0867919921875, 0.09527587890625, 0.103759765625, 0.11224365234375, 0.1207275390625, 0.12921142578125, 0.1376953125, 0.14617919921875, 0.1546630859375, 0.16314697265625, 0.171630859375, 0.18011474609375, 0.1885986328125, 0.19708251953125, 0.20556640625, 0.21405029296875, 0.2225341796875, 0.23101806640625, 0.239501953125, 0.24798583984375, 0.2564697265625, 0.26495361328125, 0.2734375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.3._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 0.0, 6.0, 11.0, 16.0, 52.0, 76.0, 155.0, 275.0, 509.0, 850.0, 1503.0, 2176.0, 3400.0, 4968.0, 7613.0, 10753.0, 14465.0, 20232.0, 25703.0, 30903.0, 38907.0, 45025.0, 53268.0, 57089.0, 62514.0, 64842.0, 63131.0, 58756.0, 53271.0, 47087.0, 40107.0, 33454.0, 25874.0, 21743.0, 16027.0, 11170.0, 8382.0, 5480.0, 3619.0, 2598.0, 1513.0, 910.0, 584.0, 312.0, 146.0, 98.0, 47.0, 26.0, 16.0, 7.0, 5.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.04150390625, -0.04000091552734375, -0.0384979248046875, -0.03699493408203125, -0.035491943359375, -0.03398895263671875, -0.0324859619140625, -0.03098297119140625, -0.02947998046875, -0.02797698974609375, -0.0264739990234375, -0.02497100830078125, -0.023468017578125, -0.02196502685546875, -0.0204620361328125, -0.01895904541015625, -0.0174560546875, -0.01595306396484375, -0.0144500732421875, -0.01294708251953125, -0.011444091796875, -0.00994110107421875, -0.0084381103515625, -0.00693511962890625, -0.00543212890625, -0.00392913818359375, -0.0024261474609375, -0.00092315673828125, 0.000579833984375, 0.00208282470703125, 0.0035858154296875, 0.00508880615234375, 0.006591796875, 0.00809478759765625, 0.0095977783203125, 0.01110076904296875, 0.012603759765625, 0.01410675048828125, 0.0156097412109375, 0.01711273193359375, 0.01861572265625, 0.02011871337890625, 0.0216217041015625, 0.02312469482421875, 0.024627685546875, 0.02613067626953125, 0.0276336669921875, 0.02913665771484375, 0.0306396484375, 0.03214263916015625, 0.0336456298828125, 0.03514862060546875, 0.036651611328125, 0.03815460205078125, 0.0396575927734375, 0.04116058349609375, 0.04266357421875, 0.04416656494140625, 0.0456695556640625, 0.04717254638671875, 0.048675537109375, 0.05017852783203125, 0.0516815185546875, 0.05318450927734375, 0.0546875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.4._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 3.0, 2.0, 2.0, 8.0, 9.0, 11.0, 21.0, 15.0, 23.0, 48.0, 45.0, 56.0, 81.0, 106.0, 135.0, 215.0, 385.0, 852.0, 2093.0, 5297.0, 14485.0, 44923.0, 125393.0, 322096.0, 820781.0, 1796678.0, 3453384.0, 4432121.0, 3087312.0, 1543868.0, 687893.0, 275832.0, 105920.0, 36427.0, 12370.0, 4600.0, 1785.0, 788.0, 368.0, 204.0, 144.0, 100.0, 77.0, 65.0, 51.0, 35.0, 24.0, 20.0, 18.0, 14.0, 7.0, 8.0, 5.0, 3.0, 3.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.306640625, -0.296875, -0.287109375, -0.27734375, -0.267578125, -0.2578125, -0.248046875, -0.23828125, -0.228515625, -0.21875, -0.208984375, -0.19921875, -0.189453125, -0.1796875, -0.169921875, -0.16015625, -0.150390625, -0.140625, -0.130859375, -0.12109375, -0.111328125, -0.1015625, -0.091796875, -0.08203125, -0.072265625, -0.0625, -0.052734375, -0.04296875, -0.033203125, -0.0234375, -0.013671875, -0.00390625, 0.005859375, 0.015625, 0.025390625, 0.03515625, 0.044921875, 0.0546875, 0.064453125, 0.07421875, 0.083984375, 0.09375, 0.103515625, 0.11328125, 0.123046875, 0.1328125, 0.142578125, 0.15234375, 0.162109375, 0.171875, 0.181640625, 0.19140625, 0.201171875, 0.2109375, 0.220703125, 0.23046875, 0.240234375, 0.25, 0.259765625, 0.26953125, 0.279296875, 0.2890625, 0.298828125, 0.30859375, 0.318359375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.4._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 3.0, 3.0, 3.0, 6.0, 10.0, 15.0, 13.0, 27.0, 29.0, 45.0, 37.0, 44.0, 62.0, 112.0, 125.0, 193.0, 400.0, 803.0, 1664.0, 3748.0, 6953.0, 15522.0, 29202.0, 59786.0, 110118.0, 186974.0, 327376.0, 501441.0, 702127.0, 750043.0, 580351.0, 388983.0, 234131.0, 138732.0, 76937.0, 39348.0, 19244.0, 10193.0, 4667.0, 2341.0, 1053.0, 533.0, 253.0, 172.0, 130.0, 90.0, 64.0, 43.0, 33.0, 32.0, 20.0, 19.0, 14.0, 8.0, 9.0, 7.0, 3.0, 2.0, 1.0, 3.0, 1.0], "bins": [-0.322265625, -0.312255859375, -0.30224609375, -0.292236328125, -0.2822265625, -0.272216796875, -0.26220703125, -0.252197265625, -0.2421875, -0.232177734375, -0.22216796875, -0.212158203125, -0.2021484375, -0.192138671875, -0.18212890625, -0.172119140625, -0.162109375, -0.152099609375, -0.14208984375, -0.132080078125, -0.1220703125, -0.112060546875, -0.10205078125, -0.092041015625, -0.08203125, -0.072021484375, -0.06201171875, -0.052001953125, -0.0419921875, -0.031982421875, -0.02197265625, -0.011962890625, -0.001953125, 0.008056640625, 0.01806640625, 0.028076171875, 0.0380859375, 0.048095703125, 0.05810546875, 0.068115234375, 0.078125, 0.088134765625, 0.09814453125, 0.108154296875, 0.1181640625, 0.128173828125, 0.13818359375, 0.148193359375, 0.158203125, 0.168212890625, 0.17822265625, 0.188232421875, 0.1982421875, 0.208251953125, 0.21826171875, 0.228271484375, 0.23828125, 0.248291015625, 0.25830078125, 0.268310546875, 0.2783203125, 0.288330078125, 0.29833984375, 0.308349609375, 0.318359375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.4._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [4.0, 1.0, 3.0, 3.0, 13.0, 22.0, 41.0, 58.0, 120.0, 202.0, 254.0, 486.0, 864.0, 1302.0, 2204.0, 3158.0, 4717.0, 6995.0, 9541.0, 14152.0, 18413.0, 24069.0, 31553.0, 39345.0, 46975.0, 55386.0, 61385.0, 67913.0, 70690.0, 67221.0, 61030.0, 53612.0, 45790.0, 38350.0, 30950.0, 23815.0, 17994.0, 12434.0, 9243.0, 6761.0, 4398.0, 3138.0, 1847.0, 1228.0, 869.0, 477.0, 278.0, 135.0, 105.0, 70.0, 23.0, 21.0, 6.0, 5.0, 1.0, 7.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.053466796875, -0.051586151123046875, -0.04970550537109375, -0.047824859619140625, -0.0459442138671875, -0.044063568115234375, -0.04218292236328125, -0.040302276611328125, -0.038421630859375, -0.036540985107421875, -0.03466033935546875, -0.032779693603515625, -0.0308990478515625, -0.029018402099609375, -0.02713775634765625, -0.025257110595703125, -0.02337646484375, -0.021495819091796875, -0.01961517333984375, -0.017734527587890625, -0.0158538818359375, -0.013973236083984375, -0.01209259033203125, -0.010211944580078125, -0.008331298828125, -0.006450653076171875, -0.00457000732421875, -0.002689361572265625, -0.0008087158203125, 0.001071929931640625, 0.00295257568359375, 0.004833221435546875, 0.0067138671875, 0.008594512939453125, 0.01047515869140625, 0.012355804443359375, 0.0142364501953125, 0.016117095947265625, 0.01799774169921875, 0.019878387451171875, 0.021759033203125, 0.023639678955078125, 0.02552032470703125, 0.027400970458984375, 0.0292816162109375, 0.031162261962890625, 0.03304290771484375, 0.034923553466796875, 0.03680419921875, 0.038684844970703125, 0.04056549072265625, 0.042446136474609375, 0.0443267822265625, 0.046207427978515625, 0.04808807373046875, 0.049968719482421875, 0.051849365234375, 0.053730010986328125, 0.05561065673828125, 0.057491302490234375, 0.0593719482421875, 0.061252593994140625, 0.06313323974609375, 0.06501388549804688, 0.06689453125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.5._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 0.0, 2.0, 3.0, 5.0, 10.0, 17.0, 23.0, 25.0, 37.0, 57.0, 73.0, 89.0, 157.0, 217.0, 297.0, 451.0, 740.0, 1348.0, 2271.0, 4854.0, 10237.0, 20996.0, 49857.0, 98647.0, 196090.0, 390514.0, 758084.0, 1298931.0, 2212235.0, 3196774.0, 3237040.0, 2299692.0, 1377653.0, 797063.0, 408728.0, 209197.0, 108447.0, 50871.0, 23491.0, 10610.0, 5128.0, 2621.0, 1362.0, 685.0, 463.0, 344.0, 222.0, 153.0, 149.0, 85.0, 55.0, 45.0, 17.0, 19.0, 13.0, 7.0, 6.0, 1.0, 2.0, 3.0], "bins": [-0.244140625, -0.236968994140625, -0.22979736328125, -0.222625732421875, -0.2154541015625, -0.208282470703125, -0.20111083984375, -0.193939208984375, -0.186767578125, -0.179595947265625, -0.17242431640625, -0.165252685546875, -0.1580810546875, -0.150909423828125, -0.14373779296875, -0.136566162109375, -0.12939453125, -0.122222900390625, -0.11505126953125, -0.107879638671875, -0.1007080078125, -0.093536376953125, -0.08636474609375, -0.079193115234375, -0.072021484375, -0.064849853515625, -0.05767822265625, -0.050506591796875, -0.0433349609375, -0.036163330078125, -0.02899169921875, -0.021820068359375, -0.0146484375, -0.007476806640625, -0.00030517578125, 0.006866455078125, 0.0140380859375, 0.021209716796875, 0.02838134765625, 0.035552978515625, 0.042724609375, 0.049896240234375, 0.05706787109375, 0.064239501953125, 0.0714111328125, 0.078582763671875, 0.08575439453125, 0.092926025390625, 0.10009765625, 0.107269287109375, 0.11444091796875, 0.121612548828125, 0.1287841796875, 0.135955810546875, 0.14312744140625, 0.150299072265625, 0.157470703125, 0.164642333984375, 0.17181396484375, 0.178985595703125, 0.1861572265625, 0.193328857421875, 0.20050048828125, 0.207672119140625, 0.21484375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.5._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 1.0, 4.0, 2.0, 2.0, 5.0, 3.0, 10.0, 13.0, 14.0, 14.0, 25.0, 30.0, 38.0, 59.0, 61.0, 75.0, 130.0, 128.0, 218.0, 420.0, 1221.0, 3717.0, 11280.0, 33313.0, 93507.0, 225471.0, 495213.0, 954377.0, 1128546.0, 690896.0, 326009.0, 146339.0, 53532.0, 19666.0, 6137.0, 2072.0, 732.0, 334.0, 176.0, 118.0, 96.0, 64.0, 48.0, 30.0, 32.0, 26.0, 17.0, 24.0, 14.0, 12.0, 7.0, 8.0, 1.0, 1.0, 0.0, 3.0, 3.0, 0.0, 5.0, 0.0, 2.0], "bins": [-0.4765625, -0.4613037109375, -0.446044921875, -0.4307861328125, -0.41552734375, -0.4002685546875, -0.385009765625, -0.3697509765625, -0.3544921875, -0.3392333984375, -0.323974609375, -0.3087158203125, -0.29345703125, -0.2781982421875, -0.262939453125, -0.2476806640625, -0.232421875, -0.2171630859375, -0.201904296875, -0.1866455078125, -0.17138671875, -0.1561279296875, -0.140869140625, -0.1256103515625, -0.1103515625, -0.0950927734375, -0.079833984375, -0.0645751953125, -0.04931640625, -0.0340576171875, -0.018798828125, -0.0035400390625, 0.01171875, 0.0269775390625, 0.042236328125, 0.0574951171875, 0.07275390625, 0.0880126953125, 0.103271484375, 0.1185302734375, 0.1337890625, 0.1490478515625, 0.164306640625, 0.1795654296875, 0.19482421875, 0.2100830078125, 0.225341796875, 0.2406005859375, 0.255859375, 0.2711181640625, 0.286376953125, 0.3016357421875, 0.31689453125, 0.3321533203125, 0.347412109375, 0.3626708984375, 0.3779296875, 0.3931884765625, 0.408447265625, 0.4237060546875, 0.43896484375, 0.4542236328125, 0.469482421875, 0.4847412109375, 0.5]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.5._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 2.0, 4.0, 4.0, 8.0, 21.0, 53.0, 117.0, 220.0, 363.0, 727.0, 1252.0, 2253.0, 3484.0, 5773.0, 9500.0, 14508.0, 21024.0, 29460.0, 38729.0, 51920.0, 63520.0, 75075.0, 80727.0, 83189.0, 78037.0, 70015.0, 57005.0, 45656.0, 33586.0, 25189.0, 17870.0, 11920.0, 7343.0, 4628.0, 2802.0, 1613.0, 981.0, 537.0, 310.0, 123.0, 57.0, 45.0, 17.0, 6.0, 2.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.04296875, -0.041332244873046875, -0.03969573974609375, -0.038059234619140625, -0.0364227294921875, -0.034786224365234375, -0.03314971923828125, -0.031513214111328125, -0.029876708984375, -0.028240203857421875, -0.02660369873046875, -0.024967193603515625, -0.0233306884765625, -0.021694183349609375, -0.02005767822265625, -0.018421173095703125, -0.01678466796875, -0.015148162841796875, -0.01351165771484375, -0.011875152587890625, -0.0102386474609375, -0.008602142333984375, -0.00696563720703125, -0.005329132080078125, -0.003692626953125, -0.002056121826171875, -0.00041961669921875, 0.001216888427734375, 0.0028533935546875, 0.004489898681640625, 0.00612640380859375, 0.007762908935546875, 0.0093994140625, 0.011035919189453125, 0.01267242431640625, 0.014308929443359375, 0.0159454345703125, 0.017581939697265625, 0.01921844482421875, 0.020854949951171875, 0.022491455078125, 0.024127960205078125, 0.02576446533203125, 0.027400970458984375, 0.0290374755859375, 0.030673980712890625, 0.03231048583984375, 0.033946990966796875, 0.03558349609375, 0.037220001220703125, 0.03885650634765625, 0.040493011474609375, 0.0421295166015625, 0.043766021728515625, 0.04540252685546875, 0.047039031982421875, 0.048675537109375, 0.050312042236328125, 0.05194854736328125, 0.053585052490234375, 0.0552215576171875, 0.056858062744140625, 0.05849456787109375, 0.060131072998046875, 0.061767578125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.6._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 0.0, 0.0, 2.0, 7.0, 4.0, 3.0, 9.0, 5.0, 16.0, 29.0, 52.0, 77.0, 142.0, 312.0, 492.0, 967.0, 1994.0, 4258.0, 8028.0, 17843.0, 38288.0, 87338.0, 177152.0, 369644.0, 761132.0, 1400380.0, 2361583.0, 3237331.0, 3198466.0, 2315229.0, 1360787.0, 752611.0, 352725.0, 177627.0, 81955.0, 37493.0, 16895.0, 8337.0, 4086.0, 1783.0, 1030.0, 513.0, 253.0, 140.0, 75.0, 37.0, 27.0, 18.0, 8.0, 9.0, 9.0, 4.0, 2.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.236328125, -0.22869873046875, -0.2210693359375, -0.21343994140625, -0.205810546875, -0.19818115234375, -0.1905517578125, -0.18292236328125, -0.17529296875, -0.16766357421875, -0.1600341796875, -0.15240478515625, -0.144775390625, -0.13714599609375, -0.1295166015625, -0.12188720703125, -0.1142578125, -0.10662841796875, -0.0989990234375, -0.09136962890625, -0.083740234375, -0.07611083984375, -0.0684814453125, -0.06085205078125, -0.05322265625, -0.04559326171875, -0.0379638671875, -0.03033447265625, -0.022705078125, -0.01507568359375, -0.0074462890625, 0.00018310546875, 0.0078125, 0.01544189453125, 0.0230712890625, 0.03070068359375, 0.038330078125, 0.04595947265625, 0.0535888671875, 0.06121826171875, 0.06884765625, 0.07647705078125, 0.0841064453125, 0.09173583984375, 0.099365234375, 0.10699462890625, 0.1146240234375, 0.12225341796875, 0.1298828125, 0.13751220703125, 0.1451416015625, 0.15277099609375, 0.160400390625, 0.16802978515625, 0.1756591796875, 0.18328857421875, 0.19091796875, 0.19854736328125, 0.2061767578125, 0.21380615234375, 0.221435546875, 0.22906494140625, 0.2366943359375, 0.24432373046875, 0.251953125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.6._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 1.0, 1.0, 1.0, 1.0, 6.0, 12.0, 7.0, 23.0, 27.0, 31.0, 42.0, 77.0, 99.0, 152.0, 291.0, 486.0, 868.0, 1617.0, 2949.0, 4954.0, 9510.0, 17033.0, 28698.0, 54588.0, 88660.0, 140411.0, 226763.0, 334426.0, 448409.0, 550671.0, 579431.0, 521608.0, 402083.0, 296848.0, 191467.0, 119195.0, 74888.0, 44743.0, 22429.0, 13761.0, 7668.0, 4101.0, 2264.0, 1271.0, 684.0, 416.0, 238.0, 141.0, 82.0, 66.0, 33.0, 28.0, 10.0, 12.0, 7.0, 4.0, 4.0, 3.0, 1.0, 0.0, 1.0], "bins": [-0.271484375, -0.263336181640625, -0.25518798828125, -0.247039794921875, -0.2388916015625, -0.230743408203125, -0.22259521484375, -0.214447021484375, -0.206298828125, -0.198150634765625, -0.19000244140625, -0.181854248046875, -0.1737060546875, -0.165557861328125, -0.15740966796875, -0.149261474609375, -0.14111328125, -0.132965087890625, -0.12481689453125, -0.116668701171875, -0.1085205078125, -0.100372314453125, -0.09222412109375, -0.084075927734375, -0.075927734375, -0.067779541015625, -0.05963134765625, -0.051483154296875, -0.0433349609375, -0.035186767578125, -0.02703857421875, -0.018890380859375, -0.0107421875, -0.002593994140625, 0.00555419921875, 0.013702392578125, 0.0218505859375, 0.029998779296875, 0.03814697265625, 0.046295166015625, 0.054443359375, 0.062591552734375, 0.07073974609375, 0.078887939453125, 0.0870361328125, 0.095184326171875, 0.10333251953125, 0.111480712890625, 0.11962890625, 0.127777099609375, 0.13592529296875, 0.144073486328125, 0.1522216796875, 0.160369873046875, 0.16851806640625, 0.176666259765625, 0.184814453125, 0.192962646484375, 0.20111083984375, 0.209259033203125, 0.2174072265625, 0.225555419921875, 0.23370361328125, 0.241851806640625, 0.25]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.6._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 5.0, 1.0, 7.0, 12.0, 22.0, 32.0, 74.0, 114.0, 147.0, 266.0, 477.0, 646.0, 970.0, 1713.0, 2151.0, 2874.0, 4534.0, 5785.0, 7594.0, 10710.0, 13569.0, 16657.0, 21315.0, 25752.0, 29136.0, 36242.0, 40196.0, 44475.0, 50225.0, 52969.0, 55514.0, 55064.0, 52717.0, 49632.0, 43765.0, 39421.0, 35412.0, 28284.0, 25183.0, 20761.0, 16103.0, 13069.0, 10215.0, 7107.0, 5390.0, 4371.0, 2818.0, 2035.0, 1538.0, 906.0, 636.0, 416.0, 239.0, 154.0, 119.0, 72.0, 25.0, 14.0, 14.0, 6.0, 6.0, 0.0, 2.0], "bins": [-0.044189453125, -0.042804718017578125, -0.04141998291015625, -0.040035247802734375, -0.0386505126953125, -0.037265777587890625, -0.03588104248046875, -0.034496307373046875, -0.033111572265625, -0.031726837158203125, -0.03034210205078125, -0.028957366943359375, -0.0275726318359375, -0.026187896728515625, -0.02480316162109375, -0.023418426513671875, -0.02203369140625, -0.020648956298828125, -0.01926422119140625, -0.017879486083984375, -0.0164947509765625, -0.015110015869140625, -0.01372528076171875, -0.012340545654296875, -0.010955810546875, -0.009571075439453125, -0.00818634033203125, -0.006801605224609375, -0.0054168701171875, -0.004032135009765625, -0.00264739990234375, -0.001262664794921875, 0.0001220703125, 0.001506805419921875, 0.00289154052734375, 0.004276275634765625, 0.0056610107421875, 0.007045745849609375, 0.00843048095703125, 0.009815216064453125, 0.011199951171875, 0.012584686279296875, 0.01396942138671875, 0.015354156494140625, 0.0167388916015625, 0.018123626708984375, 0.01950836181640625, 0.020893096923828125, 0.02227783203125, 0.023662567138671875, 0.02504730224609375, 0.026432037353515625, 0.0278167724609375, 0.029201507568359375, 0.03058624267578125, 0.031970977783203125, 0.033355712890625, 0.034740447998046875, 0.03612518310546875, 0.037509918212890625, 0.0388946533203125, 0.040279388427734375, 0.04166412353515625, 0.043048858642578125, 0.04443359375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.7._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 1.0, 4.0, 5.0, 10.0, 18.0, 34.0, 71.0, 137.0, 259.0, 561.0, 1175.0, 2535.0, 5369.0, 11609.0, 25247.0, 53979.0, 110362.0, 217119.0, 428895.0, 804135.0, 1397824.0, 2273620.0, 3087709.0, 3062494.0, 2250388.0, 1399535.0, 797118.0, 417531.0, 218049.0, 113123.0, 52395.0, 24868.0, 10760.0, 5340.0, 2520.0, 1192.0, 648.0, 302.0, 135.0, 57.0, 31.0, 20.0, 6.0, 11.0, 4.0, 2.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.2333984375, -0.226531982421875, -0.21966552734375, -0.212799072265625, -0.2059326171875, -0.199066162109375, -0.19219970703125, -0.185333251953125, -0.178466796875, -0.171600341796875, -0.16473388671875, -0.157867431640625, -0.1510009765625, -0.144134521484375, -0.13726806640625, -0.130401611328125, -0.12353515625, -0.116668701171875, -0.10980224609375, -0.102935791015625, -0.0960693359375, -0.089202880859375, -0.08233642578125, -0.075469970703125, -0.068603515625, -0.061737060546875, -0.05487060546875, -0.048004150390625, -0.0411376953125, -0.034271240234375, -0.02740478515625, -0.020538330078125, -0.013671875, -0.006805419921875, 6.103515625e-05, 0.006927490234375, 0.0137939453125, 0.020660400390625, 0.02752685546875, 0.034393310546875, 0.041259765625, 0.048126220703125, 0.05499267578125, 0.061859130859375, 0.0687255859375, 0.075592041015625, 0.08245849609375, 0.089324951171875, 0.09619140625, 0.103057861328125, 0.10992431640625, 0.116790771484375, 0.1236572265625, 0.130523681640625, 0.13739013671875, 0.144256591796875, 0.151123046875, 0.157989501953125, 0.16485595703125, 0.171722412109375, 0.1785888671875, 0.185455322265625, 0.19232177734375, 0.199188232421875, 0.2060546875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.7._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 5.0, 4.0, 3.0, 5.0, 18.0, 33.0, 43.0, 76.0, 129.0, 182.0, 332.0, 609.0, 1172.0, 1972.0, 4017.0, 6108.0, 11399.0, 17708.0, 30467.0, 49198.0, 76786.0, 113613.0, 165646.0, 241129.0, 319179.0, 416224.0, 484282.0, 501003.0, 459665.0, 377779.0, 291230.0, 209862.0, 144484.0, 99466.0, 66455.0, 41774.0, 25743.0, 14694.0, 9551.0, 5236.0, 3219.0, 1619.0, 992.0, 509.0, 300.0, 132.0, 95.0, 54.0, 35.0, 31.0, 10.0, 8.0, 9.0, 2.0, 0.0, 3.0], "bins": [-0.25, -0.242919921875, -0.23583984375, -0.228759765625, -0.2216796875, -0.214599609375, -0.20751953125, -0.200439453125, -0.193359375, -0.186279296875, -0.17919921875, -0.172119140625, -0.1650390625, -0.157958984375, -0.15087890625, -0.143798828125, -0.13671875, -0.129638671875, -0.12255859375, -0.115478515625, -0.1083984375, -0.101318359375, -0.09423828125, -0.087158203125, -0.080078125, -0.072998046875, -0.06591796875, -0.058837890625, -0.0517578125, -0.044677734375, -0.03759765625, -0.030517578125, -0.0234375, -0.016357421875, -0.00927734375, -0.002197265625, 0.0048828125, 0.011962890625, 0.01904296875, 0.026123046875, 0.033203125, 0.040283203125, 0.04736328125, 0.054443359375, 0.0615234375, 0.068603515625, 0.07568359375, 0.082763671875, 0.08984375, 0.096923828125, 0.10400390625, 0.111083984375, 0.1181640625, 0.125244140625, 0.13232421875, 0.139404296875, 0.146484375, 0.153564453125, 0.16064453125, 0.167724609375, 0.1748046875, 0.181884765625, 0.18896484375, 0.196044921875, 0.203125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.7._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 3.0, 5.0, 4.0, 8.0, 14.0, 29.0, 57.0, 106.0, 199.0, 377.0, 632.0, 991.0, 1768.0, 2609.0, 4573.0, 6887.0, 10837.0, 15974.0, 22339.0, 30507.0, 39523.0, 50873.0, 60263.0, 68176.0, 72364.0, 75242.0, 73244.0, 65933.0, 57838.0, 48672.0, 37928.0, 28922.0, 21400.0, 14755.0, 10065.0, 6354.0, 3971.0, 2610.0, 1496.0, 957.0, 486.0, 289.0, 189.0, 92.0, 56.0, 27.0, 18.0, 4.0, 4.0, 3.0, 0.0, 0.0, 1.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.04931640625, -0.047794342041015625, -0.04627227783203125, -0.044750213623046875, -0.0432281494140625, -0.041706085205078125, -0.04018402099609375, -0.038661956787109375, -0.037139892578125, -0.035617828369140625, -0.03409576416015625, -0.032573699951171875, -0.0310516357421875, -0.029529571533203125, -0.02800750732421875, -0.026485443115234375, -0.02496337890625, -0.023441314697265625, -0.02191925048828125, -0.020397186279296875, -0.0188751220703125, -0.017353057861328125, -0.01583099365234375, -0.014308929443359375, -0.012786865234375, -0.011264801025390625, -0.00974273681640625, -0.008220672607421875, -0.0066986083984375, -0.005176544189453125, -0.00365447998046875, -0.002132415771484375, -0.0006103515625, 0.000911712646484375, 0.00243377685546875, 0.003955841064453125, 0.0054779052734375, 0.006999969482421875, 0.00852203369140625, 0.010044097900390625, 0.011566162109375, 0.013088226318359375, 0.01461029052734375, 0.016132354736328125, 0.0176544189453125, 0.019176483154296875, 0.02069854736328125, 0.022220611572265625, 0.02374267578125, 0.025264739990234375, 0.02678680419921875, 0.028308868408203125, 0.0298309326171875, 0.031352996826171875, 0.03287506103515625, 0.034397125244140625, 0.035919189453125, 0.037441253662109375, 0.03896331787109375, 0.040485382080078125, 0.0420074462890625, 0.043529510498046875, 0.04505157470703125, 0.046573638916015625, 0.048095703125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.8._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 4.0, 1.0, 0.0, 5.0, 2.0, 7.0, 5.0, 3.0, 4.0, 18.0, 13.0, 16.0, 28.0, 37.0, 40.0, 42.0, 66.0, 82.0, 121.0, 176.0, 309.0, 680.0, 1784.0, 5477.0, 16968.0, 65438.0, 225994.0, 791543.0, 2510575.0, 5319695.0, 4885628.0, 2062332.0, 636346.0, 183027.0, 49845.0, 13671.0, 4185.0, 1441.0, 688.0, 293.0, 147.0, 121.0, 87.0, 59.0, 45.0, 20.0, 40.0, 25.0, 20.0, 13.0, 10.0, 8.0, 6.0, 6.0, 8.0, 2.0, 4.0, 1.0, 0.0, 3.0], "bins": [-0.43359375, -0.420806884765625, -0.40802001953125, -0.395233154296875, -0.3824462890625, -0.369659423828125, -0.35687255859375, -0.344085693359375, -0.331298828125, -0.318511962890625, -0.30572509765625, -0.292938232421875, -0.2801513671875, -0.267364501953125, -0.25457763671875, -0.241790771484375, -0.22900390625, -0.216217041015625, -0.20343017578125, -0.190643310546875, -0.1778564453125, -0.165069580078125, -0.15228271484375, -0.139495849609375, -0.126708984375, -0.113922119140625, -0.10113525390625, -0.088348388671875, -0.0755615234375, -0.062774658203125, -0.04998779296875, -0.037200927734375, -0.0244140625, -0.011627197265625, 0.00115966796875, 0.013946533203125, 0.0267333984375, 0.039520263671875, 0.05230712890625, 0.065093994140625, 0.077880859375, 0.090667724609375, 0.10345458984375, 0.116241455078125, 0.1290283203125, 0.141815185546875, 0.15460205078125, 0.167388916015625, 0.18017578125, 0.192962646484375, 0.20574951171875, 0.218536376953125, 0.2313232421875, 0.244110107421875, 0.25689697265625, 0.269683837890625, 0.282470703125, 0.295257568359375, 0.30804443359375, 0.320831298828125, 0.3336181640625, 0.346405029296875, 0.35919189453125, 0.371978759765625, 0.384765625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.8._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 3.0, 3.0, 5.0, 0.0, 3.0, 4.0, 5.0, 8.0, 12.0, 25.0, 26.0, 43.0, 49.0, 79.0, 140.0, 245.0, 402.0, 838.0, 2097.0, 4493.0, 10899.0, 26442.0, 62315.0, 148671.0, 319590.0, 625243.0, 958307.0, 928234.0, 585669.0, 289139.0, 133089.0, 57520.0, 22920.0, 10162.0, 4102.0, 1674.0, 851.0, 410.0, 197.0, 127.0, 87.0, 43.0, 45.0, 35.0, 8.0, 14.0, 10.0, 8.0, 0.0, 3.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.4453125, -0.431793212890625, -0.41827392578125, -0.404754638671875, -0.3912353515625, -0.377716064453125, -0.36419677734375, -0.350677490234375, -0.337158203125, -0.323638916015625, -0.31011962890625, -0.296600341796875, -0.2830810546875, -0.269561767578125, -0.25604248046875, -0.242523193359375, -0.22900390625, -0.215484619140625, -0.20196533203125, -0.188446044921875, -0.1749267578125, -0.161407470703125, -0.14788818359375, -0.134368896484375, -0.120849609375, -0.107330322265625, -0.09381103515625, -0.080291748046875, -0.0667724609375, -0.053253173828125, -0.03973388671875, -0.026214599609375, -0.0126953125, 0.000823974609375, 0.01434326171875, 0.027862548828125, 0.0413818359375, 0.054901123046875, 0.06842041015625, 0.081939697265625, 0.095458984375, 0.108978271484375, 0.12249755859375, 0.136016845703125, 0.1495361328125, 0.163055419921875, 0.17657470703125, 0.190093994140625, 0.20361328125, 0.217132568359375, 0.23065185546875, 0.244171142578125, 0.2576904296875, 0.271209716796875, 0.28472900390625, 0.298248291015625, 0.311767578125, 0.325286865234375, 0.33880615234375, 0.352325439453125, 0.3658447265625, 0.379364013671875, 0.39288330078125, 0.406402587890625, 0.419921875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.8._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0, 2.0, 8.0, 19.0, 30.0, 38.0, 84.0, 148.0, 189.0, 351.0, 533.0, 786.0, 1173.0, 1762.0, 2605.0, 3706.0, 5524.0, 8017.0, 11464.0, 16035.0, 22031.0, 28759.0, 37874.0, 47276.0, 57566.0, 65271.0, 71402.0, 74173.0, 71089.0, 65115.0, 57115.0, 47560.0, 37627.0, 29171.0, 21908.0, 16437.0, 11355.0, 8075.0, 5483.0, 3804.0, 2665.0, 1857.0, 1219.0, 854.0, 547.0, 396.0, 215.0, 135.0, 80.0, 58.0, 29.0, 26.0, 11.0, 8.0, 3.0, 3.0, 4.0], "bins": [-0.050537109375, -0.049072265625, -0.047607421875, -0.046142578125, -0.044677734375, -0.043212890625, -0.041748046875, -0.040283203125, -0.038818359375, -0.037353515625, -0.035888671875, -0.034423828125, -0.032958984375, -0.031494140625, -0.030029296875, -0.028564453125, -0.027099609375, -0.025634765625, -0.024169921875, -0.022705078125, -0.021240234375, -0.019775390625, -0.018310546875, -0.016845703125, -0.015380859375, -0.013916015625, -0.012451171875, -0.010986328125, -0.009521484375, -0.008056640625, -0.006591796875, -0.005126953125, -0.003662109375, -0.002197265625, -0.000732421875, 0.000732421875, 0.002197265625, 0.003662109375, 0.005126953125, 0.006591796875, 0.008056640625, 0.009521484375, 0.010986328125, 0.012451171875, 0.013916015625, 0.015380859375, 0.016845703125, 0.018310546875, 0.019775390625, 0.021240234375, 0.022705078125, 0.024169921875, 0.025634765625, 0.027099609375, 0.028564453125, 0.030029296875, 0.031494140625, 0.032958984375, 0.034423828125, 0.035888671875, 0.037353515625, 0.038818359375, 0.040283203125, 0.041748046875, 0.043212890625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.9._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [4.0, 3.0, 1.0, 1.0, 2.0, 4.0, 5.0, 7.0, 22.0, 37.0, 50.0, 96.0, 202.0, 362.0, 558.0, 1098.0, 2052.0, 3653.0, 7386.0, 14160.0, 29917.0, 62253.0, 124200.0, 235062.0, 494957.0, 938935.0, 1668053.0, 2610121.0, 3269011.0, 2918663.0, 2011842.0, 1132257.0, 629928.0, 308869.0, 158149.0, 79595.0, 38721.0, 17867.0, 9087.0, 4526.0, 2407.0, 1425.0, 702.0, 418.0, 231.0, 145.0, 70.0, 33.0, 26.0, 11.0, 13.0, 7.0, 4.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 1.0], "bins": [-0.2021484375, -0.195098876953125, -0.18804931640625, -0.180999755859375, -0.1739501953125, -0.166900634765625, -0.15985107421875, -0.152801513671875, -0.145751953125, -0.138702392578125, -0.13165283203125, -0.124603271484375, -0.1175537109375, -0.110504150390625, -0.10345458984375, -0.096405029296875, -0.08935546875, -0.082305908203125, -0.07525634765625, -0.068206787109375, -0.0611572265625, -0.054107666015625, -0.04705810546875, -0.040008544921875, -0.032958984375, -0.025909423828125, -0.01885986328125, -0.011810302734375, -0.0047607421875, 0.002288818359375, 0.00933837890625, 0.016387939453125, 0.0234375, 0.030487060546875, 0.03753662109375, 0.044586181640625, 0.0516357421875, 0.058685302734375, 0.06573486328125, 0.072784423828125, 0.079833984375, 0.086883544921875, 0.09393310546875, 0.100982666015625, 0.1080322265625, 0.115081787109375, 0.12213134765625, 0.129180908203125, 0.13623046875, 0.143280029296875, 0.15032958984375, 0.157379150390625, 0.1644287109375, 0.171478271484375, 0.17852783203125, 0.185577392578125, 0.192626953125, 0.199676513671875, 0.20672607421875, 0.213775634765625, 0.2208251953125, 0.227874755859375, 0.23492431640625, 0.241973876953125, 0.2490234375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.9._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 5.0, 6.0, 1.0, 9.0, 9.0, 10.0, 18.0, 35.0, 69.0, 107.0, 193.0, 405.0, 782.0, 1688.0, 3424.0, 6946.0, 13442.0, 24866.0, 48930.0, 89670.0, 153907.0, 267186.0, 424707.0, 624440.0, 725774.0, 655349.0, 464904.0, 296862.0, 172770.0, 101334.0, 56031.0, 28783.0, 15682.0, 7976.0, 4157.0, 1951.0, 935.0, 422.0, 222.0, 128.0, 59.0, 33.0, 21.0, 15.0, 8.0, 7.0, 9.0, 3.0, 2.0, 4.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.26953125, -0.259765625, -0.25, -0.240234375, -0.23046875, -0.220703125, -0.2109375, -0.201171875, -0.19140625, -0.181640625, -0.171875, -0.162109375, -0.15234375, -0.142578125, -0.1328125, -0.123046875, -0.11328125, -0.103515625, -0.09375, -0.083984375, -0.07421875, -0.064453125, -0.0546875, -0.044921875, -0.03515625, -0.025390625, -0.015625, -0.005859375, 0.00390625, 0.013671875, 0.0234375, 0.033203125, 0.04296875, 0.052734375, 0.0625, 0.072265625, 0.08203125, 0.091796875, 0.1015625, 0.111328125, 0.12109375, 0.130859375, 0.140625, 0.150390625, 0.16015625, 0.169921875, 0.1796875, 0.189453125, 0.19921875, 0.208984375, 0.21875, 0.228515625, 0.23828125, 0.248046875, 0.2578125, 0.267578125, 0.27734375, 0.287109375, 0.296875, 0.306640625, 0.31640625, 0.326171875, 0.3359375, 0.345703125, 0.35546875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.9._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 4.0, 2.0, 1.0, 5.0, 17.0, 13.0, 51.0, 82.0, 93.0, 164.0, 285.0, 432.0, 662.0, 1110.0, 1386.0, 1932.0, 2745.0, 3939.0, 5765.0, 7559.0, 10428.0, 15196.0, 22264.0, 31652.0, 44739.0, 58864.0, 75665.0, 89657.0, 97337.0, 88355.0, 75228.0, 57003.0, 43790.0, 30954.0, 22133.0, 14264.0, 10128.0, 7513.0, 5440.0, 3945.0, 2901.0, 1863.0, 1329.0, 907.0, 683.0, 469.0, 281.0, 182.0, 87.0, 59.0, 62.0, 19.0, 21.0, 8.0, 1.0, 3.0], "bins": [-0.08984375, -0.08737945556640625, -0.0849151611328125, -0.08245086669921875, -0.079986572265625, -0.07752227783203125, -0.0750579833984375, -0.07259368896484375, -0.07012939453125, -0.06766510009765625, -0.0652008056640625, -0.06273651123046875, -0.060272216796875, -0.05780792236328125, -0.0553436279296875, -0.05287933349609375, -0.0504150390625, -0.04795074462890625, -0.0454864501953125, -0.04302215576171875, -0.040557861328125, -0.03809356689453125, -0.0356292724609375, -0.03316497802734375, -0.03070068359375, -0.02823638916015625, -0.0257720947265625, -0.02330780029296875, -0.020843505859375, -0.01837921142578125, -0.0159149169921875, -0.01345062255859375, -0.010986328125, -0.00852203369140625, -0.0060577392578125, -0.00359344482421875, -0.001129150390625, 0.00133514404296875, 0.0037994384765625, 0.00626373291015625, 0.00872802734375, 0.01119232177734375, 0.0136566162109375, 0.01612091064453125, 0.018585205078125, 0.02104949951171875, 0.0235137939453125, 0.02597808837890625, 0.0284423828125, 0.03090667724609375, 0.0333709716796875, 0.03583526611328125, 0.038299560546875, 0.04076385498046875, 0.0432281494140625, 0.04569244384765625, 0.04815673828125, 0.05062103271484375, 0.0530853271484375, 0.05554962158203125, 0.058013916015625, 0.06047821044921875, 0.0629425048828125, 0.06540679931640625, 0.06787109375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.10._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 1.0, 2.0, 5.0, 4.0, 13.0, 22.0, 23.0, 29.0, 28.0, 46.0, 70.0, 72.0, 134.0, 199.0, 328.0, 793.0, 1921.0, 5074.0, 14964.0, 42869.0, 124225.0, 342355.0, 947860.0, 2362641.0, 4705157.0, 4571680.0, 2270650.0, 886817.0, 316819.0, 120334.0, 39540.0, 14082.0, 4985.0, 1719.0, 774.0, 351.0, 182.0, 121.0, 86.0, 54.0, 52.0, 35.0, 20.0, 19.0, 17.0, 11.0, 7.0, 2.0, 4.0, 5.0, 2.0, 2.0, 1.0, 2.0, 1.0, 1.0], "bins": [-0.365234375, -0.354156494140625, -0.34307861328125, -0.332000732421875, -0.3209228515625, -0.309844970703125, -0.29876708984375, -0.287689208984375, -0.276611328125, -0.265533447265625, -0.25445556640625, -0.243377685546875, -0.2322998046875, -0.221221923828125, -0.21014404296875, -0.199066162109375, -0.18798828125, -0.176910400390625, -0.16583251953125, -0.154754638671875, -0.1436767578125, -0.132598876953125, -0.12152099609375, -0.110443115234375, -0.099365234375, -0.088287353515625, -0.07720947265625, -0.066131591796875, -0.0550537109375, -0.043975830078125, -0.03289794921875, -0.021820068359375, -0.0107421875, 0.000335693359375, 0.01141357421875, 0.022491455078125, 0.0335693359375, 0.044647216796875, 0.05572509765625, 0.066802978515625, 0.077880859375, 0.088958740234375, 0.10003662109375, 0.111114501953125, 0.1221923828125, 0.133270263671875, 0.14434814453125, 0.155426025390625, 0.16650390625, 0.177581787109375, 0.18865966796875, 0.199737548828125, 0.2108154296875, 0.221893310546875, 0.23297119140625, 0.244049072265625, 0.255126953125, 0.266204833984375, 0.27728271484375, 0.288360595703125, 0.2994384765625, 0.310516357421875, 0.32159423828125, 0.332672119140625, 0.34375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.10._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 2.0, 2.0, 5.0, 12.0, 7.0, 11.0, 12.0, 29.0, 37.0, 38.0, 65.0, 78.0, 102.0, 185.0, 247.0, 435.0, 1028.0, 2095.0, 5107.0, 10371.0, 24548.0, 52843.0, 114934.0, 219355.0, 431182.0, 709957.0, 897499.0, 754900.0, 474411.0, 251610.0, 130403.0, 60657.0, 28345.0, 12930.0, 5771.0, 2442.0, 1168.0, 554.0, 278.0, 174.0, 134.0, 78.0, 65.0, 57.0, 35.0, 31.0, 24.0, 11.0, 11.0, 4.0, 9.0, 2.0, 4.0, 1.0, 1.0, 2.0], "bins": [-0.41796875, -0.4058837890625, -0.393798828125, -0.3817138671875, -0.36962890625, -0.3575439453125, -0.345458984375, -0.3333740234375, -0.3212890625, -0.3092041015625, -0.297119140625, -0.2850341796875, -0.27294921875, -0.2608642578125, -0.248779296875, -0.2366943359375, -0.224609375, -0.2125244140625, -0.200439453125, -0.1883544921875, -0.17626953125, -0.1641845703125, -0.152099609375, -0.1400146484375, -0.1279296875, -0.1158447265625, -0.103759765625, -0.0916748046875, -0.07958984375, -0.0675048828125, -0.055419921875, -0.0433349609375, -0.03125, -0.0191650390625, -0.007080078125, 0.0050048828125, 0.01708984375, 0.0291748046875, 0.041259765625, 0.0533447265625, 0.0654296875, 0.0775146484375, 0.089599609375, 0.1016845703125, 0.11376953125, 0.1258544921875, 0.137939453125, 0.1500244140625, 0.162109375, 0.1741943359375, 0.186279296875, 0.1983642578125, 0.21044921875, 0.2225341796875, 0.234619140625, 0.2467041015625, 0.2587890625, 0.2708740234375, 0.282958984375, 0.2950439453125, 0.30712890625, 0.3192138671875, 0.331298828125, 0.3433837890625, 0.35546875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.10._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 3.0, 2.0, 6.0, 9.0, 25.0, 22.0, 54.0, 74.0, 108.0, 166.0, 267.0, 344.0, 544.0, 739.0, 1107.0, 1602.0, 2126.0, 2738.0, 3790.0, 5172.0, 6947.0, 9309.0, 12984.0, 18374.0, 23386.0, 30048.0, 38373.0, 48723.0, 56782.0, 64266.0, 69658.0, 71336.0, 68162.0, 61390.0, 53351.0, 45114.0, 34800.0, 27435.0, 20937.0, 16423.0, 11527.0, 8253.0, 6101.0, 4658.0, 3473.0, 2557.0, 1879.0, 1475.0, 958.0, 648.0, 481.0, 332.0, 212.0, 169.0, 81.0, 71.0, 39.0, 25.0, 17.0, 10.0, 5.0, 9.0, 2.0], "bins": [-0.051513671875, -0.04991912841796875, -0.0483245849609375, -0.04673004150390625, -0.045135498046875, -0.04354095458984375, -0.0419464111328125, -0.04035186767578125, -0.03875732421875, -0.03716278076171875, -0.0355682373046875, -0.03397369384765625, -0.032379150390625, -0.03078460693359375, -0.0291900634765625, -0.02759552001953125, -0.0260009765625, -0.02440643310546875, -0.0228118896484375, -0.02121734619140625, -0.019622802734375, -0.01802825927734375, -0.0164337158203125, -0.01483917236328125, -0.01324462890625, -0.01165008544921875, -0.0100555419921875, -0.00846099853515625, -0.006866455078125, -0.00527191162109375, -0.0036773681640625, -0.00208282470703125, -0.00048828125, 0.00110626220703125, 0.0027008056640625, 0.00429534912109375, 0.005889892578125, 0.00748443603515625, 0.0090789794921875, 0.01067352294921875, 0.01226806640625, 0.01386260986328125, 0.0154571533203125, 0.01705169677734375, 0.018646240234375, 0.02024078369140625, 0.0218353271484375, 0.02342987060546875, 0.0250244140625, 0.02661895751953125, 0.0282135009765625, 0.02980804443359375, 0.031402587890625, 0.03299713134765625, 0.0345916748046875, 0.03618621826171875, 0.03778076171875, 0.03937530517578125, 0.0409698486328125, 0.04256439208984375, 0.044158935546875, 0.04575347900390625, 0.0473480224609375, 0.04894256591796875, 0.050537109375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.11._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [5.0, 5.0, 6.0, 13.0, 19.0, 45.0, 79.0, 147.0, 246.0, 404.0, 743.0, 1434.0, 2744.0, 4157.0, 7683.0, 13956.0, 26402.0, 46432.0, 76123.0, 135737.0, 217834.0, 382567.0, 623730.0, 981650.0, 1530024.0, 2078694.0, 2522717.0, 2455059.0, 1980449.0, 1390276.0, 896690.0, 572893.0, 347919.0, 201686.0, 115621.0, 71401.0, 39474.0, 23435.0, 12750.0, 6908.0, 4082.0, 2059.0, 1336.0, 674.0, 421.0, 196.0, 133.0, 67.0, 29.0, 18.0, 12.0, 14.0, 9.0, 4.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.1416015625, -0.1363372802734375, -0.131072998046875, -0.1258087158203125, -0.12054443359375, -0.1152801513671875, -0.110015869140625, -0.1047515869140625, -0.0994873046875, -0.0942230224609375, -0.088958740234375, -0.0836944580078125, -0.07843017578125, -0.0731658935546875, -0.067901611328125, -0.0626373291015625, -0.057373046875, -0.0521087646484375, -0.046844482421875, -0.0415802001953125, -0.03631591796875, -0.0310516357421875, -0.025787353515625, -0.0205230712890625, -0.0152587890625, -0.0099945068359375, -0.004730224609375, 0.0005340576171875, 0.00579833984375, 0.0110626220703125, 0.016326904296875, 0.0215911865234375, 0.02685546875, 0.0321197509765625, 0.037384033203125, 0.0426483154296875, 0.04791259765625, 0.0531768798828125, 0.058441162109375, 0.0637054443359375, 0.0689697265625, 0.0742340087890625, 0.079498291015625, 0.0847625732421875, 0.09002685546875, 0.0952911376953125, 0.100555419921875, 0.1058197021484375, 0.111083984375, 0.1163482666015625, 0.121612548828125, 0.1268768310546875, 0.13214111328125, 0.1374053955078125, 0.142669677734375, 0.1479339599609375, 0.1531982421875, 0.1584625244140625, 0.163726806640625, 0.1689910888671875, 0.17425537109375, 0.1795196533203125, 0.184783935546875, 0.1900482177734375, 0.1953125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.11._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 4.0, 5.0, 3.0, 3.0, 8.0, 18.0, 17.0, 32.0, 51.0, 67.0, 108.0, 168.0, 277.0, 543.0, 1078.0, 2307.0, 4760.0, 9295.0, 18348.0, 34079.0, 64552.0, 111221.0, 184008.0, 303955.0, 456287.0, 622728.0, 682568.0, 595771.0, 424493.0, 281706.0, 170595.0, 103352.0, 58159.0, 30379.0, 16427.0, 8400.0, 4246.0, 2061.0, 1030.0, 517.0, 239.0, 159.0, 107.0, 59.0, 32.0, 28.0, 13.0, 10.0, 8.0, 6.0, 4.0, 6.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.28125, -0.272003173828125, -0.26275634765625, -0.253509521484375, -0.2442626953125, -0.235015869140625, -0.22576904296875, -0.216522216796875, -0.207275390625, -0.198028564453125, -0.18878173828125, -0.179534912109375, -0.1702880859375, -0.161041259765625, -0.15179443359375, -0.142547607421875, -0.13330078125, -0.124053955078125, -0.11480712890625, -0.105560302734375, -0.0963134765625, -0.087066650390625, -0.07781982421875, -0.068572998046875, -0.059326171875, -0.050079345703125, -0.04083251953125, -0.031585693359375, -0.0223388671875, -0.013092041015625, -0.00384521484375, 0.005401611328125, 0.0146484375, 0.023895263671875, 0.03314208984375, 0.042388916015625, 0.0516357421875, 0.060882568359375, 0.07012939453125, 0.079376220703125, 0.088623046875, 0.097869873046875, 0.10711669921875, 0.116363525390625, 0.1256103515625, 0.134857177734375, 0.14410400390625, 0.153350830078125, 0.16259765625, 0.171844482421875, 0.18109130859375, 0.190338134765625, 0.1995849609375, 0.208831787109375, 0.21807861328125, 0.227325439453125, 0.236572265625, 0.245819091796875, 0.25506591796875, 0.264312744140625, 0.2735595703125, 0.282806396484375, 0.29205322265625, 0.301300048828125, 0.310546875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.11._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 4.0, 1.0, 1.0, 7.0, 19.0, 34.0, 65.0, 161.0, 319.0, 628.0, 1029.0, 1990.0, 3113.0, 5418.0, 8531.0, 12771.0, 18426.0, 25619.0, 34055.0, 44178.0, 53956.0, 63561.0, 71350.0, 75446.0, 74744.0, 71919.0, 63308.0, 53154.0, 43385.0, 35051.0, 25584.0, 18437.0, 12611.0, 7930.0, 5334.0, 3184.0, 1990.0, 1149.0, 573.0, 343.0, 143.0, 87.0, 26.0, 26.0, 5.0, 6.0, 2.0, 1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.044677734375, -0.04302215576171875, -0.0413665771484375, -0.03971099853515625, -0.038055419921875, -0.03639984130859375, -0.0347442626953125, -0.03308868408203125, -0.03143310546875, -0.02977752685546875, -0.0281219482421875, -0.02646636962890625, -0.024810791015625, -0.02315521240234375, -0.0214996337890625, -0.01984405517578125, -0.0181884765625, -0.01653289794921875, -0.0148773193359375, -0.01322174072265625, -0.011566162109375, -0.00991058349609375, -0.0082550048828125, -0.00659942626953125, -0.00494384765625, -0.00328826904296875, -0.0016326904296875, 2.288818359375e-05, 0.001678466796875, 0.00333404541015625, 0.0049896240234375, 0.00664520263671875, 0.00830078125, 0.00995635986328125, 0.0116119384765625, 0.01326751708984375, 0.014923095703125, 0.01657867431640625, 0.0182342529296875, 0.01988983154296875, 0.02154541015625, 0.02320098876953125, 0.0248565673828125, 0.02651214599609375, 0.028167724609375, 0.02982330322265625, 0.0314788818359375, 0.03313446044921875, 0.0347900390625, 0.03644561767578125, 0.0381011962890625, 0.03975677490234375, 0.041412353515625, 0.04306793212890625, 0.0447235107421875, 0.04637908935546875, 0.04803466796875, 0.04969024658203125, 0.0513458251953125, 0.05300140380859375, 0.054656982421875, 0.05631256103515625, 0.0579681396484375, 0.05962371826171875, 0.061279296875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.12._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 4.0, 0.0, 6.0, 3.0, 8.0, 7.0, 23.0, 31.0, 67.0, 113.0, 263.0, 503.0, 925.0, 2274.0, 4786.0, 11263.0, 27350.0, 60112.0, 134654.0, 265555.0, 563638.0, 1019376.0, 1741474.0, 2497484.0, 3009207.0, 2760919.0, 2037083.0, 1237985.0, 719147.0, 361664.0, 176469.0, 79877.0, 37423.0, 15385.0, 6642.0, 2819.0, 1407.0, 647.0, 329.0, 132.0, 57.0, 36.0, 14.0, 17.0, 11.0, 3.0, 9.0, 6.0, 1.0, 0.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.2333984375, -0.2266693115234375, -0.219940185546875, -0.2132110595703125, -0.20648193359375, -0.1997528076171875, -0.193023681640625, -0.1862945556640625, -0.1795654296875, -0.1728363037109375, -0.166107177734375, -0.1593780517578125, -0.15264892578125, -0.1459197998046875, -0.139190673828125, -0.1324615478515625, -0.125732421875, -0.1190032958984375, -0.112274169921875, -0.1055450439453125, -0.09881591796875, -0.0920867919921875, -0.085357666015625, -0.0786285400390625, -0.0718994140625, -0.0651702880859375, -0.058441162109375, -0.0517120361328125, -0.04498291015625, -0.0382537841796875, -0.031524658203125, -0.0247955322265625, -0.01806640625, -0.0113372802734375, -0.004608154296875, 0.0021209716796875, 0.00885009765625, 0.0155792236328125, 0.022308349609375, 0.0290374755859375, 0.0357666015625, 0.0424957275390625, 0.049224853515625, 0.0559539794921875, 0.06268310546875, 0.0694122314453125, 0.076141357421875, 0.0828704833984375, 0.089599609375, 0.0963287353515625, 0.103057861328125, 0.1097869873046875, 0.11651611328125, 0.1232452392578125, 0.129974365234375, 0.1367034912109375, 0.1434326171875, 0.1501617431640625, 0.156890869140625, 0.1636199951171875, 0.17034912109375, 0.1770782470703125, 0.183807373046875, 0.1905364990234375, 0.197265625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.12._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 3.0, 0.0, 2.0, 8.0, 10.0, 14.0, 15.0, 30.0, 61.0, 127.0, 203.0, 370.0, 716.0, 1320.0, 2546.0, 4375.0, 7892.0, 14449.0, 23101.0, 39764.0, 67699.0, 98776.0, 145885.0, 224089.0, 294638.0, 384807.0, 457205.0, 492291.0, 474056.0, 412036.0, 325791.0, 247876.0, 166511.0, 118245.0, 74534.0, 47984.0, 29600.0, 15697.0, 9584.0, 5399.0, 2995.0, 1649.0, 968.0, 458.0, 211.0, 121.0, 77.0, 28.0, 23.0, 23.0, 16.0, 12.0, 3.0, 4.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.205078125, -0.1981658935546875, -0.191253662109375, -0.1843414306640625, -0.17742919921875, -0.1705169677734375, -0.163604736328125, -0.1566925048828125, -0.1497802734375, -0.1428680419921875, -0.135955810546875, -0.1290435791015625, -0.12213134765625, -0.1152191162109375, -0.108306884765625, -0.1013946533203125, -0.094482421875, -0.0875701904296875, -0.080657958984375, -0.0737457275390625, -0.06683349609375, -0.0599212646484375, -0.053009033203125, -0.0460968017578125, -0.0391845703125, -0.0322723388671875, -0.025360107421875, -0.0184478759765625, -0.01153564453125, -0.0046234130859375, 0.002288818359375, 0.0092010498046875, 0.01611328125, 0.0230255126953125, 0.029937744140625, 0.0368499755859375, 0.04376220703125, 0.0506744384765625, 0.057586669921875, 0.0644989013671875, 0.0714111328125, 0.0783233642578125, 0.085235595703125, 0.0921478271484375, 0.09906005859375, 0.1059722900390625, 0.112884521484375, 0.1197967529296875, 0.126708984375, 0.1336212158203125, 0.140533447265625, 0.1474456787109375, 0.15435791015625, 0.1612701416015625, 0.168182373046875, 0.1750946044921875, 0.1820068359375, 0.1889190673828125, 0.195831298828125, 0.2027435302734375, 0.20965576171875, 0.2165679931640625, 0.223480224609375, 0.2303924560546875, 0.2373046875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.12._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 2.0, 7.0, 4.0, 17.0, 16.0, 34.0, 71.0, 79.0, 146.0, 210.0, 344.0, 572.0, 846.0, 1311.0, 1722.0, 2795.0, 4077.0, 5621.0, 7844.0, 10319.0, 13098.0, 18100.0, 22397.0, 28100.0, 33680.0, 39004.0, 44796.0, 51152.0, 54084.0, 57301.0, 58677.0, 56719.0, 53763.0, 48718.0, 43828.0, 38838.0, 31002.0, 26731.0, 21064.0, 16875.0, 13090.0, 8994.0, 7248.0, 5241.0, 3725.0, 2493.0, 1575.0, 1151.0, 761.0, 534.0, 360.0, 201.0, 117.0, 87.0, 54.0, 40.0, 12.0, 6.0, 12.0, 5.0, 3.0, 1.0, 3.0], "bins": [-0.049072265625, -0.04750823974609375, -0.0459442138671875, -0.04438018798828125, -0.042816162109375, -0.04125213623046875, -0.0396881103515625, -0.03812408447265625, -0.03656005859375, -0.03499603271484375, -0.0334320068359375, -0.03186798095703125, -0.030303955078125, -0.02873992919921875, -0.0271759033203125, -0.02561187744140625, -0.0240478515625, -0.02248382568359375, -0.0209197998046875, -0.01935577392578125, -0.017791748046875, -0.01622772216796875, -0.0146636962890625, -0.01309967041015625, -0.01153564453125, -0.00997161865234375, -0.0084075927734375, -0.00684356689453125, -0.005279541015625, -0.00371551513671875, -0.0021514892578125, -0.00058746337890625, 0.0009765625, 0.00254058837890625, 0.0041046142578125, 0.00566864013671875, 0.007232666015625, 0.00879669189453125, 0.0103607177734375, 0.01192474365234375, 0.01348876953125, 0.01505279541015625, 0.0166168212890625, 0.01818084716796875, 0.019744873046875, 0.02130889892578125, 0.0228729248046875, 0.02443695068359375, 0.0260009765625, 0.02756500244140625, 0.0291290283203125, 0.03069305419921875, 0.032257080078125, 0.03382110595703125, 0.0353851318359375, 0.03694915771484375, 0.03851318359375, 0.04007720947265625, 0.0416412353515625, 0.04320526123046875, 0.044769287109375, 0.04633331298828125, 0.0478973388671875, 0.04946136474609375, 0.051025390625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.13._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 4.0, 0.0, 7.0, 11.0, 18.0, 39.0, 70.0, 146.0, 258.0, 482.0, 830.0, 1565.0, 2923.0, 5081.0, 8719.0, 15948.0, 30784.0, 52094.0, 88206.0, 146645.0, 243893.0, 413243.0, 664663.0, 1049382.0, 1620306.0, 2228220.0, 2566921.0, 2424278.0, 1868016.0, 1275101.0, 807339.0, 511626.0, 312187.0, 179986.0, 107404.0, 65153.0, 38091.0, 21083.0, 11630.0, 6588.0, 3957.0, 2001.0, 1082.0, 568.0, 292.0, 174.0, 79.0, 51.0, 29.0, 19.0, 9.0, 4.0, 2.0, 0.0, 1.0, 1.0], "bins": [-0.1875, -0.182098388671875, -0.17669677734375, -0.171295166015625, -0.1658935546875, -0.160491943359375, -0.15509033203125, -0.149688720703125, -0.144287109375, -0.138885498046875, -0.13348388671875, -0.128082275390625, -0.1226806640625, -0.117279052734375, -0.11187744140625, -0.106475830078125, -0.10107421875, -0.095672607421875, -0.09027099609375, -0.084869384765625, -0.0794677734375, -0.074066162109375, -0.06866455078125, -0.063262939453125, -0.057861328125, -0.052459716796875, -0.04705810546875, -0.041656494140625, -0.0362548828125, -0.030853271484375, -0.02545166015625, -0.020050048828125, -0.0146484375, -0.009246826171875, -0.00384521484375, 0.001556396484375, 0.0069580078125, 0.012359619140625, 0.01776123046875, 0.023162841796875, 0.028564453125, 0.033966064453125, 0.03936767578125, 0.044769287109375, 0.0501708984375, 0.055572509765625, 0.06097412109375, 0.066375732421875, 0.07177734375, 0.077178955078125, 0.08258056640625, 0.087982177734375, 0.0933837890625, 0.098785400390625, 0.10418701171875, 0.109588623046875, 0.114990234375, 0.120391845703125, 0.12579345703125, 0.131195068359375, 0.1365966796875, 0.141998291015625, 0.14739990234375, 0.152801513671875, 0.158203125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.13._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 3.0, 2.0, 1.0, 2.0, 4.0, 9.0, 8.0, 12.0, 16.0, 21.0, 32.0, 48.0, 77.0, 112.0, 191.0, 323.0, 767.0, 1696.0, 3499.0, 7663.0, 15164.0, 31183.0, 61676.0, 112224.0, 198737.0, 350900.0, 552035.0, 751087.0, 752584.0, 558316.0, 355754.0, 197828.0, 117094.0, 63891.0, 30381.0, 16335.0, 7652.0, 3714.0, 1598.0, 758.0, 354.0, 203.0, 114.0, 67.0, 43.0, 32.0, 27.0, 20.0, 10.0, 8.0, 8.0, 10.0, 2.0, 2.0, 0.0, 5.0], "bins": [-0.376953125, -0.366485595703125, -0.35601806640625, -0.345550537109375, -0.3350830078125, -0.324615478515625, -0.31414794921875, -0.303680419921875, -0.293212890625, -0.282745361328125, -0.27227783203125, -0.261810302734375, -0.2513427734375, -0.240875244140625, -0.23040771484375, -0.219940185546875, -0.20947265625, -0.199005126953125, -0.18853759765625, -0.178070068359375, -0.1676025390625, -0.157135009765625, -0.14666748046875, -0.136199951171875, -0.125732421875, -0.115264892578125, -0.10479736328125, -0.094329833984375, -0.0838623046875, -0.073394775390625, -0.06292724609375, -0.052459716796875, -0.0419921875, -0.031524658203125, -0.02105712890625, -0.010589599609375, -0.0001220703125, 0.010345458984375, 0.02081298828125, 0.031280517578125, 0.041748046875, 0.052215576171875, 0.06268310546875, 0.073150634765625, 0.0836181640625, 0.094085693359375, 0.10455322265625, 0.115020751953125, 0.12548828125, 0.135955810546875, 0.14642333984375, 0.156890869140625, 0.1673583984375, 0.177825927734375, 0.18829345703125, 0.198760986328125, 0.209228515625, 0.219696044921875, 0.23016357421875, 0.240631103515625, 0.2510986328125, 0.261566162109375, 0.27203369140625, 0.282501220703125, 0.29296875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.13._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 1.0, 1.0, 17.0, 13.0, 31.0, 83.0, 202.0, 430.0, 961.0, 1918.0, 3679.0, 6047.0, 11468.0, 17888.0, 27237.0, 40613.0, 53910.0, 68733.0, 81493.0, 90959.0, 90715.0, 85853.0, 72884.0, 59164.0, 44519.0, 30679.0, 21110.0, 12535.0, 7784.0, 4032.0, 2385.0, 1241.0, 571.0, 294.0, 132.0, 50.0, 22.0, 11.0, 6.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0712890625, -0.06937026977539062, -0.06745147705078125, -0.06553268432617188, -0.0636138916015625, -0.061695098876953125, -0.05977630615234375, -0.057857513427734375, -0.055938720703125, -0.054019927978515625, -0.05210113525390625, -0.050182342529296875, -0.0482635498046875, -0.046344757080078125, -0.04442596435546875, -0.042507171630859375, -0.04058837890625, -0.038669586181640625, -0.03675079345703125, -0.034832000732421875, -0.0329132080078125, -0.030994415283203125, -0.02907562255859375, -0.027156829833984375, -0.025238037109375, -0.023319244384765625, -0.02140045166015625, -0.019481658935546875, -0.0175628662109375, -0.015644073486328125, -0.01372528076171875, -0.011806488037109375, -0.0098876953125, -0.007968902587890625, -0.00605010986328125, -0.004131317138671875, -0.0022125244140625, -0.000293731689453125, 0.00162506103515625, 0.003543853759765625, 0.005462646484375, 0.007381439208984375, 0.00930023193359375, 0.011219024658203125, 0.0131378173828125, 0.015056610107421875, 0.01697540283203125, 0.018894195556640625, 0.02081298828125, 0.022731781005859375, 0.02465057373046875, 0.026569366455078125, 0.0284881591796875, 0.030406951904296875, 0.03232574462890625, 0.034244537353515625, 0.036163330078125, 0.038082122802734375, 0.04000091552734375, 0.041919708251953125, 0.0438385009765625, 0.045757293701171875, 0.04767608642578125, 0.049594879150390625, 0.051513671875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.14._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 2.0, 0.0, 2.0, 5.0, 2.0, 4.0, 7.0, 12.0, 11.0, 19.0, 28.0, 34.0, 48.0, 56.0, 87.0, 83.0, 136.0, 172.0, 271.0, 446.0, 913.0, 2602.0, 8272.0, 25042.0, 77767.0, 217612.0, 613843.0, 1490538.0, 3180427.0, 4624615.0, 3582500.0, 1784119.0, 749263.0, 271020.0, 99359.0, 31813.0, 9950.0, 3442.0, 1202.0, 500.0, 239.0, 171.0, 132.0, 86.0, 95.0, 49.0, 64.0, 34.0, 23.0, 22.0, 20.0, 16.0, 12.0, 9.0, 12.0, 0.0, 0.0, 2.0, 1.0, 1.0, 1.0, 1.0], "bins": [-0.310546875, -0.30072021484375, -0.2908935546875, -0.28106689453125, -0.271240234375, -0.26141357421875, -0.2515869140625, -0.24176025390625, -0.23193359375, -0.22210693359375, -0.2122802734375, -0.20245361328125, -0.192626953125, -0.18280029296875, -0.1729736328125, -0.16314697265625, -0.1533203125, -0.14349365234375, -0.1336669921875, -0.12384033203125, -0.114013671875, -0.10418701171875, -0.0943603515625, -0.08453369140625, -0.07470703125, -0.06488037109375, -0.0550537109375, -0.04522705078125, -0.035400390625, -0.02557373046875, -0.0157470703125, -0.00592041015625, 0.00390625, 0.01373291015625, 0.0235595703125, 0.03338623046875, 0.043212890625, 0.05303955078125, 0.0628662109375, 0.07269287109375, 0.08251953125, 0.09234619140625, 0.1021728515625, 0.11199951171875, 0.121826171875, 0.13165283203125, 0.1414794921875, 0.15130615234375, 0.1611328125, 0.17095947265625, 0.1807861328125, 0.19061279296875, 0.200439453125, 0.21026611328125, 0.2200927734375, 0.22991943359375, 0.23974609375, 0.24957275390625, 0.2593994140625, 0.26922607421875, 0.279052734375, 0.28887939453125, 0.2987060546875, 0.30853271484375, 0.318359375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.14._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 3.0, 5.0, 6.0, 12.0, 6.0, 12.0, 24.0, 22.0, 34.0, 38.0, 50.0, 73.0, 76.0, 120.0, 196.0, 253.0, 448.0, 789.0, 1481.0, 3277.0, 6796.0, 13039.0, 26618.0, 55964.0, 110011.0, 208941.0, 390988.0, 629143.0, 824419.0, 764762.0, 527259.0, 302141.0, 160364.0, 85754.0, 40296.0, 21027.0, 9932.0, 4583.0, 2450.0, 1179.0, 561.0, 334.0, 225.0, 145.0, 103.0, 78.0, 59.0, 48.0, 37.0, 28.0, 23.0, 20.0, 11.0, 7.0, 11.0, 9.0, 4.0, 3.0, 0.0, 4.0], "bins": [-0.36328125, -0.352203369140625, -0.34112548828125, -0.330047607421875, -0.3189697265625, -0.307891845703125, -0.29681396484375, -0.285736083984375, -0.274658203125, -0.263580322265625, -0.25250244140625, -0.241424560546875, -0.2303466796875, -0.219268798828125, -0.20819091796875, -0.197113037109375, -0.18603515625, -0.174957275390625, -0.16387939453125, -0.152801513671875, -0.1417236328125, -0.130645751953125, -0.11956787109375, -0.108489990234375, -0.097412109375, -0.086334228515625, -0.07525634765625, -0.064178466796875, -0.0531005859375, -0.042022705078125, -0.03094482421875, -0.019866943359375, -0.0087890625, 0.002288818359375, 0.01336669921875, 0.024444580078125, 0.0355224609375, 0.046600341796875, 0.05767822265625, 0.068756103515625, 0.079833984375, 0.090911865234375, 0.10198974609375, 0.113067626953125, 0.1241455078125, 0.135223388671875, 0.14630126953125, 0.157379150390625, 0.16845703125, 0.179534912109375, 0.19061279296875, 0.201690673828125, 0.2127685546875, 0.223846435546875, 0.23492431640625, 0.246002197265625, 0.257080078125, 0.268157958984375, 0.27923583984375, 0.290313720703125, 0.3013916015625, 0.312469482421875, 0.32354736328125, 0.334625244140625, 0.345703125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.14._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 6.0, 3.0, 7.0, 23.0, 25.0, 51.0, 97.0, 147.0, 200.0, 357.0, 535.0, 867.0, 1129.0, 1821.0, 2640.0, 3692.0, 5072.0, 7058.0, 9373.0, 12307.0, 15801.0, 20309.0, 25045.0, 30083.0, 35231.0, 41028.0, 46608.0, 51474.0, 55614.0, 58239.0, 57654.0, 55480.0, 49704.0, 45979.0, 40318.0, 34206.0, 28180.0, 24456.0, 19726.0, 15663.0, 11798.0, 9085.0, 6654.0, 5034.0, 3266.0, 2503.0, 1777.0, 1188.0, 811.0, 488.0, 351.0, 183.0, 137.0, 88.0, 52.0, 25.0, 16.0, 4.0, 4.0, 1.0, 5.0], "bins": [-0.052001953125, -0.05042266845703125, -0.0488433837890625, -0.04726409912109375, -0.045684814453125, -0.04410552978515625, -0.0425262451171875, -0.04094696044921875, -0.03936767578125, -0.03778839111328125, -0.0362091064453125, -0.03462982177734375, -0.033050537109375, -0.03147125244140625, -0.0298919677734375, -0.02831268310546875, -0.0267333984375, -0.02515411376953125, -0.0235748291015625, -0.02199554443359375, -0.020416259765625, -0.01883697509765625, -0.0172576904296875, -0.01567840576171875, -0.01409912109375, -0.01251983642578125, -0.0109405517578125, -0.00936126708984375, -0.007781982421875, -0.00620269775390625, -0.0046234130859375, -0.00304412841796875, -0.00146484375, 0.00011444091796875, 0.0016937255859375, 0.00327301025390625, 0.004852294921875, 0.00643157958984375, 0.0080108642578125, 0.00959014892578125, 0.01116943359375, 0.01274871826171875, 0.0143280029296875, 0.01590728759765625, 0.017486572265625, 0.01906585693359375, 0.0206451416015625, 0.02222442626953125, 0.0238037109375, 0.02538299560546875, 0.0269622802734375, 0.02854156494140625, 0.030120849609375, 0.03170013427734375, 0.0332794189453125, 0.03485870361328125, 0.03643798828125, 0.03801727294921875, 0.0395965576171875, 0.04117584228515625, 0.042755126953125, 0.04433441162109375, 0.0459136962890625, 0.04749298095703125, 0.049072265625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.15._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 2.0, 6.0, 5.0, 5.0, 7.0, 7.0, 14.0, 23.0, 23.0, 44.0, 51.0, 109.0, 167.0, 334.0, 810.0, 1907.0, 5840.0, 15509.0, 45712.0, 140350.0, 396404.0, 1171147.0, 3081064.0, 5614742.0, 3900466.0, 1566856.0, 548064.0, 190767.0, 63105.0, 21814.0, 7370.0, 2574.0, 1010.0, 397.0, 190.0, 91.0, 62.0, 37.0, 31.0, 21.0, 14.0, 13.0, 12.0, 10.0, 5.0, 2.0, 6.0, 1.0, 0.0, 4.0, 2.0, 2.0, 0.0, 2.0], "bins": [-0.419921875, -0.407440185546875, -0.39495849609375, -0.382476806640625, -0.3699951171875, -0.357513427734375, -0.34503173828125, -0.332550048828125, -0.320068359375, -0.307586669921875, -0.29510498046875, -0.282623291015625, -0.2701416015625, -0.257659912109375, -0.24517822265625, -0.232696533203125, -0.22021484375, -0.207733154296875, -0.19525146484375, -0.182769775390625, -0.1702880859375, -0.157806396484375, -0.14532470703125, -0.132843017578125, -0.120361328125, -0.107879638671875, -0.09539794921875, -0.082916259765625, -0.0704345703125, -0.057952880859375, -0.04547119140625, -0.032989501953125, -0.0205078125, -0.008026123046875, 0.00445556640625, 0.016937255859375, 0.0294189453125, 0.041900634765625, 0.05438232421875, 0.066864013671875, 0.079345703125, 0.091827392578125, 0.10430908203125, 0.116790771484375, 0.1292724609375, 0.141754150390625, 0.15423583984375, 0.166717529296875, 0.17919921875, 0.191680908203125, 0.20416259765625, 0.216644287109375, 0.2291259765625, 0.241607666015625, 0.25408935546875, 0.266571044921875, 0.279052734375, 0.291534423828125, 0.30401611328125, 0.316497802734375, 0.3289794921875, 0.341461181640625, 0.35394287109375, 0.366424560546875, 0.37890625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.15._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 8.0, 3.0, 3.0, 11.0, 7.0, 19.0, 19.0, 33.0, 61.0, 94.0, 177.0, 231.0, 502.0, 1106.0, 2699.0, 6381.0, 14133.0, 32964.0, 71010.0, 147516.0, 273298.0, 477389.0, 702403.0, 795449.0, 680772.0, 461596.0, 263070.0, 140673.0, 67958.0, 30570.0, 13630.0, 5895.0, 2442.0, 1080.0, 490.0, 244.0, 141.0, 79.0, 45.0, 32.0, 21.0, 17.0, 9.0, 8.0, 6.0, 0.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.3828125, -0.3717041015625, -0.360595703125, -0.3494873046875, -0.33837890625, -0.3272705078125, -0.316162109375, -0.3050537109375, -0.2939453125, -0.2828369140625, -0.271728515625, -0.2606201171875, -0.24951171875, -0.2384033203125, -0.227294921875, -0.2161865234375, -0.205078125, -0.1939697265625, -0.182861328125, -0.1717529296875, -0.16064453125, -0.1495361328125, -0.138427734375, -0.1273193359375, -0.1162109375, -0.1051025390625, -0.093994140625, -0.0828857421875, -0.07177734375, -0.0606689453125, -0.049560546875, -0.0384521484375, -0.02734375, -0.0162353515625, -0.005126953125, 0.0059814453125, 0.01708984375, 0.0281982421875, 0.039306640625, 0.0504150390625, 0.0615234375, 0.0726318359375, 0.083740234375, 0.0948486328125, 0.10595703125, 0.1170654296875, 0.128173828125, 0.1392822265625, 0.150390625, 0.1614990234375, 0.172607421875, 0.1837158203125, 0.19482421875, 0.2059326171875, 0.217041015625, 0.2281494140625, 0.2392578125, 0.2503662109375, 0.261474609375, 0.2725830078125, 0.28369140625, 0.2947998046875, 0.305908203125, 0.3170166015625, 0.328125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.15._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 1.0, 0.0, 3.0, 1.0, 5.0, 4.0, 10.0, 14.0, 24.0, 46.0, 41.0, 94.0, 126.0, 226.0, 415.0, 628.0, 1239.0, 1994.0, 3196.0, 5542.0, 8663.0, 13871.0, 20682.0, 27927.0, 38321.0, 51192.0, 60484.0, 70250.0, 77983.0, 81131.0, 77322.0, 70226.0, 59203.0, 49564.0, 37112.0, 27867.0, 19393.0, 13435.0, 8520.0, 5282.0, 3004.0, 1973.0, 1084.0, 641.0, 344.0, 242.0, 119.0, 61.0, 62.0, 33.0, 26.0, 23.0, 10.0, 9.0, 2.0, 1.0, 1.0, 1.0, 2.0, 1.0, 2.0], "bins": [-0.052734375, -0.05110931396484375, -0.0494842529296875, -0.04785919189453125, -0.046234130859375, -0.04460906982421875, -0.0429840087890625, -0.04135894775390625, -0.03973388671875, -0.03810882568359375, -0.0364837646484375, -0.03485870361328125, -0.033233642578125, -0.03160858154296875, -0.0299835205078125, -0.02835845947265625, -0.0267333984375, -0.02510833740234375, -0.0234832763671875, -0.02185821533203125, -0.020233154296875, -0.01860809326171875, -0.0169830322265625, -0.01535797119140625, -0.01373291015625, -0.01210784912109375, -0.0104827880859375, -0.00885772705078125, -0.007232666015625, -0.00560760498046875, -0.0039825439453125, -0.00235748291015625, -0.000732421875, 0.00089263916015625, 0.0025177001953125, 0.00414276123046875, 0.005767822265625, 0.00739288330078125, 0.0090179443359375, 0.01064300537109375, 0.01226806640625, 0.01389312744140625, 0.0155181884765625, 0.01714324951171875, 0.018768310546875, 0.02039337158203125, 0.0220184326171875, 0.02364349365234375, 0.0252685546875, 0.02689361572265625, 0.0285186767578125, 0.03014373779296875, 0.031768798828125, 0.03339385986328125, 0.0350189208984375, 0.03664398193359375, 0.03826904296875, 0.03989410400390625, 0.0415191650390625, 0.04314422607421875, 0.044769287109375, 0.04639434814453125, 0.0480194091796875, 0.04964447021484375, 0.05126953125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.16._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 1.0, 2.0, 1.0, 6.0, 6.0, 7.0, 3.0, 15.0, 18.0, 12.0, 21.0, 24.0, 28.0, 50.0, 47.0, 73.0, 95.0, 124.0, 280.0, 608.0, 1845.0, 6233.0, 20414.0, 73090.0, 237961.0, 750117.0, 2027977.0, 4291933.0, 5012215.0, 2728693.0, 1101447.0, 364387.0, 112832.0, 32675.0, 9337.0, 2752.0, 898.0, 383.0, 169.0, 106.0, 85.0, 54.0, 55.0, 27.0, 28.0, 23.0, 10.0, 10.0, 6.0, 8.0, 8.0, 2.0, 2.0, 5.0, 0.0, 0.0, 1.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.369140625, -0.3572998046875, -0.345458984375, -0.3336181640625, -0.32177734375, -0.3099365234375, -0.298095703125, -0.2862548828125, -0.2744140625, -0.2625732421875, -0.250732421875, -0.2388916015625, -0.22705078125, -0.2152099609375, -0.203369140625, -0.1915283203125, -0.1796875, -0.1678466796875, -0.156005859375, -0.1441650390625, -0.13232421875, -0.1204833984375, -0.108642578125, -0.0968017578125, -0.0849609375, -0.0731201171875, -0.061279296875, -0.0494384765625, -0.03759765625, -0.0257568359375, -0.013916015625, -0.0020751953125, 0.009765625, 0.0216064453125, 0.033447265625, 0.0452880859375, 0.05712890625, 0.0689697265625, 0.080810546875, 0.0926513671875, 0.1044921875, 0.1163330078125, 0.128173828125, 0.1400146484375, 0.15185546875, 0.1636962890625, 0.175537109375, 0.1873779296875, 0.19921875, 0.2110595703125, 0.222900390625, 0.2347412109375, 0.24658203125, 0.2584228515625, 0.270263671875, 0.2821044921875, 0.2939453125, 0.3057861328125, 0.317626953125, 0.3294677734375, 0.34130859375, 0.3531494140625, 0.364990234375, 0.3768310546875, 0.388671875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.16._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 1.0, 2.0, 4.0, 2.0, 2.0, 4.0, 9.0, 9.0, 9.0, 17.0, 23.0, 31.0, 38.0, 66.0, 118.0, 197.0, 358.0, 645.0, 1460.0, 2860.0, 6025.0, 11784.0, 24649.0, 52174.0, 103844.0, 191537.0, 357078.0, 553632.0, 731835.0, 743715.0, 588184.0, 381730.0, 213773.0, 114887.0, 60227.0, 26471.0, 13697.0, 6535.0, 3300.0, 1610.0, 757.0, 375.0, 220.0, 144.0, 76.0, 46.0, 40.0, 32.0, 21.0, 8.0, 9.0, 8.0, 4.0, 7.0, 3.0, 4.0, 0.0, 0.0, 2.0, 0.0, 1.0, 2.0], "bins": [-0.328125, -0.31756591796875, -0.3070068359375, -0.29644775390625, -0.285888671875, -0.27532958984375, -0.2647705078125, -0.25421142578125, -0.24365234375, -0.23309326171875, -0.2225341796875, -0.21197509765625, -0.201416015625, -0.19085693359375, -0.1802978515625, -0.16973876953125, -0.1591796875, -0.14862060546875, -0.1380615234375, -0.12750244140625, -0.116943359375, -0.10638427734375, -0.0958251953125, -0.08526611328125, -0.07470703125, -0.06414794921875, -0.0535888671875, -0.04302978515625, -0.032470703125, -0.02191162109375, -0.0113525390625, -0.00079345703125, 0.009765625, 0.02032470703125, 0.0308837890625, 0.04144287109375, 0.052001953125, 0.06256103515625, 0.0731201171875, 0.08367919921875, 0.09423828125, 0.10479736328125, 0.1153564453125, 0.12591552734375, 0.136474609375, 0.14703369140625, 0.1575927734375, 0.16815185546875, 0.1787109375, 0.18927001953125, 0.1998291015625, 0.21038818359375, 0.220947265625, 0.23150634765625, 0.2420654296875, 0.25262451171875, 0.26318359375, 0.27374267578125, 0.2843017578125, 0.29486083984375, 0.305419921875, 0.31597900390625, 0.3265380859375, 0.33709716796875, 0.34765625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.16._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 2.0, 1.0, 3.0, 6.0, 10.0, 12.0, 12.0, 20.0, 30.0, 41.0, 48.0, 46.0, 80.0, 113.0, 191.0, 306.0, 458.0, 839.0, 1471.0, 2979.0, 5349.0, 9986.0, 17518.0, 29339.0, 45975.0, 66105.0, 88432.0, 103615.0, 109907.0, 102004.0, 85548.0, 63295.0, 42996.0, 27260.0, 15809.0, 9067.0, 4692.0, 2687.0, 1399.0, 730.0, 451.0, 254.0, 177.0, 101.0, 106.0, 53.0, 40.0, 30.0, 21.0, 24.0, 5.0, 10.0, 5.0, 2.0, 6.0, 5.0, 1.0, 1.0, 0.0, 1.0, 1.0, 2.0], "bins": [-0.080078125, -0.07744598388671875, -0.0748138427734375, -0.07218170166015625, -0.069549560546875, -0.06691741943359375, -0.0642852783203125, -0.06165313720703125, -0.05902099609375, -0.05638885498046875, -0.0537567138671875, -0.05112457275390625, -0.048492431640625, -0.04586029052734375, -0.0432281494140625, -0.04059600830078125, -0.0379638671875, -0.03533172607421875, -0.0326995849609375, -0.03006744384765625, -0.027435302734375, -0.02480316162109375, -0.0221710205078125, -0.01953887939453125, -0.01690673828125, -0.01427459716796875, -0.0116424560546875, -0.00901031494140625, -0.006378173828125, -0.00374603271484375, -0.0011138916015625, 0.00151824951171875, 0.004150390625, 0.00678253173828125, 0.0094146728515625, 0.01204681396484375, 0.014678955078125, 0.01731109619140625, 0.0199432373046875, 0.02257537841796875, 0.02520751953125, 0.02783966064453125, 0.0304718017578125, 0.03310394287109375, 0.035736083984375, 0.03836822509765625, 0.0410003662109375, 0.04363250732421875, 0.0462646484375, 0.04889678955078125, 0.0515289306640625, 0.05416107177734375, 0.056793212890625, 0.05942535400390625, 0.0620574951171875, 0.06468963623046875, 0.06732177734375, 0.06995391845703125, 0.0725860595703125, 0.07521820068359375, 0.077850341796875, 0.08048248291015625, 0.0831146240234375, 0.08574676513671875, 0.08837890625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.17._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 3.0, 1.0, 8.0, 3.0, 19.0, 21.0, 37.0, 49.0, 72.0, 122.0, 200.0, 418.0, 662.0, 1279.0, 2469.0, 4538.0, 7638.0, 15546.0, 28730.0, 56735.0, 100368.0, 184910.0, 312398.0, 559856.0, 915082.0, 1471641.0, 2186752.0, 2825157.0, 2729441.0, 2036646.0, 1338032.0, 841116.0, 506483.0, 290766.0, 160732.0, 93703.0, 49905.0, 26430.0, 13444.0, 7344.0, 3769.0, 2144.0, 1101.0, 580.0, 342.0, 190.0, 142.0, 73.0, 33.0, 21.0, 23.0, 11.0, 6.0, 11.0, 2.0, 2.0, 0.0, 1.0, 3.0, 1.0, 1.0], "bins": [-0.1923828125, -0.1861572265625, -0.179931640625, -0.1737060546875, -0.16748046875, -0.1612548828125, -0.155029296875, -0.1488037109375, -0.142578125, -0.1363525390625, -0.130126953125, -0.1239013671875, -0.11767578125, -0.1114501953125, -0.105224609375, -0.0989990234375, -0.0927734375, -0.0865478515625, -0.080322265625, -0.0740966796875, -0.06787109375, -0.0616455078125, -0.055419921875, -0.0491943359375, -0.04296875, -0.0367431640625, -0.030517578125, -0.0242919921875, -0.01806640625, -0.0118408203125, -0.005615234375, 0.0006103515625, 0.0068359375, 0.0130615234375, 0.019287109375, 0.0255126953125, 0.03173828125, 0.0379638671875, 0.044189453125, 0.0504150390625, 0.056640625, 0.0628662109375, 0.069091796875, 0.0753173828125, 0.08154296875, 0.0877685546875, 0.093994140625, 0.1002197265625, 0.1064453125, 0.1126708984375, 0.118896484375, 0.1251220703125, 0.13134765625, 0.1375732421875, 0.143798828125, 0.1500244140625, 0.15625, 0.1624755859375, 0.168701171875, 0.1749267578125, 0.18115234375, 0.1873779296875, 0.193603515625, 0.1998291015625, 0.2060546875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.17._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 0.0, 3.0, 1.0, 6.0, 10.0, 3.0, 10.0, 17.0, 13.0, 15.0, 20.0, 40.0, 66.0, 68.0, 97.0, 180.0, 288.0, 508.0, 1082.0, 2311.0, 5040.0, 11803.0, 23735.0, 52696.0, 109253.0, 197832.0, 366159.0, 575681.0, 760394.0, 754599.0, 570823.0, 362832.0, 197846.0, 104386.0, 52543.0, 23054.0, 11165.0, 5223.0, 2286.0, 983.0, 524.0, 239.0, 154.0, 83.0, 59.0, 40.0, 41.0, 24.0, 13.0, 9.0, 11.0, 14.0, 4.0, 6.0, 2.0, 3.0, 0.0, 2.0, 0.0, 2.0], "bins": [-0.359375, -0.348480224609375, -0.33758544921875, -0.326690673828125, -0.3157958984375, -0.304901123046875, -0.29400634765625, -0.283111572265625, -0.272216796875, -0.261322021484375, -0.25042724609375, -0.239532470703125, -0.2286376953125, -0.217742919921875, -0.20684814453125, -0.195953369140625, -0.18505859375, -0.174163818359375, -0.16326904296875, -0.152374267578125, -0.1414794921875, -0.130584716796875, -0.11968994140625, -0.108795166015625, -0.097900390625, -0.087005615234375, -0.07611083984375, -0.065216064453125, -0.0543212890625, -0.043426513671875, -0.03253173828125, -0.021636962890625, -0.0107421875, 0.000152587890625, 0.01104736328125, 0.021942138671875, 0.0328369140625, 0.043731689453125, 0.05462646484375, 0.065521240234375, 0.076416015625, 0.087310791015625, 0.09820556640625, 0.109100341796875, 0.1199951171875, 0.130889892578125, 0.14178466796875, 0.152679443359375, 0.16357421875, 0.174468994140625, 0.18536376953125, 0.196258544921875, 0.2071533203125, 0.218048095703125, 0.22894287109375, 0.239837646484375, 0.250732421875, 0.261627197265625, 0.27252197265625, 0.283416748046875, 0.2943115234375, 0.305206298828125, 0.31610107421875, 0.326995849609375, 0.337890625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.17._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 2.0, 6.0, 7.0, 29.0, 34.0, 62.0, 75.0, 177.0, 243.0, 444.0, 601.0, 1029.0, 1405.0, 2318.0, 2860.0, 4460.0, 5601.0, 8102.0, 10165.0, 13387.0, 17052.0, 20760.0, 25210.0, 29659.0, 35100.0, 40235.0, 44958.0, 48647.0, 53504.0, 56079.0, 55680.0, 52539.0, 48211.0, 43958.0, 39895.0, 34489.0, 29532.0, 25120.0, 20933.0, 17051.0, 13074.0, 10390.0, 7404.0, 5908.0, 3936.0, 3068.0, 2017.0, 1561.0, 941.0, 706.0, 386.0, 285.0, 147.0, 87.0, 60.0, 38.0, 28.0, 10.0, 5.0, 4.0, 0.0, 1.0, 2.0], "bins": [-0.04345703125, -0.04205322265625, -0.0406494140625, -0.03924560546875, -0.037841796875, -0.03643798828125, -0.0350341796875, -0.03363037109375, -0.0322265625, -0.03082275390625, -0.0294189453125, -0.02801513671875, -0.026611328125, -0.02520751953125, -0.0238037109375, -0.02239990234375, -0.02099609375, -0.01959228515625, -0.0181884765625, -0.01678466796875, -0.015380859375, -0.01397705078125, -0.0125732421875, -0.01116943359375, -0.009765625, -0.00836181640625, -0.0069580078125, -0.00555419921875, -0.004150390625, -0.00274658203125, -0.0013427734375, 6.103515625e-05, 0.00146484375, 0.00286865234375, 0.0042724609375, 0.00567626953125, 0.007080078125, 0.00848388671875, 0.0098876953125, 0.01129150390625, 0.0126953125, 0.01409912109375, 0.0155029296875, 0.01690673828125, 0.018310546875, 0.01971435546875, 0.0211181640625, 0.02252197265625, 0.02392578125, 0.02532958984375, 0.0267333984375, 0.02813720703125, 0.029541015625, 0.03094482421875, 0.0323486328125, 0.03375244140625, 0.03515625, 0.03656005859375, 0.0379638671875, 0.03936767578125, 0.040771484375, 0.04217529296875, 0.0435791015625, 0.04498291015625, 0.04638671875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.18._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 1.0, 1.0, 2.0, 1.0, 4.0, 8.0, 9.0, 12.0, 14.0, 7.0, 29.0, 44.0, 46.0, 81.0, 145.0, 230.0, 407.0, 863.0, 1807.0, 3332.0, 7310.0, 15288.0, 30667.0, 64726.0, 118175.0, 218328.0, 409942.0, 729720.0, 1138160.0, 1808894.0, 2413471.0, 2727436.0, 2435937.0, 1833269.0, 1182389.0, 742777.0, 408762.0, 235629.0, 121139.0, 66327.0, 31517.0, 15127.0, 7680.0, 3824.0, 1694.0, 963.0, 451.0, 222.0, 128.0, 66.0, 46.0, 32.0, 24.0, 12.0, 12.0, 5.0, 5.0, 8.0, 5.0, 1.0, 3.0], "bins": [-0.2138671875, -0.207672119140625, -0.20147705078125, -0.195281982421875, -0.1890869140625, -0.182891845703125, -0.17669677734375, -0.170501708984375, -0.164306640625, -0.158111572265625, -0.15191650390625, -0.145721435546875, -0.1395263671875, -0.133331298828125, -0.12713623046875, -0.120941162109375, -0.11474609375, -0.108551025390625, -0.10235595703125, -0.096160888671875, -0.0899658203125, -0.083770751953125, -0.07757568359375, -0.071380615234375, -0.065185546875, -0.058990478515625, -0.05279541015625, -0.046600341796875, -0.0404052734375, -0.034210205078125, -0.02801513671875, -0.021820068359375, -0.015625, -0.009429931640625, -0.00323486328125, 0.002960205078125, 0.0091552734375, 0.015350341796875, 0.02154541015625, 0.027740478515625, 0.033935546875, 0.040130615234375, 0.04632568359375, 0.052520751953125, 0.0587158203125, 0.064910888671875, 0.07110595703125, 0.077301025390625, 0.08349609375, 0.089691162109375, 0.09588623046875, 0.102081298828125, 0.1082763671875, 0.114471435546875, 0.12066650390625, 0.126861572265625, 0.133056640625, 0.139251708984375, 0.14544677734375, 0.151641845703125, 0.1578369140625, 0.164031982421875, 0.17022705078125, 0.176422119140625, 0.1826171875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.18._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [3.0, 5.0, 2.0, 3.0, 4.0, 7.0, 7.0, 10.0, 15.0, 13.0, 30.0, 42.0, 46.0, 49.0, 97.0, 148.0, 221.0, 355.0, 624.0, 1120.0, 2112.0, 4219.0, 7779.0, 15086.0, 27006.0, 52409.0, 90723.0, 146767.0, 235243.0, 354763.0, 491505.0, 598661.0, 613201.0, 521048.0, 379925.0, 260631.0, 162165.0, 100831.0, 58779.0, 32237.0, 17202.0, 9101.0, 4778.0, 2308.0, 1286.0, 614.0, 392.0, 232.0, 155.0, 109.0, 75.0, 41.0, 29.0, 27.0, 26.0, 14.0, 7.0, 5.0, 5.0, 2.0, 2.0, 1.0, 1.0, 1.0], "bins": [-0.283203125, -0.274383544921875, -0.26556396484375, -0.256744384765625, -0.2479248046875, -0.239105224609375, -0.23028564453125, -0.221466064453125, -0.212646484375, -0.203826904296875, -0.19500732421875, -0.186187744140625, -0.1773681640625, -0.168548583984375, -0.15972900390625, -0.150909423828125, -0.14208984375, -0.133270263671875, -0.12445068359375, -0.115631103515625, -0.1068115234375, -0.097991943359375, -0.08917236328125, -0.080352783203125, -0.071533203125, -0.062713623046875, -0.05389404296875, -0.045074462890625, -0.0362548828125, -0.027435302734375, -0.01861572265625, -0.009796142578125, -0.0009765625, 0.007843017578125, 0.01666259765625, 0.025482177734375, 0.0343017578125, 0.043121337890625, 0.05194091796875, 0.060760498046875, 0.069580078125, 0.078399658203125, 0.08721923828125, 0.096038818359375, 0.1048583984375, 0.113677978515625, 0.12249755859375, 0.131317138671875, 0.14013671875, 0.148956298828125, 0.15777587890625, 0.166595458984375, 0.1754150390625, 0.184234619140625, 0.19305419921875, 0.201873779296875, 0.210693359375, 0.219512939453125, 0.22833251953125, 0.237152099609375, 0.2459716796875, 0.254791259765625, 0.26361083984375, 0.272430419921875, 0.28125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.18._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 3.0, 0.0, 4.0, 4.0, 5.0, 6.0, 13.0, 27.0, 44.0, 82.0, 115.0, 204.0, 411.0, 669.0, 1396.0, 2412.0, 3743.0, 6389.0, 10351.0, 15879.0, 23499.0, 34491.0, 45211.0, 59228.0, 72847.0, 83324.0, 89315.0, 86322.0, 77026.0, 64080.0, 49032.0, 37942.0, 26835.0, 18775.0, 11444.0, 7616.0, 4677.0, 2740.0, 1508.0, 909.0, 489.0, 241.0, 146.0, 77.0, 55.0, 34.0, 17.0, 15.0, 6.0, 6.0, 3.0, 3.0, 1.0, 3.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.059814453125, -0.057651519775390625, -0.05548858642578125, -0.053325653076171875, -0.0511627197265625, -0.048999786376953125, -0.04683685302734375, -0.044673919677734375, -0.042510986328125, -0.040348052978515625, -0.03818511962890625, -0.036022186279296875, -0.0338592529296875, -0.031696319580078125, -0.02953338623046875, -0.027370452880859375, -0.02520751953125, -0.023044586181640625, -0.02088165283203125, -0.018718719482421875, -0.0165557861328125, -0.014392852783203125, -0.01222991943359375, -0.010066986083984375, -0.007904052734375, -0.005741119384765625, -0.00357818603515625, -0.001415252685546875, 0.0007476806640625, 0.002910614013671875, 0.00507354736328125, 0.007236480712890625, 0.0093994140625, 0.011562347412109375, 0.01372528076171875, 0.015888214111328125, 0.0180511474609375, 0.020214080810546875, 0.02237701416015625, 0.024539947509765625, 0.026702880859375, 0.028865814208984375, 0.03102874755859375, 0.033191680908203125, 0.0353546142578125, 0.037517547607421875, 0.03968048095703125, 0.041843414306640625, 0.04400634765625, 0.046169281005859375, 0.04833221435546875, 0.050495147705078125, 0.0526580810546875, 0.054821014404296875, 0.05698394775390625, 0.059146881103515625, 0.061309814453125, 0.06347274780273438, 0.06563568115234375, 0.06779861450195312, 0.0699615478515625, 0.07212448120117188, 0.07428741455078125, 0.07645034790039062, 0.07861328125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.19._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 3.0, 2.0, 2.0, 7.0, 7.0, 11.0, 13.0, 18.0, 26.0, 55.0, 70.0, 136.0, 298.0, 639.0, 1110.0, 2251.0, 5087.0, 9540.0, 22802.0, 50424.0, 106590.0, 218728.0, 420673.0, 830895.0, 1419029.0, 2255671.0, 2993789.0, 2994936.0, 2277720.0, 1458077.0, 837514.0, 441914.0, 223785.0, 110096.0, 52067.0, 22907.0, 10480.0, 4887.0, 2358.0, 1303.0, 601.0, 326.0, 159.0, 85.0, 47.0, 35.0, 9.0, 12.0, 6.0, 1.0, 3.0, 4.0, 1.0, 2.0, 2.0], "bins": [-0.25390625, -0.246856689453125, -0.23980712890625, -0.232757568359375, -0.2257080078125, -0.218658447265625, -0.21160888671875, -0.204559326171875, -0.197509765625, -0.190460205078125, -0.18341064453125, -0.176361083984375, -0.1693115234375, -0.162261962890625, -0.15521240234375, -0.148162841796875, -0.14111328125, -0.134063720703125, -0.12701416015625, -0.119964599609375, -0.1129150390625, -0.105865478515625, -0.09881591796875, -0.091766357421875, -0.084716796875, -0.077667236328125, -0.07061767578125, -0.063568115234375, -0.0565185546875, -0.049468994140625, -0.04241943359375, -0.035369873046875, -0.0283203125, -0.021270751953125, -0.01422119140625, -0.007171630859375, -0.0001220703125, 0.006927490234375, 0.01397705078125, 0.021026611328125, 0.028076171875, 0.035125732421875, 0.04217529296875, 0.049224853515625, 0.0562744140625, 0.063323974609375, 0.07037353515625, 0.077423095703125, 0.08447265625, 0.091522216796875, 0.09857177734375, 0.105621337890625, 0.1126708984375, 0.119720458984375, 0.12677001953125, 0.133819580078125, 0.140869140625, 0.147918701171875, 0.15496826171875, 0.162017822265625, 0.1690673828125, 0.176116943359375, 0.18316650390625, 0.190216064453125, 0.197265625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.19._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 4.0, 0.0, 4.0, 6.0, 4.0, 14.0, 14.0, 17.0, 22.0, 28.0, 55.0, 66.0, 108.0, 204.0, 336.0, 606.0, 1232.0, 2721.0, 5161.0, 10738.0, 22257.0, 44376.0, 86318.0, 147753.0, 260258.0, 406112.0, 574917.0, 673552.0, 643388.0, 506517.0, 349015.0, 205344.0, 120928.0, 66215.0, 32869.0, 16633.0, 8402.0, 4044.0, 1977.0, 926.0, 446.0, 256.0, 152.0, 106.0, 68.0, 33.0, 36.0, 18.0, 8.0, 13.0, 8.0, 6.0, 1.0, 2.0, 0.0, 3.0, 0.0, 0.0, 1.0, 2.0], "bins": [-0.30078125, -0.29132080078125, -0.2818603515625, -0.27239990234375, -0.262939453125, -0.25347900390625, -0.2440185546875, -0.23455810546875, -0.22509765625, -0.21563720703125, -0.2061767578125, -0.19671630859375, -0.187255859375, -0.17779541015625, -0.1683349609375, -0.15887451171875, -0.1494140625, -0.13995361328125, -0.1304931640625, -0.12103271484375, -0.111572265625, -0.10211181640625, -0.0926513671875, -0.08319091796875, -0.07373046875, -0.06427001953125, -0.0548095703125, -0.04534912109375, -0.035888671875, -0.02642822265625, -0.0169677734375, -0.00750732421875, 0.001953125, 0.01141357421875, 0.0208740234375, 0.03033447265625, 0.039794921875, 0.04925537109375, 0.0587158203125, 0.06817626953125, 0.07763671875, 0.08709716796875, 0.0965576171875, 0.10601806640625, 0.115478515625, 0.12493896484375, 0.1343994140625, 0.14385986328125, 0.1533203125, 0.16278076171875, 0.1722412109375, 0.18170166015625, 0.191162109375, 0.20062255859375, 0.2100830078125, 0.21954345703125, 0.22900390625, 0.23846435546875, 0.2479248046875, 0.25738525390625, 0.266845703125, 0.27630615234375, 0.2857666015625, 0.29522705078125, 0.3046875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.19._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 3.0, 7.0, 5.0, 7.0, 16.0, 26.0, 55.0, 75.0, 108.0, 194.0, 326.0, 502.0, 728.0, 1199.0, 1654.0, 2610.0, 3470.0, 5564.0, 7270.0, 11010.0, 14778.0, 21191.0, 28235.0, 37072.0, 46768.0, 56520.0, 65216.0, 71748.0, 73963.0, 72255.0, 66756.0, 58401.0, 48369.0, 38731.0, 29692.0, 22426.0, 15635.0, 11368.0, 8267.0, 5500.0, 4082.0, 2572.0, 1846.0, 1195.0, 797.0, 543.0, 322.0, 245.0, 144.0, 79.0, 60.0, 33.0, 9.0, 11.0, 6.0, 5.0, 5.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.0556640625, -0.05390167236328125, -0.0521392822265625, -0.05037689208984375, -0.048614501953125, -0.04685211181640625, -0.0450897216796875, -0.04332733154296875, -0.04156494140625, -0.03980255126953125, -0.0380401611328125, -0.03627777099609375, -0.034515380859375, -0.03275299072265625, -0.0309906005859375, -0.02922821044921875, -0.0274658203125, -0.02570343017578125, -0.0239410400390625, -0.02217864990234375, -0.020416259765625, -0.01865386962890625, -0.0168914794921875, -0.01512908935546875, -0.01336669921875, -0.01160430908203125, -0.0098419189453125, -0.00807952880859375, -0.006317138671875, -0.00455474853515625, -0.0027923583984375, -0.00102996826171875, 0.000732421875, 0.00249481201171875, 0.0042572021484375, 0.00601959228515625, 0.007781982421875, 0.00954437255859375, 0.0113067626953125, 0.01306915283203125, 0.01483154296875, 0.01659393310546875, 0.0183563232421875, 0.02011871337890625, 0.021881103515625, 0.02364349365234375, 0.0254058837890625, 0.02716827392578125, 0.0289306640625, 0.03069305419921875, 0.0324554443359375, 0.03421783447265625, 0.035980224609375, 0.03774261474609375, 0.0395050048828125, 0.04126739501953125, 0.04302978515625, 0.04479217529296875, 0.0465545654296875, 0.04831695556640625, 0.050079345703125, 0.05184173583984375, 0.0536041259765625, 0.05536651611328125, 0.05712890625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.20._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 0.0, 0.0, 2.0, 2.0, 3.0, 2.0, 2.0, 6.0, 16.0, 16.0, 15.0, 35.0, 47.0, 70.0, 149.0, 294.0, 606.0, 1309.0, 2730.0, 5785.0, 12295.0, 26646.0, 57537.0, 116739.0, 229359.0, 447547.0, 831723.0, 1419395.0, 2245003.0, 2991648.0, 2982093.0, 2237889.0, 1417657.0, 839759.0, 447755.0, 233689.0, 118969.0, 58686.0, 27640.0, 12617.0, 5916.0, 2956.0, 1337.0, 634.0, 289.0, 164.0, 82.0, 41.0, 23.0, 15.0, 6.0, 5.0, 3.0, 2.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.21875, -0.2119140625, -0.205078125, -0.1982421875, -0.19140625, -0.1845703125, -0.177734375, -0.1708984375, -0.1640625, -0.1572265625, -0.150390625, -0.1435546875, -0.13671875, -0.1298828125, -0.123046875, -0.1162109375, -0.109375, -0.1025390625, -0.095703125, -0.0888671875, -0.08203125, -0.0751953125, -0.068359375, -0.0615234375, -0.0546875, -0.0478515625, -0.041015625, -0.0341796875, -0.02734375, -0.0205078125, -0.013671875, -0.0068359375, 0.0, 0.0068359375, 0.013671875, 0.0205078125, 0.02734375, 0.0341796875, 0.041015625, 0.0478515625, 0.0546875, 0.0615234375, 0.068359375, 0.0751953125, 0.08203125, 0.0888671875, 0.095703125, 0.1025390625, 0.109375, 0.1162109375, 0.123046875, 0.1298828125, 0.13671875, 0.1435546875, 0.150390625, 0.1572265625, 0.1640625, 0.1708984375, 0.177734375, 0.1845703125, 0.19140625, 0.1982421875, 0.205078125, 0.2119140625, 0.21875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.20._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 4.0, 1.0, 0.0, 2.0, 6.0, 4.0, 11.0, 10.0, 12.0, 30.0, 33.0, 51.0, 73.0, 96.0, 119.0, 224.0, 315.0, 571.0, 1017.0, 2123.0, 4120.0, 8071.0, 16900.0, 32038.0, 65860.0, 122281.0, 207630.0, 351851.0, 520541.0, 664812.0, 684193.0, 566820.0, 397415.0, 248364.0, 138767.0, 79692.0, 40145.0, 19335.0, 10306.0, 4758.0, 2527.0, 1379.0, 696.0, 362.0, 225.0, 167.0, 92.0, 69.0, 56.0, 42.0, 21.0, 21.0, 14.0, 8.0, 6.0, 5.0, 1.0, 2.0, 2.0, 3.0, 2.0], "bins": [-0.3125, -0.303070068359375, -0.29364013671875, -0.284210205078125, -0.2747802734375, -0.265350341796875, -0.25592041015625, -0.246490478515625, -0.237060546875, -0.227630615234375, -0.21820068359375, -0.208770751953125, -0.1993408203125, -0.189910888671875, -0.18048095703125, -0.171051025390625, -0.16162109375, -0.152191162109375, -0.14276123046875, -0.133331298828125, -0.1239013671875, -0.114471435546875, -0.10504150390625, -0.095611572265625, -0.086181640625, -0.076751708984375, -0.06732177734375, -0.057891845703125, -0.0484619140625, -0.039031982421875, -0.02960205078125, -0.020172119140625, -0.0107421875, -0.001312255859375, 0.00811767578125, 0.017547607421875, 0.0269775390625, 0.036407470703125, 0.04583740234375, 0.055267333984375, 0.064697265625, 0.074127197265625, 0.08355712890625, 0.092987060546875, 0.1024169921875, 0.111846923828125, 0.12127685546875, 0.130706787109375, 0.14013671875, 0.149566650390625, 0.15899658203125, 0.168426513671875, 0.1778564453125, 0.187286376953125, 0.19671630859375, 0.206146240234375, 0.215576171875, 0.225006103515625, 0.23443603515625, 0.243865966796875, 0.2532958984375, 0.262725830078125, 0.27215576171875, 0.281585693359375, 0.291015625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.20._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 0.0, 1.0, 1.0, 2.0, 4.0, 7.0, 3.0, 8.0, 15.0, 16.0, 25.0, 34.0, 59.0, 75.0, 118.0, 218.0, 357.0, 517.0, 971.0, 1656.0, 2778.0, 4259.0, 7414.0, 11728.0, 18787.0, 27364.0, 39705.0, 54251.0, 69358.0, 81664.0, 92576.0, 92069.0, 83625.0, 70447.0, 56330.0, 41948.0, 29485.0, 18796.0, 12606.0, 8072.0, 5081.0, 3037.0, 1731.0, 996.0, 539.0, 350.0, 236.0, 128.0, 81.0, 40.0, 44.0, 14.0, 17.0, 8.0, 8.0, 9.0, 2.0, 2.0, 2.0, 2.0, 0.0, 1.0], "bins": [-0.076171875, -0.0738677978515625, -0.071563720703125, -0.0692596435546875, -0.06695556640625, -0.0646514892578125, -0.062347412109375, -0.0600433349609375, -0.0577392578125, -0.0554351806640625, -0.053131103515625, -0.0508270263671875, -0.04852294921875, -0.0462188720703125, -0.043914794921875, -0.0416107177734375, -0.039306640625, -0.0370025634765625, -0.034698486328125, -0.0323944091796875, -0.03009033203125, -0.0277862548828125, -0.025482177734375, -0.0231781005859375, -0.0208740234375, -0.0185699462890625, -0.016265869140625, -0.0139617919921875, -0.01165771484375, -0.0093536376953125, -0.007049560546875, -0.0047454833984375, -0.00244140625, -0.0001373291015625, 0.002166748046875, 0.0044708251953125, 0.00677490234375, 0.0090789794921875, 0.011383056640625, 0.0136871337890625, 0.0159912109375, 0.0182952880859375, 0.020599365234375, 0.0229034423828125, 0.02520751953125, 0.0275115966796875, 0.029815673828125, 0.0321197509765625, 0.034423828125, 0.0367279052734375, 0.039031982421875, 0.0413360595703125, 0.04364013671875, 0.0459442138671875, 0.048248291015625, 0.0505523681640625, 0.0528564453125, 0.0551605224609375, 0.057464599609375, 0.0597686767578125, 0.06207275390625, 0.0643768310546875, 0.066680908203125, 0.0689849853515625, 0.0712890625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.21._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 2.0, 2.0, 7.0, 1.0, 8.0, 1.0, 10.0, 10.0, 24.0, 30.0, 41.0, 66.0, 77.0, 151.0, 270.0, 433.0, 956.0, 1892.0, 4401.0, 9437.0, 22817.0, 50728.0, 114803.0, 243225.0, 512405.0, 952795.0, 1674147.0, 2496631.0, 3071972.0, 2863224.0, 2097411.0, 1279534.0, 718294.0, 349939.0, 170820.0, 78845.0, 35150.0, 14484.0, 6177.0, 3126.0, 1298.0, 693.0, 328.0, 177.0, 92.0, 85.0, 54.0, 37.0, 26.0, 20.0, 13.0, 12.0, 11.0, 2.0, 8.0, 0.0, 3.0, 1.0, 2.0, 1.0, 1.0, 3.0], "bins": [-0.216796875, -0.209747314453125, -0.20269775390625, -0.195648193359375, -0.1885986328125, -0.181549072265625, -0.17449951171875, -0.167449951171875, -0.160400390625, -0.153350830078125, -0.14630126953125, -0.139251708984375, -0.1322021484375, -0.125152587890625, -0.11810302734375, -0.111053466796875, -0.10400390625, -0.096954345703125, -0.08990478515625, -0.082855224609375, -0.0758056640625, -0.068756103515625, -0.06170654296875, -0.054656982421875, -0.047607421875, -0.040557861328125, -0.03350830078125, -0.026458740234375, -0.0194091796875, -0.012359619140625, -0.00531005859375, 0.001739501953125, 0.0087890625, 0.015838623046875, 0.02288818359375, 0.029937744140625, 0.0369873046875, 0.044036865234375, 0.05108642578125, 0.058135986328125, 0.065185546875, 0.072235107421875, 0.07928466796875, 0.086334228515625, 0.0933837890625, 0.100433349609375, 0.10748291015625, 0.114532470703125, 0.12158203125, 0.128631591796875, 0.13568115234375, 0.142730712890625, 0.1497802734375, 0.156829833984375, 0.16387939453125, 0.170928955078125, 0.177978515625, 0.185028076171875, 0.19207763671875, 0.199127197265625, 0.2061767578125, 0.213226318359375, 0.22027587890625, 0.227325439453125, 0.234375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.21._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 0.0, 5.0, 4.0, 6.0, 15.0, 17.0, 34.0, 40.0, 63.0, 96.0, 129.0, 207.0, 374.0, 594.0, 1351.0, 3092.0, 6793.0, 14046.0, 32458.0, 66979.0, 134500.0, 241351.0, 408599.0, 588127.0, 712423.0, 683376.0, 530799.0, 355745.0, 203345.0, 108209.0, 55351.0, 24559.0, 11660.0, 5116.0, 2407.0, 1097.0, 532.0, 298.0, 160.0, 134.0, 59.0, 46.0, 40.0, 18.0, 15.0, 8.0, 5.0, 4.0, 4.0, 1.0, 3.0, 1.0, 0.0, 1.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.3125, -0.3023681640625, -0.292236328125, -0.2821044921875, -0.27197265625, -0.2618408203125, -0.251708984375, -0.2415771484375, -0.2314453125, -0.2213134765625, -0.211181640625, -0.2010498046875, -0.19091796875, -0.1807861328125, -0.170654296875, -0.1605224609375, -0.150390625, -0.1402587890625, -0.130126953125, -0.1199951171875, -0.10986328125, -0.0997314453125, -0.089599609375, -0.0794677734375, -0.0693359375, -0.0592041015625, -0.049072265625, -0.0389404296875, -0.02880859375, -0.0186767578125, -0.008544921875, 0.0015869140625, 0.01171875, 0.0218505859375, 0.031982421875, 0.0421142578125, 0.05224609375, 0.0623779296875, 0.072509765625, 0.0826416015625, 0.0927734375, 0.1029052734375, 0.113037109375, 0.1231689453125, 0.13330078125, 0.1434326171875, 0.153564453125, 0.1636962890625, 0.173828125, 0.1839599609375, 0.194091796875, 0.2042236328125, 0.21435546875, 0.2244873046875, 0.234619140625, 0.2447509765625, 0.2548828125, 0.2650146484375, 0.275146484375, 0.2852783203125, 0.29541015625, 0.3055419921875, 0.315673828125, 0.3258056640625, 0.3359375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.21._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 2.0, 3.0, 3.0, 6.0, 5.0, 15.0, 25.0, 60.0, 124.0, 195.0, 370.0, 615.0, 1003.0, 1783.0, 2764.0, 4413.0, 6543.0, 9443.0, 13196.0, 18354.0, 24351.0, 31260.0, 37881.0, 45653.0, 52831.0, 59512.0, 63740.0, 66521.0, 64520.0, 62160.0, 55478.0, 47910.0, 41629.0, 33885.0, 25823.0, 20917.0, 15191.0, 10962.0, 7315.0, 4650.0, 3374.0, 2145.0, 1297.0, 769.0, 462.0, 237.0, 128.0, 72.0, 38.0, 22.0, 6.0, 6.0, 4.0, 0.0, 0.0, 2.0, 2.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.04833984375, -0.046764373779296875, -0.04518890380859375, -0.043613433837890625, -0.0420379638671875, -0.040462493896484375, -0.03888702392578125, -0.037311553955078125, -0.035736083984375, -0.034160614013671875, -0.03258514404296875, -0.031009674072265625, -0.0294342041015625, -0.027858734130859375, -0.02628326416015625, -0.024707794189453125, -0.02313232421875, -0.021556854248046875, -0.01998138427734375, -0.018405914306640625, -0.0168304443359375, -0.015254974365234375, -0.01367950439453125, -0.012104034423828125, -0.010528564453125, -0.008953094482421875, -0.00737762451171875, -0.005802154541015625, -0.0042266845703125, -0.002651214599609375, -0.00107574462890625, 0.000499725341796875, 0.0020751953125, 0.003650665283203125, 0.00522613525390625, 0.006801605224609375, 0.0083770751953125, 0.009952545166015625, 0.01152801513671875, 0.013103485107421875, 0.014678955078125, 0.016254425048828125, 0.01782989501953125, 0.019405364990234375, 0.0209808349609375, 0.022556304931640625, 0.02413177490234375, 0.025707244873046875, 0.02728271484375, 0.028858184814453125, 0.03043365478515625, 0.032009124755859375, 0.0335845947265625, 0.035160064697265625, 0.03673553466796875, 0.038311004638671875, 0.039886474609375, 0.041461944580078125, 0.04303741455078125, 0.044612884521484375, 0.0461883544921875, 0.047763824462890625, 0.04933929443359375, 0.050914764404296875, 0.052490234375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.22._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 4.0, 2.0, 2.0, 5.0, 4.0, 6.0, 21.0, 8.0, 11.0, 21.0, 33.0, 56.0, 64.0, 126.0, 208.0, 356.0, 664.0, 1620.0, 3470.0, 8298.0, 21105.0, 49699.0, 109205.0, 226735.0, 452581.0, 848591.0, 1401069.0, 2182903.0, 2834468.0, 2882984.0, 2297639.0, 1542201.0, 926720.0, 515762.0, 244731.0, 125746.0, 58043.0, 24622.0, 9978.0, 3904.0, 1875.0, 737.0, 373.0, 215.0, 123.0, 75.0, 47.0, 32.0, 26.0, 17.0, 12.0, 5.0, 3.0, 5.0, 0.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.2236328125, -0.217071533203125, -0.21051025390625, -0.203948974609375, -0.1973876953125, -0.190826416015625, -0.18426513671875, -0.177703857421875, -0.171142578125, -0.164581298828125, -0.15802001953125, -0.151458740234375, -0.1448974609375, -0.138336181640625, -0.13177490234375, -0.125213623046875, -0.11865234375, -0.112091064453125, -0.10552978515625, -0.098968505859375, -0.0924072265625, -0.085845947265625, -0.07928466796875, -0.072723388671875, -0.066162109375, -0.059600830078125, -0.05303955078125, -0.046478271484375, -0.0399169921875, -0.033355712890625, -0.02679443359375, -0.020233154296875, -0.013671875, -0.007110595703125, -0.00054931640625, 0.006011962890625, 0.0125732421875, 0.019134521484375, 0.02569580078125, 0.032257080078125, 0.038818359375, 0.045379638671875, 0.05194091796875, 0.058502197265625, 0.0650634765625, 0.071624755859375, 0.07818603515625, 0.084747314453125, 0.09130859375, 0.097869873046875, 0.10443115234375, 0.110992431640625, 0.1175537109375, 0.124114990234375, 0.13067626953125, 0.137237548828125, 0.143798828125, 0.150360107421875, 0.15692138671875, 0.163482666015625, 0.1700439453125, 0.176605224609375, 0.18316650390625, 0.189727783203125, 0.1962890625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.22._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 1.0, 2.0, 3.0, 3.0, 3.0, 5.0, 12.0, 4.0, 14.0, 18.0, 29.0, 34.0, 56.0, 77.0, 93.0, 157.0, 301.0, 544.0, 1084.0, 2485.0, 5471.0, 12651.0, 29242.0, 64354.0, 130276.0, 241782.0, 427044.0, 632539.0, 766105.0, 710443.0, 515474.0, 317501.0, 171839.0, 89654.0, 41407.0, 18688.0, 8136.0, 3404.0, 1613.0, 709.0, 373.0, 238.0, 125.0, 99.0, 76.0, 43.0, 31.0, 16.0, 13.0, 12.0, 3.0, 8.0, 2.0, 1.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.33203125, -0.321563720703125, -0.31109619140625, -0.300628662109375, -0.2901611328125, -0.279693603515625, -0.26922607421875, -0.258758544921875, -0.248291015625, -0.237823486328125, -0.22735595703125, -0.216888427734375, -0.2064208984375, -0.195953369140625, -0.18548583984375, -0.175018310546875, -0.16455078125, -0.154083251953125, -0.14361572265625, -0.133148193359375, -0.1226806640625, -0.112213134765625, -0.10174560546875, -0.091278076171875, -0.080810546875, -0.070343017578125, -0.05987548828125, -0.049407958984375, -0.0389404296875, -0.028472900390625, -0.01800537109375, -0.007537841796875, 0.0029296875, 0.013397216796875, 0.02386474609375, 0.034332275390625, 0.0447998046875, 0.055267333984375, 0.06573486328125, 0.076202392578125, 0.086669921875, 0.097137451171875, 0.10760498046875, 0.118072509765625, 0.1285400390625, 0.139007568359375, 0.14947509765625, 0.159942626953125, 0.17041015625, 0.180877685546875, 0.19134521484375, 0.201812744140625, 0.2122802734375, 0.222747802734375, 0.23321533203125, 0.243682861328125, 0.254150390625, 0.264617919921875, 0.27508544921875, 0.285552978515625, 0.2960205078125, 0.306488037109375, 0.31695556640625, 0.327423095703125, 0.337890625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.22._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 5.0, 1.0, 3.0, 10.0, 7.0, 12.0, 23.0, 31.0, 53.0, 95.0, 133.0, 252.0, 415.0, 748.0, 1554.0, 3055.0, 5440.0, 9423.0, 15677.0, 25231.0, 35800.0, 50660.0, 63963.0, 77170.0, 88729.0, 92266.0, 86722.0, 76897.0, 62492.0, 49319.0, 34630.0, 24149.0, 14665.0, 9027.0, 5119.0, 2824.0, 1421.0, 699.0, 401.0, 218.0, 133.0, 79.0, 44.0, 18.0, 14.0, 18.0, 11.0, 7.0, 5.0, 4.0, 3.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.08642578125, -0.083587646484375, -0.08074951171875, -0.077911376953125, -0.0750732421875, -0.072235107421875, -0.06939697265625, -0.066558837890625, -0.063720703125, -0.060882568359375, -0.05804443359375, -0.055206298828125, -0.0523681640625, -0.049530029296875, -0.04669189453125, -0.043853759765625, -0.041015625, -0.038177490234375, -0.03533935546875, -0.032501220703125, -0.0296630859375, -0.026824951171875, -0.02398681640625, -0.021148681640625, -0.018310546875, -0.015472412109375, -0.01263427734375, -0.009796142578125, -0.0069580078125, -0.004119873046875, -0.00128173828125, 0.001556396484375, 0.00439453125, 0.007232666015625, 0.01007080078125, 0.012908935546875, 0.0157470703125, 0.018585205078125, 0.02142333984375, 0.024261474609375, 0.027099609375, 0.029937744140625, 0.03277587890625, 0.035614013671875, 0.0384521484375, 0.041290283203125, 0.04412841796875, 0.046966552734375, 0.0498046875, 0.052642822265625, 0.05548095703125, 0.058319091796875, 0.0611572265625, 0.063995361328125, 0.06683349609375, 0.069671630859375, 0.072509765625, 0.075347900390625, 0.07818603515625, 0.081024169921875, 0.0838623046875, 0.086700439453125, 0.08953857421875, 0.092376708984375, 0.09521484375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.23._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 5.0, 7.0, 4.0, 7.0, 12.0, 13.0, 23.0, 18.0, 38.0, 26.0, 71.0, 101.0, 113.0, 217.0, 288.0, 672.0, 1353.0, 3010.0, 7995.0, 20860.0, 53574.0, 131980.0, 282365.0, 593772.0, 1121066.0, 1872783.0, 2742420.0, 3222349.0, 2723660.0, 1849741.0, 1078206.0, 579342.0, 282777.0, 122832.0, 52056.0, 20088.0, 7507.0, 3129.0, 1290.0, 561.0, 324.0, 157.0, 116.0, 91.0, 50.0, 41.0, 33.0, 18.0, 20.0, 15.0, 6.0, 4.0, 3.0, 1.0, 1.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.2314453125, -0.2243194580078125, -0.217193603515625, -0.2100677490234375, -0.20294189453125, -0.1958160400390625, -0.188690185546875, -0.1815643310546875, -0.1744384765625, -0.1673126220703125, -0.160186767578125, -0.1530609130859375, -0.14593505859375, -0.1388092041015625, -0.131683349609375, -0.1245574951171875, -0.117431640625, -0.1103057861328125, -0.103179931640625, -0.0960540771484375, -0.08892822265625, -0.0818023681640625, -0.074676513671875, -0.0675506591796875, -0.0604248046875, -0.0532989501953125, -0.046173095703125, -0.0390472412109375, -0.03192138671875, -0.0247955322265625, -0.017669677734375, -0.0105438232421875, -0.00341796875, 0.0037078857421875, 0.010833740234375, 0.0179595947265625, 0.02508544921875, 0.0322113037109375, 0.039337158203125, 0.0464630126953125, 0.0535888671875, 0.0607147216796875, 0.067840576171875, 0.0749664306640625, 0.08209228515625, 0.0892181396484375, 0.096343994140625, 0.1034698486328125, 0.110595703125, 0.1177215576171875, 0.124847412109375, 0.1319732666015625, 0.13909912109375, 0.1462249755859375, 0.153350830078125, 0.1604766845703125, 0.1676025390625, 0.1747283935546875, 0.181854248046875, 0.1889801025390625, 0.19610595703125, 0.2032318115234375, 0.210357666015625, 0.2174835205078125, 0.224609375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.23._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 2.0, 4.0, 7.0, 4.0, 12.0, 13.0, 26.0, 31.0, 43.0, 84.0, 102.0, 162.0, 274.0, 548.0, 1229.0, 3207.0, 8789.0, 21899.0, 59836.0, 139983.0, 289959.0, 531743.0, 795058.0, 867359.0, 685768.0, 419103.0, 207513.0, 98837.0, 38460.0, 14641.0, 5601.0, 2101.0, 850.0, 399.0, 230.0, 157.0, 86.0, 64.0, 38.0, 21.0, 18.0, 20.0, 8.0, 4.0, 2.0, 1.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.3046875, -0.292633056640625, -0.28057861328125, -0.268524169921875, -0.2564697265625, -0.244415283203125, -0.23236083984375, -0.220306396484375, -0.208251953125, -0.196197509765625, -0.18414306640625, -0.172088623046875, -0.1600341796875, -0.147979736328125, -0.13592529296875, -0.123870849609375, -0.11181640625, -0.099761962890625, -0.08770751953125, -0.075653076171875, -0.0635986328125, -0.051544189453125, -0.03948974609375, -0.027435302734375, -0.015380859375, -0.003326416015625, 0.00872802734375, 0.020782470703125, 0.0328369140625, 0.044891357421875, 0.05694580078125, 0.069000244140625, 0.0810546875, 0.093109130859375, 0.10516357421875, 0.117218017578125, 0.1292724609375, 0.141326904296875, 0.15338134765625, 0.165435791015625, 0.177490234375, 0.189544677734375, 0.20159912109375, 0.213653564453125, 0.2257080078125, 0.237762451171875, 0.24981689453125, 0.261871337890625, 0.27392578125, 0.285980224609375, 0.29803466796875, 0.310089111328125, 0.3221435546875, 0.334197998046875, 0.34625244140625, 0.358306884765625, 0.370361328125, 0.382415771484375, 0.39447021484375, 0.406524658203125, 0.4185791015625, 0.430633544921875, 0.44268798828125, 0.454742431640625, 0.466796875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.23._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 2.0, 0.0, 0.0, 2.0, 5.0, 4.0, 5.0, 3.0, 5.0, 13.0, 11.0, 8.0, 27.0, 39.0, 76.0, 116.0, 212.0, 364.0, 766.0, 1387.0, 2550.0, 4652.0, 8163.0, 13157.0, 20169.0, 29666.0, 41493.0, 53895.0, 66214.0, 77203.0, 85222.0, 86146.0, 79625.0, 70089.0, 58040.0, 45675.0, 33561.0, 23195.0, 15453.0, 9752.0, 5655.0, 3197.0, 1788.0, 922.0, 489.0, 279.0, 142.0, 81.0, 53.0, 33.0, 24.0, 17.0, 9.0, 13.0, 4.0, 3.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.10302734375, -0.10009765625, -0.09716796875, -0.09423828125, -0.09130859375, -0.08837890625, -0.08544921875, -0.08251953125, -0.07958984375, -0.07666015625, -0.07373046875, -0.07080078125, -0.06787109375, -0.06494140625, -0.06201171875, -0.05908203125, -0.05615234375, -0.05322265625, -0.05029296875, -0.04736328125, -0.04443359375, -0.04150390625, -0.03857421875, -0.03564453125, -0.03271484375, -0.02978515625, -0.02685546875, -0.02392578125, -0.02099609375, -0.01806640625, -0.01513671875, -0.01220703125, -0.00927734375, -0.00634765625, -0.00341796875, -0.00048828125, 0.00244140625, 0.00537109375, 0.00830078125, 0.01123046875, 0.01416015625, 0.01708984375, 0.02001953125, 0.02294921875, 0.02587890625, 0.02880859375, 0.03173828125, 0.03466796875, 0.03759765625, 0.04052734375, 0.04345703125, 0.04638671875, 0.04931640625, 0.05224609375, 0.05517578125, 0.05810546875, 0.06103515625, 0.06396484375, 0.06689453125, 0.06982421875, 0.07275390625, 0.07568359375, 0.07861328125, 0.08154296875, 0.08447265625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.24._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 3.0, 1.0, 3.0, 3.0, 6.0, 10.0, 10.0, 9.0, 24.0, 34.0, 48.0, 76.0, 99.0, 114.0, 172.0, 354.0, 570.0, 1075.0, 2244.0, 5145.0, 13502.0, 33269.0, 77864.0, 180709.0, 363509.0, 751505.0, 1301051.0, 2140295.0, 2994812.0, 3143102.0, 2427243.0, 1564948.0, 896041.0, 473918.0, 224209.0, 105735.0, 44262.0, 18015.0, 6956.0, 3217.0, 1376.0, 610.0, 344.0, 230.0, 157.0, 87.0, 68.0, 56.0, 39.0, 24.0, 15.0, 9.0, 12.0, 7.0, 5.0, 3.0, 1.0, 2.0, 3.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.2080078125, -0.20111083984375, -0.1942138671875, -0.18731689453125, -0.180419921875, -0.17352294921875, -0.1666259765625, -0.15972900390625, -0.15283203125, -0.14593505859375, -0.1390380859375, -0.13214111328125, -0.125244140625, -0.11834716796875, -0.1114501953125, -0.10455322265625, -0.09765625, -0.09075927734375, -0.0838623046875, -0.07696533203125, -0.070068359375, -0.06317138671875, -0.0562744140625, -0.04937744140625, -0.04248046875, -0.03558349609375, -0.0286865234375, -0.02178955078125, -0.014892578125, -0.00799560546875, -0.0010986328125, 0.00579833984375, 0.0126953125, 0.01959228515625, 0.0264892578125, 0.03338623046875, 0.040283203125, 0.04718017578125, 0.0540771484375, 0.06097412109375, 0.06787109375, 0.07476806640625, 0.0816650390625, 0.08856201171875, 0.095458984375, 0.10235595703125, 0.1092529296875, 0.11614990234375, 0.123046875, 0.12994384765625, 0.1368408203125, 0.14373779296875, 0.150634765625, 0.15753173828125, 0.1644287109375, 0.17132568359375, 0.17822265625, 0.18511962890625, 0.1920166015625, 0.19891357421875, 0.205810546875, 0.21270751953125, 0.2196044921875, 0.22650146484375, 0.2333984375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.24._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 5.0, 5.0, 3.0, 2.0, 5.0, 7.0, 13.0, 5.0, 16.0, 31.0, 41.0, 51.0, 55.0, 89.0, 129.0, 153.0, 281.0, 439.0, 889.0, 1994.0, 4321.0, 9415.0, 20986.0, 46423.0, 98424.0, 186597.0, 360019.0, 592231.0, 807625.0, 792874.0, 572979.0, 344346.0, 181930.0, 91431.0, 44011.0, 19901.0, 8773.0, 3871.0, 1721.0, 860.0, 513.0, 261.0, 155.0, 120.0, 84.0, 57.0, 37.0, 33.0, 25.0, 14.0, 17.0, 9.0, 7.0, 7.0, 1.0, 2.0, 0.0, 2.0, 2.0, 1.0, 2.0, 1.0], "bins": [-0.3203125, -0.309967041015625, -0.29962158203125, -0.289276123046875, -0.2789306640625, -0.268585205078125, -0.25823974609375, -0.247894287109375, -0.237548828125, -0.227203369140625, -0.21685791015625, -0.206512451171875, -0.1961669921875, -0.185821533203125, -0.17547607421875, -0.165130615234375, -0.15478515625, -0.144439697265625, -0.13409423828125, -0.123748779296875, -0.1134033203125, -0.103057861328125, -0.09271240234375, -0.082366943359375, -0.072021484375, -0.061676025390625, -0.05133056640625, -0.040985107421875, -0.0306396484375, -0.020294189453125, -0.00994873046875, 0.000396728515625, 0.0107421875, 0.021087646484375, 0.03143310546875, 0.041778564453125, 0.0521240234375, 0.062469482421875, 0.07281494140625, 0.083160400390625, 0.093505859375, 0.103851318359375, 0.11419677734375, 0.124542236328125, 0.1348876953125, 0.145233154296875, 0.15557861328125, 0.165924072265625, 0.17626953125, 0.186614990234375, 0.19696044921875, 0.207305908203125, 0.2176513671875, 0.227996826171875, 0.23834228515625, 0.248687744140625, 0.259033203125, 0.269378662109375, 0.27972412109375, 0.290069580078125, 0.3004150390625, 0.310760498046875, 0.32110595703125, 0.331451416015625, 0.341796875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.24._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 5.0, 5.0, 10.0, 27.0, 59.0, 85.0, 153.0, 282.0, 487.0, 796.0, 1337.0, 1955.0, 3078.0, 4572.0, 6798.0, 9610.0, 13015.0, 16111.0, 21977.0, 27231.0, 34050.0, 40973.0, 45391.0, 52735.0, 58092.0, 63035.0, 63952.0, 61013.0, 55832.0, 49872.0, 42822.0, 37922.0, 31581.0, 23697.0, 19859.0, 15245.0, 11475.0, 8305.0, 5487.0, 3895.0, 2528.0, 1699.0, 1053.0, 669.0, 375.0, 238.0, 132.0, 71.0, 39.0, 24.0, 7.0, 6.0, 5.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0732421875, -0.07082366943359375, -0.0684051513671875, -0.06598663330078125, -0.063568115234375, -0.06114959716796875, -0.0587310791015625, -0.05631256103515625, -0.05389404296875, -0.05147552490234375, -0.0490570068359375, -0.04663848876953125, -0.044219970703125, -0.04180145263671875, -0.0393829345703125, -0.03696441650390625, -0.0345458984375, -0.03212738037109375, -0.0297088623046875, -0.02729034423828125, -0.024871826171875, -0.02245330810546875, -0.0200347900390625, -0.01761627197265625, -0.01519775390625, -0.01277923583984375, -0.0103607177734375, -0.00794219970703125, -0.005523681640625, -0.00310516357421875, -0.0006866455078125, 0.00173187255859375, 0.004150390625, 0.00656890869140625, 0.0089874267578125, 0.01140594482421875, 0.013824462890625, 0.01624298095703125, 0.0186614990234375, 0.02108001708984375, 0.02349853515625, 0.02591705322265625, 0.0283355712890625, 0.03075408935546875, 0.033172607421875, 0.03559112548828125, 0.0380096435546875, 0.04042816162109375, 0.0428466796875, 0.04526519775390625, 0.0476837158203125, 0.05010223388671875, 0.052520751953125, 0.05493927001953125, 0.0573577880859375, 0.05977630615234375, 0.06219482421875, 0.06461334228515625, 0.0670318603515625, 0.06945037841796875, 0.071868896484375, 0.07428741455078125, 0.0767059326171875, 0.07912445068359375, 0.08154296875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.25._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 1.0, 3.0, 0.0, 4.0, 10.0, 6.0, 10.0, 14.0, 22.0, 34.0, 47.0, 74.0, 90.0, 105.0, 215.0, 268.0, 487.0, 754.0, 1472.0, 3219.0, 7973.0, 21690.0, 55649.0, 125890.0, 293822.0, 638187.0, 1169223.0, 2029104.0, 2904984.0, 3235134.0, 2652332.0, 1737795.0, 996748.0, 508413.0, 221506.0, 102264.0, 41727.0, 15995.0, 6178.0, 2509.0, 1265.0, 682.0, 432.0, 283.0, 156.0, 134.0, 81.0, 58.0, 50.0, 30.0, 23.0, 24.0, 4.0, 8.0, 7.0, 5.0, 6.0, 3.0, 2.0, 0.0, 2.0], "bins": [-0.2275390625, -0.2205047607421875, -0.213470458984375, -0.2064361572265625, -0.19940185546875, -0.1923675537109375, -0.185333251953125, -0.1782989501953125, -0.1712646484375, -0.1642303466796875, -0.157196044921875, -0.1501617431640625, -0.14312744140625, -0.1360931396484375, -0.129058837890625, -0.1220245361328125, -0.114990234375, -0.1079559326171875, -0.100921630859375, -0.0938873291015625, -0.08685302734375, -0.0798187255859375, -0.072784423828125, -0.0657501220703125, -0.0587158203125, -0.0516815185546875, -0.044647216796875, -0.0376129150390625, -0.03057861328125, -0.0235443115234375, -0.016510009765625, -0.0094757080078125, -0.00244140625, 0.0045928955078125, 0.011627197265625, 0.0186614990234375, 0.02569580078125, 0.0327301025390625, 0.039764404296875, 0.0467987060546875, 0.0538330078125, 0.0608673095703125, 0.067901611328125, 0.0749359130859375, 0.08197021484375, 0.0890045166015625, 0.096038818359375, 0.1030731201171875, 0.110107421875, 0.1171417236328125, 0.124176025390625, 0.1312103271484375, 0.13824462890625, 0.1452789306640625, 0.152313232421875, 0.1593475341796875, 0.1663818359375, 0.1734161376953125, 0.180450439453125, 0.1874847412109375, 0.19451904296875, 0.2015533447265625, 0.208587646484375, 0.2156219482421875, 0.22265625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.25._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 5.0, 1.0, 3.0, 3.0, 4.0, 8.0, 6.0, 6.0, 20.0, 23.0, 26.0, 51.0, 55.0, 79.0, 149.0, 196.0, 368.0, 870.0, 2269.0, 6503.0, 17729.0, 50626.0, 120144.0, 263687.0, 516573.0, 876796.0, 998688.0, 671262.0, 368221.0, 176455.0, 76830.0, 29457.0, 10639.0, 3881.0, 1244.0, 580.0, 290.0, 169.0, 116.0, 74.0, 56.0, 39.0, 19.0, 19.0, 13.0, 16.0, 15.0, 5.0, 5.0, 4.0, 2.0, 2.0, 1.0], "bins": [-0.458984375, -0.4466552734375, -0.434326171875, -0.4219970703125, -0.40966796875, -0.3973388671875, -0.385009765625, -0.3726806640625, -0.3603515625, -0.3480224609375, -0.335693359375, -0.3233642578125, -0.31103515625, -0.2987060546875, -0.286376953125, -0.2740478515625, -0.26171875, -0.2493896484375, -0.237060546875, -0.2247314453125, -0.21240234375, -0.2000732421875, -0.187744140625, -0.1754150390625, -0.1630859375, -0.1507568359375, -0.138427734375, -0.1260986328125, -0.11376953125, -0.1014404296875, -0.089111328125, -0.0767822265625, -0.064453125, -0.0521240234375, -0.039794921875, -0.0274658203125, -0.01513671875, -0.0028076171875, 0.009521484375, 0.0218505859375, 0.0341796875, 0.0465087890625, 0.058837890625, 0.0711669921875, 0.08349609375, 0.0958251953125, 0.108154296875, 0.1204833984375, 0.1328125, 0.1451416015625, 0.157470703125, 0.1697998046875, 0.18212890625, 0.1944580078125, 0.206787109375, 0.2191162109375, 0.2314453125, 0.2437744140625, 0.256103515625, 0.2684326171875, 0.28076171875, 0.2930908203125, 0.305419921875, 0.3177490234375, 0.330078125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.25._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 5.0, 3.0, 6.0, 5.0, 5.0, 6.0, 12.0, 17.0, 19.0, 34.0, 50.0, 60.0, 93.0, 140.0, 226.0, 429.0, 699.0, 1235.0, 2355.0, 4624.0, 8059.0, 14252.0, 23065.0, 33404.0, 49081.0, 62221.0, 77458.0, 89478.0, 93124.0, 89271.0, 80048.0, 64480.0, 51183.0, 35319.0, 24146.0, 15155.0, 8815.0, 5062.0, 2718.0, 1335.0, 770.0, 462.0, 228.0, 168.0, 104.0, 66.0, 50.0, 36.0, 19.0, 19.0, 17.0, 9.0, 7.0, 3.0, 7.0, 5.0, 2.0, 5.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.099609375, -0.09635162353515625, -0.0930938720703125, -0.08983612060546875, -0.086578369140625, -0.08332061767578125, -0.0800628662109375, -0.07680511474609375, -0.07354736328125, -0.07028961181640625, -0.0670318603515625, -0.06377410888671875, -0.060516357421875, -0.05725860595703125, -0.0540008544921875, -0.05074310302734375, -0.0474853515625, -0.04422760009765625, -0.0409698486328125, -0.03771209716796875, -0.034454345703125, -0.03119659423828125, -0.0279388427734375, -0.02468109130859375, -0.02142333984375, -0.01816558837890625, -0.0149078369140625, -0.01165008544921875, -0.008392333984375, -0.00513458251953125, -0.0018768310546875, 0.00138092041015625, 0.004638671875, 0.00789642333984375, 0.0111541748046875, 0.01441192626953125, 0.017669677734375, 0.02092742919921875, 0.0241851806640625, 0.02744293212890625, 0.03070068359375, 0.03395843505859375, 0.0372161865234375, 0.04047393798828125, 0.043731689453125, 0.04698944091796875, 0.0502471923828125, 0.05350494384765625, 0.0567626953125, 0.06002044677734375, 0.0632781982421875, 0.06653594970703125, 0.069793701171875, 0.07305145263671875, 0.0763092041015625, 0.07956695556640625, 0.08282470703125, 0.08608245849609375, 0.0893402099609375, 0.09259796142578125, 0.095855712890625, 0.09911346435546875, 0.1023712158203125, 0.10562896728515625, 0.10888671875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.26._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 1.0, 3.0, 2.0, 3.0, 6.0, 6.0, 10.0, 11.0, 20.0, 33.0, 49.0, 88.0, 101.0, 150.0, 214.0, 352.0, 506.0, 1004.0, 1857.0, 4236.0, 10169.0, 25415.0, 59901.0, 128244.0, 276378.0, 584621.0, 1053885.0, 1801378.0, 2636877.0, 3120975.0, 2750411.0, 1941542.0, 1149261.0, 651246.0, 308811.0, 151840.0, 66234.0, 30270.0, 11846.0, 4390.0, 2246.0, 1102.0, 607.0, 288.0, 202.0, 140.0, 100.0, 52.0, 35.0, 31.0, 16.0, 11.0, 10.0, 5.0, 6.0, 5.0, 3.0, 4.0, 2.0, 0.0, 2.0], "bins": [-0.2197265625, -0.212982177734375, -0.20623779296875, -0.199493408203125, -0.1927490234375, -0.186004638671875, -0.17926025390625, -0.172515869140625, -0.165771484375, -0.159027099609375, -0.15228271484375, -0.145538330078125, -0.1387939453125, -0.132049560546875, -0.12530517578125, -0.118560791015625, -0.11181640625, -0.105072021484375, -0.09832763671875, -0.091583251953125, -0.0848388671875, -0.078094482421875, -0.07135009765625, -0.064605712890625, -0.057861328125, -0.051116943359375, -0.04437255859375, -0.037628173828125, -0.0308837890625, -0.024139404296875, -0.01739501953125, -0.010650634765625, -0.00390625, 0.002838134765625, 0.00958251953125, 0.016326904296875, 0.0230712890625, 0.029815673828125, 0.03656005859375, 0.043304443359375, 0.050048828125, 0.056793212890625, 0.06353759765625, 0.070281982421875, 0.0770263671875, 0.083770751953125, 0.09051513671875, 0.097259521484375, 0.10400390625, 0.110748291015625, 0.11749267578125, 0.124237060546875, 0.1309814453125, 0.137725830078125, 0.14447021484375, 0.151214599609375, 0.157958984375, 0.164703369140625, 0.17144775390625, 0.178192138671875, 0.1849365234375, 0.191680908203125, 0.19842529296875, 0.205169677734375, 0.2119140625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.26._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 4.0, 3.0, 5.0, 3.0, 6.0, 16.0, 13.0, 30.0, 25.0, 43.0, 58.0, 90.0, 109.0, 195.0, 260.0, 460.0, 880.0, 1918.0, 4306.0, 10323.0, 21180.0, 49114.0, 103855.0, 200535.0, 383754.0, 635534.0, 854099.0, 790999.0, 531496.0, 305441.0, 156455.0, 77149.0, 36573.0, 15795.0, 7192.0, 3268.0, 1448.0, 628.0, 350.0, 196.0, 133.0, 114.0, 74.0, 48.0, 33.0, 33.0, 10.0, 12.0, 10.0, 10.0, 3.0, 4.0, 5.0, 1.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.357421875, -0.346527099609375, -0.33563232421875, -0.324737548828125, -0.3138427734375, -0.302947998046875, -0.29205322265625, -0.281158447265625, -0.270263671875, -0.259368896484375, -0.24847412109375, -0.237579345703125, -0.2266845703125, -0.215789794921875, -0.20489501953125, -0.194000244140625, -0.18310546875, -0.172210693359375, -0.16131591796875, -0.150421142578125, -0.1395263671875, -0.128631591796875, -0.11773681640625, -0.106842041015625, -0.095947265625, -0.085052490234375, -0.07415771484375, -0.063262939453125, -0.0523681640625, -0.041473388671875, -0.03057861328125, -0.019683837890625, -0.0087890625, 0.002105712890625, 0.01300048828125, 0.023895263671875, 0.0347900390625, 0.045684814453125, 0.05657958984375, 0.067474365234375, 0.078369140625, 0.089263916015625, 0.10015869140625, 0.111053466796875, 0.1219482421875, 0.132843017578125, 0.14373779296875, 0.154632568359375, 0.16552734375, 0.176422119140625, 0.18731689453125, 0.198211669921875, 0.2091064453125, 0.220001220703125, 0.23089599609375, 0.241790771484375, 0.252685546875, 0.263580322265625, 0.27447509765625, 0.285369873046875, 0.2962646484375, 0.307159423828125, 0.31805419921875, 0.328948974609375, 0.33984375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.26._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 7.0, 4.0, 10.0, 17.0, 30.0, 41.0, 95.0, 141.0, 272.0, 436.0, 670.0, 1182.0, 1886.0, 2822.0, 4551.0, 6245.0, 9493.0, 12072.0, 17349.0, 22069.0, 28982.0, 35901.0, 43221.0, 51321.0, 57284.0, 63891.0, 66066.0, 66389.0, 61785.0, 55855.0, 48875.0, 41989.0, 35183.0, 27768.0, 21896.0, 15824.0, 12444.0, 8248.0, 6190.0, 3982.0, 2770.0, 1570.0, 1191.0, 675.0, 394.0, 251.0, 148.0, 86.0, 51.0, 25.0, 15.0, 6.0, 3.0, 3.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.0703125, -0.06817626953125, -0.0660400390625, -0.06390380859375, -0.061767578125, -0.05963134765625, -0.0574951171875, -0.05535888671875, -0.05322265625, -0.05108642578125, -0.0489501953125, -0.04681396484375, -0.044677734375, -0.04254150390625, -0.0404052734375, -0.03826904296875, -0.0361328125, -0.03399658203125, -0.0318603515625, -0.02972412109375, -0.027587890625, -0.02545166015625, -0.0233154296875, -0.02117919921875, -0.01904296875, -0.01690673828125, -0.0147705078125, -0.01263427734375, -0.010498046875, -0.00836181640625, -0.0062255859375, -0.00408935546875, -0.001953125, 0.00018310546875, 0.0023193359375, 0.00445556640625, 0.006591796875, 0.00872802734375, 0.0108642578125, 0.01300048828125, 0.01513671875, 0.01727294921875, 0.0194091796875, 0.02154541015625, 0.023681640625, 0.02581787109375, 0.0279541015625, 0.03009033203125, 0.0322265625, 0.03436279296875, 0.0364990234375, 0.03863525390625, 0.040771484375, 0.04290771484375, 0.0450439453125, 0.04718017578125, 0.04931640625, 0.05145263671875, 0.0535888671875, 0.05572509765625, 0.057861328125, 0.05999755859375, 0.0621337890625, 0.06427001953125, 0.06640625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.27._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [3.0, 0.0, 2.0, 0.0, 3.0, 5.0, 7.0, 7.0, 11.0, 17.0, 15.0, 22.0, 35.0, 50.0, 49.0, 123.0, 202.0, 260.0, 372.0, 621.0, 1148.0, 2525.0, 5504.0, 14263.0, 35655.0, 84796.0, 199117.0, 413952.0, 826611.0, 1487835.0, 2383969.0, 3170477.0, 3093615.0, 2250370.0, 1372793.0, 756110.0, 370405.0, 178199.0, 75098.0, 31055.0, 12075.0, 4769.0, 2346.0, 1040.0, 587.0, 350.0, 237.0, 157.0, 121.0, 65.0, 46.0, 37.0, 24.0, 22.0, 19.0, 3.0, 4.0, 4.0, 3.0, 2.0, 2.0, 1.0, 0.0, 1.0], "bins": [-0.2197265625, -0.2128448486328125, -0.205963134765625, -0.1990814208984375, -0.19219970703125, -0.1853179931640625, -0.178436279296875, -0.1715545654296875, -0.1646728515625, -0.1577911376953125, -0.150909423828125, -0.1440277099609375, -0.13714599609375, -0.1302642822265625, -0.123382568359375, -0.1165008544921875, -0.109619140625, -0.1027374267578125, -0.095855712890625, -0.0889739990234375, -0.08209228515625, -0.0752105712890625, -0.068328857421875, -0.0614471435546875, -0.0545654296875, -0.0476837158203125, -0.040802001953125, -0.0339202880859375, -0.02703857421875, -0.0201568603515625, -0.013275146484375, -0.0063934326171875, 0.00048828125, 0.0073699951171875, 0.014251708984375, 0.0211334228515625, 0.02801513671875, 0.0348968505859375, 0.041778564453125, 0.0486602783203125, 0.0555419921875, 0.0624237060546875, 0.069305419921875, 0.0761871337890625, 0.08306884765625, 0.0899505615234375, 0.096832275390625, 0.1037139892578125, 0.110595703125, 0.1174774169921875, 0.124359130859375, 0.1312408447265625, 0.13812255859375, 0.1450042724609375, 0.151885986328125, 0.1587677001953125, 0.1656494140625, 0.1725311279296875, 0.179412841796875, 0.1862945556640625, 0.19317626953125, 0.2000579833984375, 0.206939697265625, 0.2138214111328125, 0.220703125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.27._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 8.0, 6.0, 11.0, 9.0, 24.0, 22.0, 31.0, 51.0, 74.0, 110.0, 139.0, 212.0, 335.0, 569.0, 1075.0, 2621.0, 6093.0, 16307.0, 39361.0, 99511.0, 213746.0, 443503.0, 787473.0, 1023307.0, 764145.0, 431754.0, 203075.0, 96241.0, 38717.0, 14650.0, 5870.0, 2639.0, 1115.0, 554.0, 310.0, 191.0, 133.0, 93.0, 59.0, 44.0, 32.0, 19.0, 19.0, 10.0, 9.0, 6.0, 3.0, 2.0, 3.0, 1.0, 2.0, 0.0, 1.0, 1.0, 2.0], "bins": [-0.40234375, -0.38995361328125, -0.3775634765625, -0.36517333984375, -0.352783203125, -0.34039306640625, -0.3280029296875, -0.31561279296875, -0.30322265625, -0.29083251953125, -0.2784423828125, -0.26605224609375, -0.253662109375, -0.24127197265625, -0.2288818359375, -0.21649169921875, -0.2041015625, -0.19171142578125, -0.1793212890625, -0.16693115234375, -0.154541015625, -0.14215087890625, -0.1297607421875, -0.11737060546875, -0.10498046875, -0.09259033203125, -0.0802001953125, -0.06781005859375, -0.055419921875, -0.04302978515625, -0.0306396484375, -0.01824951171875, -0.005859375, 0.00653076171875, 0.0189208984375, 0.03131103515625, 0.043701171875, 0.05609130859375, 0.0684814453125, 0.08087158203125, 0.09326171875, 0.10565185546875, 0.1180419921875, 0.13043212890625, 0.142822265625, 0.15521240234375, 0.1676025390625, 0.17999267578125, 0.1923828125, 0.20477294921875, 0.2171630859375, 0.22955322265625, 0.241943359375, 0.25433349609375, 0.2667236328125, 0.27911376953125, 0.29150390625, 0.30389404296875, 0.3162841796875, 0.32867431640625, 0.341064453125, 0.35345458984375, 0.3658447265625, 0.37823486328125, 0.390625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.27._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 3.0, 7.0, 7.0, 11.0, 14.0, 24.0, 41.0, 72.0, 89.0, 163.0, 273.0, 587.0, 1258.0, 3619.0, 10006.0, 28470.0, 61303.0, 110292.0, 156514.0, 171093.0, 137163.0, 87055.0, 43190.0, 18177.0, 6217.0, 2168.0, 857.0, 411.0, 227.0, 130.0, 97.0, 51.0, 26.0, 19.0, 14.0, 9.0, 7.0, 4.0, 2.0, 2.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.1884765625, -0.1818084716796875, -0.175140380859375, -0.1684722900390625, -0.16180419921875, -0.1551361083984375, -0.148468017578125, -0.1417999267578125, -0.1351318359375, -0.1284637451171875, -0.121795654296875, -0.1151275634765625, -0.10845947265625, -0.1017913818359375, -0.095123291015625, -0.0884552001953125, -0.081787109375, -0.0751190185546875, -0.068450927734375, -0.0617828369140625, -0.05511474609375, -0.0484466552734375, -0.041778564453125, -0.0351104736328125, -0.0284423828125, -0.0217742919921875, -0.015106201171875, -0.0084381103515625, -0.00177001953125, 0.0048980712890625, 0.011566162109375, 0.0182342529296875, 0.02490234375, 0.0315704345703125, 0.038238525390625, 0.0449066162109375, 0.05157470703125, 0.0582427978515625, 0.064910888671875, 0.0715789794921875, 0.0782470703125, 0.0849151611328125, 0.091583251953125, 0.0982513427734375, 0.10491943359375, 0.1115875244140625, 0.118255615234375, 0.1249237060546875, 0.131591796875, 0.1382598876953125, 0.144927978515625, 0.1515960693359375, 0.15826416015625, 0.1649322509765625, 0.171600341796875, 0.1782684326171875, 0.1849365234375, 0.1916046142578125, 0.198272705078125, 0.2049407958984375, 0.21160888671875, 0.2182769775390625, 0.224945068359375, 0.2316131591796875, 0.23828125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.28._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 0.0, 6.0, 8.0, 10.0, 15.0, 16.0, 26.0, 24.0, 46.0, 75.0, 111.0, 192.0, 365.0, 668.0, 1606.0, 4154.0, 13914.0, 42531.0, 131441.0, 347645.0, 872846.0, 1790290.0, 3073219.0, 3918276.0, 3175001.0, 1876328.0, 935423.0, 378377.0, 143535.0, 47674.0, 15188.0, 4653.0, 1851.0, 678.0, 405.0, 193.0, 140.0, 82.0, 58.0, 49.0, 39.0, 11.0, 7.0, 11.0, 4.0, 6.0, 5.0, 3.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0], "bins": [-0.265625, -0.2572021484375, -0.248779296875, -0.2403564453125, -0.23193359375, -0.2235107421875, -0.215087890625, -0.2066650390625, -0.1982421875, -0.1898193359375, -0.181396484375, -0.1729736328125, -0.16455078125, -0.1561279296875, -0.147705078125, -0.1392822265625, -0.130859375, -0.1224365234375, -0.114013671875, -0.1055908203125, -0.09716796875, -0.0887451171875, -0.080322265625, -0.0718994140625, -0.0634765625, -0.0550537109375, -0.046630859375, -0.0382080078125, -0.02978515625, -0.0213623046875, -0.012939453125, -0.0045166015625, 0.00390625, 0.0123291015625, 0.020751953125, 0.0291748046875, 0.03759765625, 0.0460205078125, 0.054443359375, 0.0628662109375, 0.0712890625, 0.0797119140625, 0.088134765625, 0.0965576171875, 0.10498046875, 0.1134033203125, 0.121826171875, 0.1302490234375, 0.138671875, 0.1470947265625, 0.155517578125, 0.1639404296875, 0.17236328125, 0.1807861328125, 0.189208984375, 0.1976318359375, 0.2060546875, 0.2144775390625, 0.222900390625, 0.2313232421875, 0.23974609375, 0.2481689453125, 0.256591796875, 0.2650146484375, 0.2734375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.28._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 1.0, 3.0, 2.0, 2.0, 9.0, 7.0, 12.0, 24.0, 14.0, 31.0, 37.0, 69.0, 91.0, 115.0, 171.0, 230.0, 313.0, 532.0, 889.0, 1867.0, 3728.0, 8309.0, 18999.0, 41329.0, 83336.0, 154990.0, 281756.0, 431171.0, 605352.0, 696318.0, 646312.0, 494439.0, 329768.0, 191660.0, 103710.0, 52824.0, 25026.0, 10662.0, 4913.0, 2203.0, 1141.0, 658.0, 361.0, 283.0, 173.0, 125.0, 87.0, 67.0, 52.0, 44.0, 25.0, 20.0, 10.0, 9.0, 8.0, 4.0, 2.0, 1.0, 2.0, 1.0, 3.0, 1.0], "bins": [-0.279296875, -0.270477294921875, -0.26165771484375, -0.252838134765625, -0.2440185546875, -0.235198974609375, -0.22637939453125, -0.217559814453125, -0.208740234375, -0.199920654296875, -0.19110107421875, -0.182281494140625, -0.1734619140625, -0.164642333984375, -0.15582275390625, -0.147003173828125, -0.13818359375, -0.129364013671875, -0.12054443359375, -0.111724853515625, -0.1029052734375, -0.094085693359375, -0.08526611328125, -0.076446533203125, -0.067626953125, -0.058807373046875, -0.04998779296875, -0.041168212890625, -0.0323486328125, -0.023529052734375, -0.01470947265625, -0.005889892578125, 0.0029296875, 0.011749267578125, 0.02056884765625, 0.029388427734375, 0.0382080078125, 0.047027587890625, 0.05584716796875, 0.064666748046875, 0.073486328125, 0.082305908203125, 0.09112548828125, 0.099945068359375, 0.1087646484375, 0.117584228515625, 0.12640380859375, 0.135223388671875, 0.14404296875, 0.152862548828125, 0.16168212890625, 0.170501708984375, 0.1793212890625, 0.188140869140625, 0.19696044921875, 0.205780029296875, 0.214599609375, 0.223419189453125, 0.23223876953125, 0.241058349609375, 0.2498779296875, 0.258697509765625, 0.26751708984375, 0.276336669921875, 0.28515625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.28._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 4.0, 0.0, 7.0, 7.0, 19.0, 41.0, 62.0, 104.0, 132.0, 279.0, 410.0, 579.0, 967.0, 1329.0, 2074.0, 2975.0, 3666.0, 5442.0, 7245.0, 9774.0, 13251.0, 15886.0, 19699.0, 23972.0, 29066.0, 34882.0, 39352.0, 43867.0, 49133.0, 53094.0, 56831.0, 57259.0, 53737.0, 49727.0, 44822.0, 39656.0, 36287.0, 30578.0, 24977.0, 20573.0, 16340.0, 14189.0, 10220.0, 7700.0, 5903.0, 3906.0, 3142.0, 2272.0, 1362.0, 1021.0, 631.0, 483.0, 282.0, 171.0, 122.0, 56.0, 50.0, 26.0, 15.0, 11.0, 6.0, 1.0, 4.0], "bins": [-0.0751953125, -0.07285308837890625, -0.0705108642578125, -0.06816864013671875, -0.065826416015625, -0.06348419189453125, -0.0611419677734375, -0.05879974365234375, -0.05645751953125, -0.05411529541015625, -0.0517730712890625, -0.04943084716796875, -0.047088623046875, -0.04474639892578125, -0.0424041748046875, -0.04006195068359375, -0.0377197265625, -0.03537750244140625, -0.0330352783203125, -0.03069305419921875, -0.028350830078125, -0.02600860595703125, -0.0236663818359375, -0.02132415771484375, -0.01898193359375, -0.01663970947265625, -0.0142974853515625, -0.01195526123046875, -0.009613037109375, -0.00727081298828125, -0.0049285888671875, -0.00258636474609375, -0.000244140625, 0.00209808349609375, 0.0044403076171875, 0.00678253173828125, 0.009124755859375, 0.01146697998046875, 0.0138092041015625, 0.01615142822265625, 0.01849365234375, 0.02083587646484375, 0.0231781005859375, 0.02552032470703125, 0.027862548828125, 0.03020477294921875, 0.0325469970703125, 0.03488922119140625, 0.0372314453125, 0.03957366943359375, 0.0419158935546875, 0.04425811767578125, 0.046600341796875, 0.04894256591796875, 0.0512847900390625, 0.05362701416015625, 0.05596923828125, 0.05831146240234375, 0.0606536865234375, 0.06299591064453125, 0.065338134765625, 0.06768035888671875, 0.0700225830078125, 0.07236480712890625, 0.07470703125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.29._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 4.0, 4.0, 10.0, 6.0, 6.0, 17.0, 21.0, 40.0, 74.0, 84.0, 146.0, 268.0, 381.0, 647.0, 1246.0, 2476.0, 6504.0, 20480.0, 64968.0, 189888.0, 523140.0, 1228650.0, 2434578.0, 3686600.0, 3755140.0, 2590991.0, 1350650.0, 588797.0, 218252.0, 75488.0, 24150.0, 7518.0, 2865.0, 1312.0, 732.0, 364.0, 278.0, 158.0, 87.0, 62.0, 38.0, 31.0, 14.0, 14.0, 4.0, 13.0, 7.0, 1.0, 2.0, 2.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.291015625, -0.282470703125, -0.27392578125, -0.265380859375, -0.2568359375, -0.248291015625, -0.23974609375, -0.231201171875, -0.22265625, -0.214111328125, -0.20556640625, -0.197021484375, -0.1884765625, -0.179931640625, -0.17138671875, -0.162841796875, -0.154296875, -0.145751953125, -0.13720703125, -0.128662109375, -0.1201171875, -0.111572265625, -0.10302734375, -0.094482421875, -0.0859375, -0.077392578125, -0.06884765625, -0.060302734375, -0.0517578125, -0.043212890625, -0.03466796875, -0.026123046875, -0.017578125, -0.009033203125, -0.00048828125, 0.008056640625, 0.0166015625, 0.025146484375, 0.03369140625, 0.042236328125, 0.05078125, 0.059326171875, 0.06787109375, 0.076416015625, 0.0849609375, 0.093505859375, 0.10205078125, 0.110595703125, 0.119140625, 0.127685546875, 0.13623046875, 0.144775390625, 0.1533203125, 0.161865234375, 0.17041015625, 0.178955078125, 0.1875, 0.196044921875, 0.20458984375, 0.213134765625, 0.2216796875, 0.230224609375, 0.23876953125, 0.247314453125, 0.255859375]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.29._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 4.0, 5.0, 4.0, 6.0, 8.0, 16.0, 23.0, 34.0, 41.0, 77.0, 77.0, 157.0, 222.0, 400.0, 996.0, 2354.0, 6023.0, 15577.0, 40300.0, 98757.0, 209001.0, 428759.0, 694673.0, 885494.0, 789536.0, 521626.0, 277074.0, 130974.0, 55250.0, 22016.0, 8693.0, 3424.0, 1289.0, 608.0, 287.0, 139.0, 108.0, 61.0, 66.0, 33.0, 28.0, 27.0, 12.0, 14.0, 9.0, 8.0, 2.0, 3.0, 2.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.404296875, -0.39263916015625, -0.3809814453125, -0.36932373046875, -0.357666015625, -0.34600830078125, -0.3343505859375, -0.32269287109375, -0.31103515625, -0.29937744140625, -0.2877197265625, -0.27606201171875, -0.264404296875, -0.25274658203125, -0.2410888671875, -0.22943115234375, -0.2177734375, -0.20611572265625, -0.1944580078125, -0.18280029296875, -0.171142578125, -0.15948486328125, -0.1478271484375, -0.13616943359375, -0.12451171875, -0.11285400390625, -0.1011962890625, -0.08953857421875, -0.077880859375, -0.06622314453125, -0.0545654296875, -0.04290771484375, -0.03125, -0.01959228515625, -0.0079345703125, 0.00372314453125, 0.015380859375, 0.02703857421875, 0.0386962890625, 0.05035400390625, 0.06201171875, 0.07366943359375, 0.0853271484375, 0.09698486328125, 0.108642578125, 0.12030029296875, 0.1319580078125, 0.14361572265625, 0.1552734375, 0.16693115234375, 0.1785888671875, 0.19024658203125, 0.201904296875, 0.21356201171875, 0.2252197265625, 0.23687744140625, 0.24853515625, 0.26019287109375, 0.2718505859375, 0.28350830078125, 0.295166015625, 0.30682373046875, 0.3184814453125, 0.33013916015625, 0.341796875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.29._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 8.0, 16.0, 20.0, 30.0, 59.0, 106.0, 188.0, 264.0, 467.0, 671.0, 1187.0, 1630.0, 2407.0, 3776.0, 5074.0, 7085.0, 10054.0, 13018.0, 16881.0, 20944.0, 25652.0, 30969.0, 36771.0, 41031.0, 46096.0, 50602.0, 54737.0, 56339.0, 56749.0, 53075.0, 49885.0, 44680.0, 40615.0, 35604.0, 29677.0, 24582.0, 20087.0, 16108.0, 12150.0, 9208.0, 6677.0, 4615.0, 3485.0, 2143.0, 1489.0, 1071.0, 638.0, 424.0, 243.0, 151.0, 107.0, 55.0, 31.0, 22.0, 9.0, 8.0, 1.0, 1.0, 1.0, 3.0], "bins": [-0.07373046875, -0.07141876220703125, -0.0691070556640625, -0.06679534912109375, -0.064483642578125, -0.06217193603515625, -0.0598602294921875, -0.05754852294921875, -0.05523681640625, -0.05292510986328125, -0.0506134033203125, -0.04830169677734375, -0.045989990234375, -0.04367828369140625, -0.0413665771484375, -0.03905487060546875, -0.0367431640625, -0.03443145751953125, -0.0321197509765625, -0.02980804443359375, -0.027496337890625, -0.02518463134765625, -0.0228729248046875, -0.02056121826171875, -0.01824951171875, -0.01593780517578125, -0.0136260986328125, -0.01131439208984375, -0.009002685546875, -0.00669097900390625, -0.0043792724609375, -0.00206756591796875, 0.000244140625, 0.00255584716796875, 0.0048675537109375, 0.00717926025390625, 0.009490966796875, 0.01180267333984375, 0.0141143798828125, 0.01642608642578125, 0.01873779296875, 0.02104949951171875, 0.0233612060546875, 0.02567291259765625, 0.027984619140625, 0.03029632568359375, 0.0326080322265625, 0.03491973876953125, 0.0372314453125, 0.03954315185546875, 0.0418548583984375, 0.04416656494140625, 0.046478271484375, 0.04878997802734375, 0.0511016845703125, 0.05341339111328125, 0.05572509765625, 0.05803680419921875, 0.0603485107421875, 0.06266021728515625, 0.064971923828125, 0.06728363037109375, 0.0695953369140625, 0.07190704345703125, 0.07421875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.30._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 2.0, 3.0, 2.0, 6.0, 8.0, 22.0, 25.0, 35.0, 46.0, 66.0, 108.0, 168.0, 263.0, 481.0, 778.0, 1379.0, 2612.0, 5756.0, 14643.0, 37923.0, 93887.0, 237117.0, 573647.0, 1199919.0, 2247865.0, 3683371.0, 3829962.0, 2447954.0, 1303924.0, 645519.0, 266525.0, 111239.0, 42343.0, 16291.0, 6778.0, 2869.0, 1527.0, 848.0, 446.0, 314.0, 212.0, 123.0, 81.0, 33.0, 34.0, 19.0, 13.0, 7.0, 8.0, 4.0, 3.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.22265625, -0.214996337890625, -0.20733642578125, -0.199676513671875, -0.1920166015625, -0.184356689453125, -0.17669677734375, -0.169036865234375, -0.161376953125, -0.153717041015625, -0.14605712890625, -0.138397216796875, -0.1307373046875, -0.123077392578125, -0.11541748046875, -0.107757568359375, -0.10009765625, -0.092437744140625, -0.08477783203125, -0.077117919921875, -0.0694580078125, -0.061798095703125, -0.05413818359375, -0.046478271484375, -0.038818359375, -0.031158447265625, -0.02349853515625, -0.015838623046875, -0.0081787109375, -0.000518798828125, 0.00714111328125, 0.014801025390625, 0.0224609375, 0.030120849609375, 0.03778076171875, 0.045440673828125, 0.0531005859375, 0.060760498046875, 0.06842041015625, 0.076080322265625, 0.083740234375, 0.091400146484375, 0.09906005859375, 0.106719970703125, 0.1143798828125, 0.122039794921875, 0.12969970703125, 0.137359619140625, 0.14501953125, 0.152679443359375, 0.16033935546875, 0.167999267578125, 0.1756591796875, 0.183319091796875, 0.19097900390625, 0.198638916015625, 0.206298828125, 0.213958740234375, 0.22161865234375, 0.229278564453125, 0.2369384765625, 0.244598388671875, 0.25225830078125, 0.259918212890625, 0.267578125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.30._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [6.0, 2.0, 9.0, 4.0, 5.0, 9.0, 24.0, 13.0, 33.0, 40.0, 63.0, 79.0, 90.0, 122.0, 146.0, 244.0, 270.0, 405.0, 619.0, 1013.0, 1809.0, 3609.0, 8195.0, 19555.0, 50472.0, 119564.0, 266787.0, 522604.0, 833336.0, 933014.0, 702455.0, 398262.0, 188815.0, 83227.0, 33510.0, 13338.0, 5604.0, 2673.0, 1434.0, 836.0, 494.0, 420.0, 274.0, 199.0, 152.0, 140.0, 84.0, 68.0, 38.0, 38.0, 30.0, 23.0, 16.0, 16.0, 3.0, 1.0, 4.0, 1.0, 1.0, 1.0, 3.0, 1.0, 1.0, 1.0], "bins": [-0.3046875, -0.294281005859375, -0.28387451171875, -0.273468017578125, -0.2630615234375, -0.252655029296875, -0.24224853515625, -0.231842041015625, -0.221435546875, -0.211029052734375, -0.20062255859375, -0.190216064453125, -0.1798095703125, -0.169403076171875, -0.15899658203125, -0.148590087890625, -0.13818359375, -0.127777099609375, -0.11737060546875, -0.106964111328125, -0.0965576171875, -0.086151123046875, -0.07574462890625, -0.065338134765625, -0.054931640625, -0.044525146484375, -0.03411865234375, -0.023712158203125, -0.0133056640625, -0.002899169921875, 0.00750732421875, 0.017913818359375, 0.0283203125, 0.038726806640625, 0.04913330078125, 0.059539794921875, 0.0699462890625, 0.080352783203125, 0.09075927734375, 0.101165771484375, 0.111572265625, 0.121978759765625, 0.13238525390625, 0.142791748046875, 0.1531982421875, 0.163604736328125, 0.17401123046875, 0.184417724609375, 0.19482421875, 0.205230712890625, 0.21563720703125, 0.226043701171875, 0.2364501953125, 0.246856689453125, 0.25726318359375, 0.267669677734375, 0.278076171875, 0.288482666015625, 0.29888916015625, 0.309295654296875, 0.3197021484375, 0.330108642578125, 0.34051513671875, 0.350921630859375, 0.361328125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.30._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 1.0, 3.0, 3.0, 14.0, 31.0, 37.0, 83.0, 127.0, 202.0, 288.0, 511.0, 774.0, 1187.0, 1748.0, 2413.0, 3681.0, 5092.0, 6814.0, 9305.0, 11969.0, 15086.0, 19818.0, 23138.0, 28934.0, 33554.0, 39332.0, 44146.0, 49444.0, 54314.0, 59276.0, 60172.0, 55751.0, 50639.0, 45238.0, 40505.0, 36226.0, 28940.0, 25699.0, 20297.0, 16973.0, 13373.0, 9189.0, 7467.0, 5559.0, 3954.0, 2858.0, 1900.0, 1332.0, 911.0, 543.0, 320.0, 219.0, 104.0, 70.0, 48.0, 25.0, 21.0, 4.0, 6.0, 3.0, 4.0, 0.0, 2.0], "bins": [-0.08984375, -0.0869598388671875, -0.084075927734375, -0.0811920166015625, -0.07830810546875, -0.0754241943359375, -0.072540283203125, -0.0696563720703125, -0.0667724609375, -0.0638885498046875, -0.061004638671875, -0.0581207275390625, -0.05523681640625, -0.0523529052734375, -0.049468994140625, -0.0465850830078125, -0.043701171875, -0.0408172607421875, -0.037933349609375, -0.0350494384765625, -0.03216552734375, -0.0292816162109375, -0.026397705078125, -0.0235137939453125, -0.0206298828125, -0.0177459716796875, -0.014862060546875, -0.0119781494140625, -0.00909423828125, -0.0062103271484375, -0.003326416015625, -0.0004425048828125, 0.00244140625, 0.0053253173828125, 0.008209228515625, 0.0110931396484375, 0.01397705078125, 0.0168609619140625, 0.019744873046875, 0.0226287841796875, 0.0255126953125, 0.0283966064453125, 0.031280517578125, 0.0341644287109375, 0.03704833984375, 0.0399322509765625, 0.042816162109375, 0.0457000732421875, 0.048583984375, 0.0514678955078125, 0.054351806640625, 0.0572357177734375, 0.06011962890625, 0.0630035400390625, 0.065887451171875, 0.0687713623046875, 0.0716552734375, 0.0745391845703125, 0.077423095703125, 0.0803070068359375, 0.08319091796875, 0.0860748291015625, 0.088958740234375, 0.0918426513671875, 0.0947265625]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.31._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 8.0, 4.0, 5.0, 8.0, 11.0, 13.0, 21.0, 30.0, 44.0, 50.0, 71.0, 117.0, 181.0, 305.0, 612.0, 1162.0, 3007.0, 8618.0, 34345.0, 140484.0, 546428.0, 1808674.0, 4347468.0, 5462189.0, 2974661.0, 1060631.0, 288481.0, 73182.0, 17572.0, 5035.0, 1843.0, 803.0, 469.0, 240.0, 151.0, 79.0, 66.0, 45.0, 32.0, 17.0, 11.0, 6.0, 6.0, 3.0, 4.0, 2.0, 1.0, 2.0, 2.0, 2.0, 2.0, 1.0, 2.0, 1.0], "bins": [-0.4140625, -0.401611328125, -0.38916015625, -0.376708984375, -0.3642578125, -0.351806640625, -0.33935546875, -0.326904296875, -0.314453125, -0.302001953125, -0.28955078125, -0.277099609375, -0.2646484375, -0.252197265625, -0.23974609375, -0.227294921875, -0.21484375, -0.202392578125, -0.18994140625, -0.177490234375, -0.1650390625, -0.152587890625, -0.14013671875, -0.127685546875, -0.115234375, -0.102783203125, -0.09033203125, -0.077880859375, -0.0654296875, -0.052978515625, -0.04052734375, -0.028076171875, -0.015625, -0.003173828125, 0.00927734375, 0.021728515625, 0.0341796875, 0.046630859375, 0.05908203125, 0.071533203125, 0.083984375, 0.096435546875, 0.10888671875, 0.121337890625, 0.1337890625, 0.146240234375, 0.15869140625, 0.171142578125, 0.18359375, 0.196044921875, 0.20849609375, 0.220947265625, 0.2333984375, 0.245849609375, 0.25830078125, 0.270751953125, 0.283203125, 0.295654296875, 0.30810546875, 0.320556640625, 0.3330078125, 0.345458984375, 0.35791015625, 0.370361328125, 0.3828125]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.31._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 4.0, 3.0, 4.0, 9.0, 5.0, 10.0, 18.0, 28.0, 31.0, 38.0, 68.0, 103.0, 110.0, 191.0, 300.0, 510.0, 1107.0, 2489.0, 8208.0, 28691.0, 100774.0, 279654.0, 658670.0, 1070752.0, 1042040.0, 613691.0, 260027.0, 89749.0, 25291.0, 7078.0, 2397.0, 941.0, 471.0, 247.0, 201.0, 107.0, 78.0, 56.0, 44.0, 26.0, 22.0, 11.0, 13.0, 6.0, 4.0, 7.0, 1.0, 6.0, 2.0, 1.0, 2.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.373046875, -0.358673095703125, -0.34429931640625, -0.329925537109375, -0.3155517578125, -0.301177978515625, -0.28680419921875, -0.272430419921875, -0.258056640625, -0.243682861328125, -0.22930908203125, -0.214935302734375, -0.2005615234375, -0.186187744140625, -0.17181396484375, -0.157440185546875, -0.14306640625, -0.128692626953125, -0.11431884765625, -0.099945068359375, -0.0855712890625, -0.071197509765625, -0.05682373046875, -0.042449951171875, -0.028076171875, -0.013702392578125, 0.00067138671875, 0.015045166015625, 0.0294189453125, 0.043792724609375, 0.05816650390625, 0.072540283203125, 0.0869140625, 0.101287841796875, 0.11566162109375, 0.130035400390625, 0.1444091796875, 0.158782958984375, 0.17315673828125, 0.187530517578125, 0.201904296875, 0.216278076171875, 0.23065185546875, 0.245025634765625, 0.2593994140625, 0.273773193359375, 0.28814697265625, 0.302520751953125, 0.31689453125, 0.331268310546875, 0.34564208984375, 0.360015869140625, 0.3743896484375, 0.388763427734375, 0.40313720703125, 0.417510986328125, 0.431884765625, 0.446258544921875, 0.46063232421875, 0.475006103515625, 0.4893798828125, 0.503753662109375, 0.51812744140625, 0.532501220703125, 0.546875]}, "parameters/_orig_mod._fsdp_wrapped_module.model.layers.31._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 10.0, 13.0, 22.0, 48.0, 69.0, 96.0, 219.0, 350.0, 658.0, 1118.0, 2028.0, 2932.0, 4548.0, 6930.0, 10565.0, 13891.0, 20579.0, 26093.0, 34123.0, 40971.0, 50050.0, 57448.0, 64471.0, 69289.0, 70052.0, 66254.0, 60027.0, 53426.0, 45249.0, 35936.0, 29372.0, 22109.0, 16744.0, 11150.0, 8428.0, 5312.0, 3520.0, 2288.0, 1385.0, 798.0, 473.0, 245.0, 173.0, 99.0, 41.0, 25.0, 24.0, 8.0, 7.0, 3.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.10498046875, -0.10150146484375, -0.0980224609375, -0.09454345703125, -0.091064453125, -0.08758544921875, -0.0841064453125, -0.08062744140625, -0.0771484375, -0.07366943359375, -0.0701904296875, -0.06671142578125, -0.063232421875, -0.05975341796875, -0.0562744140625, -0.05279541015625, -0.04931640625, -0.04583740234375, -0.0423583984375, -0.03887939453125, -0.035400390625, -0.03192138671875, -0.0284423828125, -0.02496337890625, -0.021484375, -0.01800537109375, -0.0145263671875, -0.01104736328125, -0.007568359375, -0.00408935546875, -0.0006103515625, 0.00286865234375, 0.00634765625, 0.00982666015625, 0.0133056640625, 0.01678466796875, 0.020263671875, 0.02374267578125, 0.0272216796875, 0.03070068359375, 0.0341796875, 0.03765869140625, 0.0411376953125, 0.04461669921875, 0.048095703125, 0.05157470703125, 0.0550537109375, 0.05853271484375, 0.06201171875, 0.06549072265625, 0.0689697265625, 0.07244873046875, 0.075927734375, 0.07940673828125, 0.0828857421875, 0.08636474609375, 0.08984375, 0.09332275390625, 0.0968017578125, 0.10028076171875, 0.103759765625, 0.10723876953125, 0.1107177734375, 0.11419677734375, 0.11767578125]}, "parameters/_orig_mod.model.embed_tokens.weight": {"_type": "histogram", "values": [2.0, 1.0, 6.0, 8.0, 9.0, 14.0, 25.0, 47.0, 88.0, 314.0, 1316.0, 5746.0, 20883.0, 84104.0, 277676.0, 782856.0, 2006050.0, 4315809.0, 8093267.0, 12519296.0, 16666752.0, 16777216.0, 16177924.0, 11827574.0, 7497016.0, 3872541.0, 1775839.0, 681293.0, 236928.0, 72507.0, 20733.0, 6064.0, 2135.0, 891.0, 401.0, 206.0, 82.0, 64.0, 42.0, 30.0, 13.0, 9.0, 8.0, 4.0, 5.0, 3.0, 3.0, 4.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.08203125, -0.0782012939453125, -0.074371337890625, -0.0705413818359375, -0.06671142578125, -0.0628814697265625, -0.059051513671875, -0.0552215576171875, -0.0513916015625, -0.0475616455078125, -0.043731689453125, -0.0399017333984375, -0.03607177734375, -0.0322418212890625, -0.028411865234375, -0.0245819091796875, -0.020751953125, -0.0169219970703125, -0.013092041015625, -0.0092620849609375, -0.00543212890625, -0.0016021728515625, 0.002227783203125, 0.0060577392578125, 0.0098876953125, 0.0137176513671875, 0.017547607421875, 0.0213775634765625, 0.02520751953125, 0.0290374755859375, 0.032867431640625, 0.0366973876953125, 0.04052734375, 0.0443572998046875, 0.048187255859375, 0.0520172119140625, 0.05584716796875, 0.0596771240234375, 0.063507080078125, 0.0673370361328125, 0.0711669921875, 0.0749969482421875, 0.078826904296875, 0.0826568603515625, 0.08648681640625, 0.0903167724609375, 0.094146728515625, 0.0979766845703125, 0.101806640625, 0.1056365966796875, 0.109466552734375, 0.1132965087890625, 0.11712646484375, 0.1209564208984375, 0.124786376953125, 0.1286163330078125, 0.1324462890625, 0.1362762451171875, 0.140106201171875, 0.1439361572265625, 0.14776611328125, 0.1515960693359375, 0.155426025390625, 0.1592559814453125, 0.1630859375]}, "parameters/_orig_mod.model.layers.0._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 2.0, 1.0, 0.0, 3.0, 5.0, 3.0, 3.0, 13.0, 13.0, 19.0, 25.0, 33.0, 34.0, 58.0, 88.0, 141.0, 276.0, 464.0, 868.0, 1442.0, 2926.0, 6519.0, 16346.0, 49035.0, 188027.0, 897749.0, 9269153.0, 5273066.0, 820347.0, 178426.0, 44064.0, 15983.0, 5973.0, 2729.0, 1481.0, 797.0, 424.0, 220.0, 148.0, 92.0, 71.0, 50.0, 24.0, 18.0, 14.0, 14.0, 7.0, 2.0, 3.0, 5.0, 2.0, 0.0, 2.0, 1.0, 1.0, 0.0, 2.0], "bins": [-0.7734375, -0.75067138671875, -0.7279052734375, -0.70513916015625, -0.682373046875, -0.65960693359375, -0.6368408203125, -0.61407470703125, -0.59130859375, -0.56854248046875, -0.5457763671875, -0.52301025390625, -0.500244140625, -0.47747802734375, -0.4547119140625, -0.43194580078125, -0.4091796875, -0.38641357421875, -0.3636474609375, -0.34088134765625, -0.318115234375, -0.29534912109375, -0.2725830078125, -0.24981689453125, -0.22705078125, -0.20428466796875, -0.1815185546875, -0.15875244140625, -0.135986328125, -0.11322021484375, -0.0904541015625, -0.06768798828125, -0.044921875, -0.02215576171875, 0.0006103515625, 0.02337646484375, 0.046142578125, 0.06890869140625, 0.0916748046875, 0.11444091796875, 0.13720703125, 0.15997314453125, 0.1827392578125, 0.20550537109375, 0.228271484375, 0.25103759765625, 0.2738037109375, 0.29656982421875, 0.3193359375, 0.34210205078125, 0.3648681640625, 0.38763427734375, 0.410400390625, 0.43316650390625, 0.4559326171875, 0.47869873046875, 0.50146484375, 0.52423095703125, 0.5469970703125, 0.56976318359375, 0.592529296875, 0.61529541015625, 0.6380615234375, 0.66082763671875, 0.68359375]}, "parameters/_orig_mod.model.layers.0._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [3.0, 5.0, 1.0, 2.0, 1.0, 5.0, 3.0, 4.0, 9.0, 21.0, 22.0, 18.0, 76.0, 69.0, 93.0, 134.0, 277.0, 435.0, 736.0, 1234.0, 2154.0, 3716.0, 7577.0, 16016.0, 38378.0, 110340.0, 344922.0, 2197721.0, 1028765.0, 286577.0, 92032.0, 33250.0, 14808.0, 6744.0, 3376.0, 1898.0, 1090.0, 706.0, 392.0, 246.0, 138.0, 104.0, 61.0, 50.0, 32.0, 20.0, 11.0, 12.0, 5.0, 0.0, 2.0, 3.0, 1.0, 1.0, 4.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.6015625, -0.58001708984375, -0.5584716796875, -0.53692626953125, -0.515380859375, -0.49383544921875, -0.4722900390625, -0.45074462890625, -0.42919921875, -0.40765380859375, -0.3861083984375, -0.36456298828125, -0.343017578125, -0.32147216796875, -0.2999267578125, -0.27838134765625, -0.2568359375, -0.23529052734375, -0.2137451171875, -0.19219970703125, -0.170654296875, -0.14910888671875, -0.1275634765625, -0.10601806640625, -0.08447265625, -0.06292724609375, -0.0413818359375, -0.01983642578125, 0.001708984375, 0.02325439453125, 0.0447998046875, 0.06634521484375, 0.087890625, 0.10943603515625, 0.1309814453125, 0.15252685546875, 0.174072265625, 0.19561767578125, 0.2171630859375, 0.23870849609375, 0.26025390625, 0.28179931640625, 0.3033447265625, 0.32489013671875, 0.346435546875, 0.36798095703125, 0.3895263671875, 0.41107177734375, 0.4326171875, 0.45416259765625, 0.4757080078125, 0.49725341796875, 0.518798828125, 0.54034423828125, 0.5618896484375, 0.58343505859375, 0.60498046875, 0.62652587890625, 0.6480712890625, 0.66961669921875, 0.691162109375, 0.71270751953125, 0.7342529296875, 0.75579833984375, 0.77734375]}, "parameters/_orig_mod.model.layers.0._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 1.0, 2.0, 3.0, 4.0, 5.0, 10.0, 12.0, 25.0, 42.0, 55.0, 79.0, 132.0, 210.0, 306.0, 474.0, 752.0, 1033.0, 1458.0, 2060.0, 3383.0, 4528.0, 6758.0, 8849.0, 12823.0, 17139.0, 27259.0, 33202.0, 50723.0, 61432.0, 92277.0, 116158.0, 109477.0, 81027.0, 57629.0, 41793.0, 29762.0, 24658.0, 15743.0, 11023.0, 8528.0, 5842.0, 4110.0, 2782.0, 2063.0, 1315.0, 912.0, 620.0, 448.0, 257.0, 187.0, 111.0, 85.0, 36.0, 31.0, 17.0, 15.0, 7.0, 0.0, 2.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.043701171875, -0.042327880859375, -0.04095458984375, -0.039581298828125, -0.0382080078125, -0.036834716796875, -0.03546142578125, -0.034088134765625, -0.03271484375, -0.031341552734375, -0.02996826171875, -0.028594970703125, -0.0272216796875, -0.025848388671875, -0.02447509765625, -0.023101806640625, -0.021728515625, -0.020355224609375, -0.01898193359375, -0.017608642578125, -0.0162353515625, -0.014862060546875, -0.01348876953125, -0.012115478515625, -0.0107421875, -0.009368896484375, -0.00799560546875, -0.006622314453125, -0.0052490234375, -0.003875732421875, -0.00250244140625, -0.001129150390625, 0.000244140625, 0.001617431640625, 0.00299072265625, 0.004364013671875, 0.0057373046875, 0.007110595703125, 0.00848388671875, 0.009857177734375, 0.01123046875, 0.012603759765625, 0.01397705078125, 0.015350341796875, 0.0167236328125, 0.018096923828125, 0.01947021484375, 0.020843505859375, 0.022216796875, 0.023590087890625, 0.02496337890625, 0.026336669921875, 0.0277099609375, 0.029083251953125, 0.03045654296875, 0.031829833984375, 0.033203125, 0.034576416015625, 0.03594970703125, 0.037322998046875, 0.0386962890625, 0.040069580078125, 0.04144287109375, 0.042816162109375, 0.044189453125]}, "parameters/_orig_mod.model.layers.1._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 2.0, 1.0, 1.0, 4.0, 6.0, 14.0, 19.0, 30.0, 51.0, 104.0, 186.0, 285.0, 686.0, 1703.0, 4899.0, 16039.0, 55044.0, 201305.0, 660293.0, 2166509.0, 6659395.0, 4862364.0, 1500672.0, 458901.0, 134978.0, 37016.0, 11041.0, 3301.0, 1282.0, 489.0, 261.0, 134.0, 84.0, 44.0, 28.0, 10.0, 10.0, 3.0, 4.0, 3.0, 2.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.5390625, -0.523590087890625, -0.50811767578125, -0.492645263671875, -0.4771728515625, -0.461700439453125, -0.44622802734375, -0.430755615234375, -0.415283203125, -0.399810791015625, -0.38433837890625, -0.368865966796875, -0.3533935546875, -0.337921142578125, -0.32244873046875, -0.306976318359375, -0.29150390625, -0.276031494140625, -0.26055908203125, -0.245086669921875, -0.2296142578125, -0.214141845703125, -0.19866943359375, -0.183197021484375, -0.167724609375, -0.152252197265625, -0.13677978515625, -0.121307373046875, -0.1058349609375, -0.090362548828125, -0.07489013671875, -0.059417724609375, -0.0439453125, -0.028472900390625, -0.01300048828125, 0.002471923828125, 0.0179443359375, 0.033416748046875, 0.04888916015625, 0.064361572265625, 0.079833984375, 0.095306396484375, 0.11077880859375, 0.126251220703125, 0.1417236328125, 0.157196044921875, 0.17266845703125, 0.188140869140625, 0.20361328125, 0.219085693359375, 0.23455810546875, 0.250030517578125, 0.2655029296875, 0.280975341796875, 0.29644775390625, 0.311920166015625, 0.327392578125, 0.342864990234375, 0.35833740234375, 0.373809814453125, 0.3892822265625, 0.404754638671875, 0.42022705078125, 0.435699462890625, 0.451171875]}, "parameters/_orig_mod.model.layers.1._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 3.0, 3.0, 6.0, 2.0, 4.0, 12.0, 13.0, 15.0, 42.0, 51.0, 82.0, 131.0, 172.0, 285.0, 437.0, 717.0, 979.0, 1743.0, 2961.0, 4933.0, 8410.0, 15376.0, 24949.0, 48232.0, 86473.0, 147258.0, 262479.0, 439819.0, 827400.0, 1033131.0, 541201.0, 320264.0, 182165.0, 106352.0, 61465.0, 32835.0, 18084.0, 10604.0, 6004.0, 3614.0, 2005.0, 1260.0, 826.0, 523.0, 311.0, 238.0, 148.0, 90.0, 59.0, 44.0, 32.0, 22.0, 19.0, 6.0, 3.0, 4.0, 4.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.337890625, -0.327392578125, -0.31689453125, -0.306396484375, -0.2958984375, -0.285400390625, -0.27490234375, -0.264404296875, -0.25390625, -0.243408203125, -0.23291015625, -0.222412109375, -0.2119140625, -0.201416015625, -0.19091796875, -0.180419921875, -0.169921875, -0.159423828125, -0.14892578125, -0.138427734375, -0.1279296875, -0.117431640625, -0.10693359375, -0.096435546875, -0.0859375, -0.075439453125, -0.06494140625, -0.054443359375, -0.0439453125, -0.033447265625, -0.02294921875, -0.012451171875, -0.001953125, 0.008544921875, 0.01904296875, 0.029541015625, 0.0400390625, 0.050537109375, 0.06103515625, 0.071533203125, 0.08203125, 0.092529296875, 0.10302734375, 0.113525390625, 0.1240234375, 0.134521484375, 0.14501953125, 0.155517578125, 0.166015625, 0.176513671875, 0.18701171875, 0.197509765625, 0.2080078125, 0.218505859375, 0.22900390625, 0.239501953125, 0.25, 0.260498046875, 0.27099609375, 0.281494140625, 0.2919921875, 0.302490234375, 0.31298828125, 0.323486328125, 0.333984375]}, "parameters/_orig_mod.model.layers.1._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [4.0, 1.0, 0.0, 5.0, 11.0, 11.0, 23.0, 63.0, 60.0, 135.0, 209.0, 369.0, 528.0, 840.0, 1230.0, 1982.0, 2855.0, 3957.0, 5639.0, 7237.0, 10132.0, 13733.0, 16751.0, 21544.0, 26331.0, 34866.0, 37426.0, 42834.0, 48872.0, 51939.0, 57143.0, 60334.0, 56299.0, 53678.0, 50349.0, 43718.0, 37177.0, 36659.0, 25746.0, 21775.0, 17577.0, 13318.0, 10439.0, 8047.0, 5066.0, 4224.0, 2852.0, 1902.0, 1364.0, 946.0, 589.0, 348.0, 227.0, 141.0, 76.0, 38.0, 30.0, 9.0, 9.0, 5.0, 3.0, 2.0, 1.0, 2.0], "bins": [-0.041748046875, -0.040424346923828125, -0.03910064697265625, -0.037776947021484375, -0.0364532470703125, -0.035129547119140625, -0.03380584716796875, -0.032482147216796875, -0.031158447265625, -0.029834747314453125, -0.02851104736328125, -0.027187347412109375, -0.0258636474609375, -0.024539947509765625, -0.02321624755859375, -0.021892547607421875, -0.02056884765625, -0.019245147705078125, -0.01792144775390625, -0.016597747802734375, -0.0152740478515625, -0.013950347900390625, -0.01262664794921875, -0.011302947998046875, -0.009979248046875, -0.008655548095703125, -0.00733184814453125, -0.006008148193359375, -0.0046844482421875, -0.003360748291015625, -0.00203704833984375, -0.000713348388671875, 0.0006103515625, 0.001934051513671875, 0.00325775146484375, 0.004581451416015625, 0.0059051513671875, 0.007228851318359375, 0.00855255126953125, 0.009876251220703125, 0.011199951171875, 0.012523651123046875, 0.01384735107421875, 0.015171051025390625, 0.0164947509765625, 0.017818450927734375, 0.01914215087890625, 0.020465850830078125, 0.02178955078125, 0.023113250732421875, 0.02443695068359375, 0.025760650634765625, 0.0270843505859375, 0.028408050537109375, 0.02973175048828125, 0.031055450439453125, 0.032379150390625, 0.033702850341796875, 0.03502655029296875, 0.036350250244140625, 0.0376739501953125, 0.038997650146484375, 0.04032135009765625, 0.041645050048828125, 0.04296875]}, "parameters/_orig_mod.model.layers.2._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [4.0, 0.0, 3.0, 0.0, 3.0, 5.0, 15.0, 14.0, 31.0, 24.0, 51.0, 60.0, 78.0, 136.0, 147.0, 205.0, 374.0, 479.0, 781.0, 1283.0, 2573.0, 6083.0, 12280.0, 27728.0, 66217.0, 131030.0, 265158.0, 490235.0, 930083.0, 1622362.0, 2731193.0, 3481146.0, 2976358.0, 1826830.0, 1040983.0, 576085.0, 292895.0, 156192.0, 74592.0, 34179.0, 15059.0, 6822.0, 3285.0, 1547.0, 909.0, 532.0, 359.0, 230.0, 159.0, 92.0, 112.0, 56.0, 54.0, 33.0, 24.0, 17.0, 8.0, 8.0, 8.0, 0.0, 2.0, 0.0, 2.0, 3.0], "bins": [-0.234375, -0.226959228515625, -0.21954345703125, -0.212127685546875, -0.2047119140625, -0.197296142578125, -0.18988037109375, -0.182464599609375, -0.175048828125, -0.167633056640625, -0.16021728515625, -0.152801513671875, -0.1453857421875, -0.137969970703125, -0.13055419921875, -0.123138427734375, -0.11572265625, -0.108306884765625, -0.10089111328125, -0.093475341796875, -0.0860595703125, -0.078643798828125, -0.07122802734375, -0.063812255859375, -0.056396484375, -0.048980712890625, -0.04156494140625, -0.034149169921875, -0.0267333984375, -0.019317626953125, -0.01190185546875, -0.004486083984375, 0.0029296875, 0.010345458984375, 0.01776123046875, 0.025177001953125, 0.0325927734375, 0.040008544921875, 0.04742431640625, 0.054840087890625, 0.062255859375, 0.069671630859375, 0.07708740234375, 0.084503173828125, 0.0919189453125, 0.099334716796875, 0.10675048828125, 0.114166259765625, 0.12158203125, 0.128997802734375, 0.13641357421875, 0.143829345703125, 0.1512451171875, 0.158660888671875, 0.16607666015625, 0.173492431640625, 0.180908203125, 0.188323974609375, 0.19573974609375, 0.203155517578125, 0.2105712890625, 0.217987060546875, 0.22540283203125, 0.232818603515625, 0.240234375]}, "parameters/_orig_mod.model.layers.2._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 2.0, 1.0, 4.0, 7.0, 12.0, 17.0, 18.0, 36.0, 40.0, 52.0, 52.0, 79.0, 145.0, 236.0, 351.0, 624.0, 1177.0, 2383.0, 4216.0, 8753.0, 15093.0, 27493.0, 49545.0, 86583.0, 141129.0, 236840.0, 401560.0, 662853.0, 844772.0, 688785.0, 419944.0, 250450.0, 146020.0, 89562.0, 51306.0, 29159.0, 15819.0, 9147.0, 4710.0, 2279.0, 1326.0, 583.0, 401.0, 224.0, 159.0, 105.0, 69.0, 63.0, 27.0, 33.0, 18.0, 17.0, 6.0, 5.0, 5.0, 2.0, 3.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.306640625, -0.296600341796875, -0.28656005859375, -0.276519775390625, -0.2664794921875, -0.256439208984375, -0.24639892578125, -0.236358642578125, -0.226318359375, -0.216278076171875, -0.20623779296875, -0.196197509765625, -0.1861572265625, -0.176116943359375, -0.16607666015625, -0.156036376953125, -0.14599609375, -0.135955810546875, -0.12591552734375, -0.115875244140625, -0.1058349609375, -0.095794677734375, -0.08575439453125, -0.075714111328125, -0.065673828125, -0.055633544921875, -0.04559326171875, -0.035552978515625, -0.0255126953125, -0.015472412109375, -0.00543212890625, 0.004608154296875, 0.0146484375, 0.024688720703125, 0.03472900390625, 0.044769287109375, 0.0548095703125, 0.064849853515625, 0.07489013671875, 0.084930419921875, 0.094970703125, 0.105010986328125, 0.11505126953125, 0.125091552734375, 0.1351318359375, 0.145172119140625, 0.15521240234375, 0.165252685546875, 0.17529296875, 0.185333251953125, 0.19537353515625, 0.205413818359375, 0.2154541015625, 0.225494384765625, 0.23553466796875, 0.245574951171875, 0.255615234375, 0.265655517578125, 0.27569580078125, 0.285736083984375, 0.2957763671875, 0.305816650390625, 0.31585693359375, 0.325897216796875, 0.3359375]}, "parameters/_orig_mod.model.layers.2._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 2.0, 7.0, 10.0, 20.0, 34.0, 59.0, 92.0, 166.0, 256.0, 512.0, 771.0, 1056.0, 1788.0, 2350.0, 3972.0, 4981.0, 7588.0, 10052.0, 13629.0, 17534.0, 21853.0, 26961.0, 36814.0, 38848.0, 45626.0, 51020.0, 54224.0, 56802.0, 58154.0, 56258.0, 53033.0, 50423.0, 44190.0, 37476.0, 36059.0, 25818.0, 20872.0, 16596.0, 12785.0, 9376.0, 6817.0, 5006.0, 3520.0, 2253.0, 1527.0, 880.0, 666.0, 381.0, 253.0, 125.0, 75.0, 43.0, 29.0, 11.0, 12.0, 4.0, 4.0, 3.0], "bins": [-0.044189453125, -0.042903900146484375, -0.04161834716796875, -0.040332794189453125, -0.0390472412109375, -0.037761688232421875, -0.03647613525390625, -0.035190582275390625, -0.033905029296875, -0.032619476318359375, -0.03133392333984375, -0.030048370361328125, -0.0287628173828125, -0.027477264404296875, -0.02619171142578125, -0.024906158447265625, -0.02362060546875, -0.022335052490234375, -0.02104949951171875, -0.019763946533203125, -0.0184783935546875, -0.017192840576171875, -0.01590728759765625, -0.014621734619140625, -0.013336181640625, -0.012050628662109375, -0.01076507568359375, -0.009479522705078125, -0.0081939697265625, -0.006908416748046875, -0.00562286376953125, -0.004337310791015625, -0.0030517578125, -0.001766204833984375, -0.00048065185546875, 0.000804901123046875, 0.0020904541015625, 0.003376007080078125, 0.00466156005859375, 0.005947113037109375, 0.007232666015625, 0.008518218994140625, 0.00980377197265625, 0.011089324951171875, 0.0123748779296875, 0.013660430908203125, 0.01494598388671875, 0.016231536865234375, 0.01751708984375, 0.018802642822265625, 0.02008819580078125, 0.021373748779296875, 0.0226593017578125, 0.023944854736328125, 0.02523040771484375, 0.026515960693359375, 0.027801513671875, 0.029087066650390625, 0.03037261962890625, 0.031658172607421875, 0.0329437255859375, 0.034229278564453125, 0.03551483154296875, 0.036800384521484375, 0.0380859375]}, "parameters/_orig_mod.model.layers.3._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 2.0, 1.0, 1.0, 5.0, 2.0, 3.0, 6.0, 10.0, 10.0, 14.0, 27.0, 51.0, 60.0, 142.0, 261.0, 495.0, 897.0, 1613.0, 3582.0, 7069.0, 14740.0, 29770.0, 63718.0, 119196.0, 218878.0, 398372.0, 723198.0, 1133177.0, 1811246.0, 2572799.0, 2791408.0, 2518919.0, 1762384.0, 1092722.0, 694704.0, 379492.0, 208556.0, 113770.0, 60737.0, 28032.0, 13871.0, 6530.0, 3273.0, 1526.0, 875.0, 461.0, 278.0, 124.0, 79.0, 46.0, 25.0, 20.0, 14.0, 7.0, 5.0, 2.0, 2.0, 1.0, 2.0, 0.0, 0.0, 2.0], "bins": [-0.2080078125, -0.20159912109375, -0.1951904296875, -0.18878173828125, -0.182373046875, -0.17596435546875, -0.1695556640625, -0.16314697265625, -0.15673828125, -0.15032958984375, -0.1439208984375, -0.13751220703125, -0.131103515625, -0.12469482421875, -0.1182861328125, -0.11187744140625, -0.10546875, -0.09906005859375, -0.0926513671875, -0.08624267578125, -0.079833984375, -0.07342529296875, -0.0670166015625, -0.06060791015625, -0.05419921875, -0.04779052734375, -0.0413818359375, -0.03497314453125, -0.028564453125, -0.02215576171875, -0.0157470703125, -0.00933837890625, -0.0029296875, 0.00347900390625, 0.0098876953125, 0.01629638671875, 0.022705078125, 0.02911376953125, 0.0355224609375, 0.04193115234375, 0.04833984375, 0.05474853515625, 0.0611572265625, 0.06756591796875, 0.073974609375, 0.08038330078125, 0.0867919921875, 0.09320068359375, 0.099609375, 0.10601806640625, 0.1124267578125, 0.11883544921875, 0.125244140625, 0.13165283203125, 0.1380615234375, 0.14447021484375, 0.15087890625, 0.15728759765625, 0.1636962890625, 0.17010498046875, 0.176513671875, 0.18292236328125, 0.1893310546875, 0.19573974609375, 0.2021484375]}, "parameters/_orig_mod.model.layers.3._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 6.0, 16.0, 22.0, 34.0, 89.0, 133.0, 222.0, 404.0, 732.0, 1430.0, 2497.0, 4308.0, 7906.0, 12968.0, 21478.0, 38487.0, 65480.0, 106307.0, 160357.0, 265095.0, 376760.0, 537602.0, 642977.0, 613638.0, 459500.0, 323662.0, 207980.0, 135860.0, 86293.0, 51458.0, 28358.0, 18325.0, 10083.0, 6287.0, 3255.0, 1906.0, 1086.0, 601.0, 280.0, 187.0, 111.0, 63.0, 21.0, 15.0, 8.0, 3.0, 0.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.26953125, -0.26104736328125, -0.2525634765625, -0.24407958984375, -0.235595703125, -0.22711181640625, -0.2186279296875, -0.21014404296875, -0.20166015625, -0.19317626953125, -0.1846923828125, -0.17620849609375, -0.167724609375, -0.15924072265625, -0.1507568359375, -0.14227294921875, -0.1337890625, -0.12530517578125, -0.1168212890625, -0.10833740234375, -0.099853515625, -0.09136962890625, -0.0828857421875, -0.07440185546875, -0.06591796875, -0.05743408203125, -0.0489501953125, -0.04046630859375, -0.031982421875, -0.02349853515625, -0.0150146484375, -0.00653076171875, 0.001953125, 0.01043701171875, 0.0189208984375, 0.02740478515625, 0.035888671875, 0.04437255859375, 0.0528564453125, 0.06134033203125, 0.06982421875, 0.07830810546875, 0.0867919921875, 0.09527587890625, 0.103759765625, 0.11224365234375, 0.1207275390625, 0.12921142578125, 0.1376953125, 0.14617919921875, 0.1546630859375, 0.16314697265625, 0.171630859375, 0.18011474609375, 0.1885986328125, 0.19708251953125, 0.20556640625, 0.21405029296875, 0.2225341796875, 0.23101806640625, 0.239501953125, 0.24798583984375, 0.2564697265625, 0.26495361328125, 0.2734375]}, "parameters/_orig_mod.model.layers.3._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 0.0, 6.0, 11.0, 16.0, 52.0, 76.0, 155.0, 275.0, 509.0, 850.0, 1503.0, 2176.0, 3400.0, 4968.0, 7613.0, 10747.0, 14536.0, 20250.0, 25663.0, 30925.0, 41908.0, 45428.0, 56258.0, 53816.0, 60956.0, 62542.0, 60944.0, 56013.0, 55974.0, 46796.0, 42914.0, 33720.0, 25961.0, 21754.0, 16046.0, 11056.0, 8496.0, 5480.0, 3619.0, 2598.0, 1513.0, 910.0, 584.0, 312.0, 146.0, 98.0, 47.0, 26.0, 16.0, 7.0, 5.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.04150390625, -0.04000091552734375, -0.0384979248046875, -0.03699493408203125, -0.035491943359375, -0.03398895263671875, -0.0324859619140625, -0.03098297119140625, -0.02947998046875, -0.02797698974609375, -0.0264739990234375, -0.02497100830078125, -0.023468017578125, -0.02196502685546875, -0.0204620361328125, -0.01895904541015625, -0.0174560546875, -0.01595306396484375, -0.0144500732421875, -0.01294708251953125, -0.011444091796875, -0.00994110107421875, -0.0084381103515625, -0.00693511962890625, -0.00543212890625, -0.00392913818359375, -0.0024261474609375, -0.00092315673828125, 0.000579833984375, 0.00208282470703125, 0.0035858154296875, 0.00508880615234375, 0.006591796875, 0.00809478759765625, 0.0095977783203125, 0.01110076904296875, 0.012603759765625, 0.01410675048828125, 0.0156097412109375, 0.01711273193359375, 0.01861572265625, 0.02011871337890625, 0.0216217041015625, 0.02312469482421875, 0.024627685546875, 0.02613067626953125, 0.0276336669921875, 0.02913665771484375, 0.0306396484375, 0.03214263916015625, 0.0336456298828125, 0.03514862060546875, 0.036651611328125, 0.03815460205078125, 0.0396575927734375, 0.04116058349609375, 0.04266357421875, 0.04416656494140625, 0.0456695556640625, 0.04717254638671875, 0.048675537109375, 0.05017852783203125, 0.0516815185546875, 0.05318450927734375, 0.0546875]}, "parameters/_orig_mod.model.layers.4._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 3.0, 2.0, 2.0, 8.0, 9.0, 11.0, 21.0, 15.0, 23.0, 48.0, 45.0, 56.0, 81.0, 106.0, 135.0, 215.0, 385.0, 852.0, 2093.0, 5297.0, 14485.0, 44923.0, 125393.0, 322096.0, 820781.0, 1797193.0, 3530044.0, 4321295.0, 3117817.0, 1547014.0, 687893.0, 275832.0, 105920.0, 36427.0, 12370.0, 4600.0, 1785.0, 788.0, 368.0, 204.0, 144.0, 100.0, 77.0, 65.0, 51.0, 35.0, 24.0, 20.0, 18.0, 14.0, 7.0, 8.0, 5.0, 3.0, 3.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.306640625, -0.296875, -0.287109375, -0.27734375, -0.267578125, -0.2578125, -0.248046875, -0.23828125, -0.228515625, -0.21875, -0.208984375, -0.19921875, -0.189453125, -0.1796875, -0.169921875, -0.16015625, -0.150390625, -0.140625, -0.130859375, -0.12109375, -0.111328125, -0.1015625, -0.091796875, -0.08203125, -0.072265625, -0.0625, -0.052734375, -0.04296875, -0.033203125, -0.0234375, -0.013671875, -0.00390625, 0.005859375, 0.015625, 0.025390625, 0.03515625, 0.044921875, 0.0546875, 0.064453125, 0.07421875, 0.083984375, 0.09375, 0.103515625, 0.11328125, 0.123046875, 0.1328125, 0.142578125, 0.15234375, 0.162109375, 0.171875, 0.181640625, 0.19140625, 0.201171875, 0.2109375, 0.220703125, 0.23046875, 0.240234375, 0.25, 0.259765625, 0.26953125, 0.279296875, 0.2890625, 0.298828125, 0.30859375, 0.318359375]}, "parameters/_orig_mod.model.layers.4._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 3.0, 3.0, 3.0, 6.0, 10.0, 15.0, 13.0, 27.0, 29.0, 45.0, 37.0, 44.0, 62.0, 112.0, 125.0, 193.0, 400.0, 803.0, 1664.0, 3748.0, 6953.0, 15522.0, 29202.0, 59786.0, 110118.0, 186974.0, 327376.0, 501527.0, 706702.0, 743526.0, 582207.0, 388983.0, 234131.0, 138732.0, 76937.0, 39348.0, 19244.0, 10193.0, 4667.0, 2341.0, 1053.0, 533.0, 253.0, 172.0, 130.0, 90.0, 64.0, 43.0, 33.0, 32.0, 20.0, 19.0, 14.0, 8.0, 9.0, 7.0, 3.0, 2.0, 1.0, 3.0, 1.0], "bins": [-0.322265625, -0.312255859375, -0.30224609375, -0.292236328125, -0.2822265625, -0.272216796875, -0.26220703125, -0.252197265625, -0.2421875, -0.232177734375, -0.22216796875, -0.212158203125, -0.2021484375, -0.192138671875, -0.18212890625, -0.172119140625, -0.162109375, -0.152099609375, -0.14208984375, -0.132080078125, -0.1220703125, -0.112060546875, -0.10205078125, -0.092041015625, -0.08203125, -0.072021484375, -0.06201171875, -0.052001953125, -0.0419921875, -0.031982421875, -0.02197265625, -0.011962890625, -0.001953125, 0.008056640625, 0.01806640625, 0.028076171875, 0.0380859375, 0.048095703125, 0.05810546875, 0.068115234375, 0.078125, 0.088134765625, 0.09814453125, 0.108154296875, 0.1181640625, 0.128173828125, 0.13818359375, 0.148193359375, 0.158203125, 0.168212890625, 0.17822265625, 0.188232421875, 0.1982421875, 0.208251953125, 0.21826171875, 0.228271484375, 0.23828125, 0.248291015625, 0.25830078125, 0.268310546875, 0.2783203125, 0.288330078125, 0.29833984375, 0.308349609375, 0.318359375]}, "parameters/_orig_mod.model.layers.4._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [4.0, 1.0, 3.0, 3.0, 13.0, 22.0, 41.0, 58.0, 120.0, 202.0, 254.0, 486.0, 864.0, 1302.0, 2204.0, 3158.0, 4717.0, 6995.0, 9541.0, 14152.0, 18459.0, 24077.0, 31516.0, 39426.0, 49270.0, 55966.0, 61426.0, 66010.0, 68778.0, 64784.0, 61888.0, 53278.0, 48450.0, 38341.0, 30988.0, 23774.0, 18060.0, 12434.0, 9243.0, 6761.0, 4398.0, 3138.0, 1847.0, 1228.0, 869.0, 477.0, 278.0, 135.0, 105.0, 70.0, 23.0, 21.0, 6.0, 5.0, 1.0, 7.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.053466796875, -0.051586151123046875, -0.04970550537109375, -0.047824859619140625, -0.0459442138671875, -0.044063568115234375, -0.04218292236328125, -0.040302276611328125, -0.038421630859375, -0.036540985107421875, -0.03466033935546875, -0.032779693603515625, -0.0308990478515625, -0.029018402099609375, -0.02713775634765625, -0.025257110595703125, -0.02337646484375, -0.021495819091796875, -0.01961517333984375, -0.017734527587890625, -0.0158538818359375, -0.013973236083984375, -0.01209259033203125, -0.010211944580078125, -0.008331298828125, -0.006450653076171875, -0.00457000732421875, -0.002689361572265625, -0.0008087158203125, 0.001071929931640625, 0.00295257568359375, 0.004833221435546875, 0.0067138671875, 0.008594512939453125, 0.01047515869140625, 0.012355804443359375, 0.0142364501953125, 0.016117095947265625, 0.01799774169921875, 0.019878387451171875, 0.021759033203125, 0.023639678955078125, 0.02552032470703125, 0.027400970458984375, 0.0292816162109375, 0.031162261962890625, 0.03304290771484375, 0.034923553466796875, 0.03680419921875, 0.038684844970703125, 0.04056549072265625, 0.042446136474609375, 0.0443267822265625, 0.046207427978515625, 0.04808807373046875, 0.049968719482421875, 0.051849365234375, 0.053730010986328125, 0.05561065673828125, 0.057491302490234375, 0.0593719482421875, 0.061252593994140625, 0.06313323974609375, 0.06501388549804688, 0.06689453125]}, "parameters/_orig_mod.model.layers.5._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 0.0, 2.0, 3.0, 5.0, 10.0, 17.0, 23.0, 25.0, 37.0, 57.0, 73.0, 89.0, 157.0, 217.0, 297.0, 451.0, 740.0, 1348.0, 2271.0, 4854.0, 10237.0, 20996.0, 49857.0, 98647.0, 196090.0, 390514.0, 758084.0, 1299096.0, 2278388.0, 3134293.0, 3182929.0, 2349790.0, 1377829.0, 797063.0, 408728.0, 209197.0, 108447.0, 50871.0, 23491.0, 10610.0, 5128.0, 2621.0, 1362.0, 685.0, 463.0, 344.0, 222.0, 153.0, 149.0, 85.0, 55.0, 45.0, 17.0, 19.0, 13.0, 7.0, 6.0, 1.0, 2.0, 3.0], "bins": [-0.244140625, -0.236968994140625, -0.22979736328125, -0.222625732421875, -0.2154541015625, -0.208282470703125, -0.20111083984375, -0.193939208984375, -0.186767578125, -0.179595947265625, -0.17242431640625, -0.165252685546875, -0.1580810546875, -0.150909423828125, -0.14373779296875, -0.136566162109375, -0.12939453125, -0.122222900390625, -0.11505126953125, -0.107879638671875, -0.1007080078125, -0.093536376953125, -0.08636474609375, -0.079193115234375, -0.072021484375, -0.064849853515625, -0.05767822265625, -0.050506591796875, -0.0433349609375, -0.036163330078125, -0.02899169921875, -0.021820068359375, -0.0146484375, -0.007476806640625, -0.00030517578125, 0.006866455078125, 0.0140380859375, 0.021209716796875, 0.02838134765625, 0.035552978515625, 0.042724609375, 0.049896240234375, 0.05706787109375, 0.064239501953125, 0.0714111328125, 0.078582763671875, 0.08575439453125, 0.092926025390625, 0.10009765625, 0.107269287109375, 0.11444091796875, 0.121612548828125, 0.1287841796875, 0.135955810546875, 0.14312744140625, 0.150299072265625, 0.157470703125, 0.164642333984375, 0.17181396484375, 0.178985595703125, 0.1861572265625, 0.193328857421875, 0.20050048828125, 0.207672119140625, 0.21484375]}, "parameters/_orig_mod.model.layers.5._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 1.0, 4.0, 2.0, 2.0, 5.0, 3.0, 10.0, 13.0, 14.0, 14.0, 25.0, 30.0, 38.0, 59.0, 61.0, 75.0, 130.0, 128.0, 218.0, 420.0, 1221.0, 3717.0, 11280.0, 33313.0, 93507.0, 225471.0, 495213.0, 965213.0, 1117530.0, 691076.0, 326009.0, 146339.0, 53532.0, 19666.0, 6137.0, 2072.0, 732.0, 334.0, 176.0, 118.0, 96.0, 64.0, 48.0, 30.0, 32.0, 26.0, 17.0, 24.0, 14.0, 12.0, 7.0, 8.0, 1.0, 1.0, 0.0, 3.0, 3.0, 0.0, 5.0, 0.0, 2.0], "bins": [-0.4765625, -0.4613037109375, -0.446044921875, -0.4307861328125, -0.41552734375, -0.4002685546875, -0.385009765625, -0.3697509765625, -0.3544921875, -0.3392333984375, -0.323974609375, -0.3087158203125, -0.29345703125, -0.2781982421875, -0.262939453125, -0.2476806640625, -0.232421875, -0.2171630859375, -0.201904296875, -0.1866455078125, -0.17138671875, -0.1561279296875, -0.140869140625, -0.1256103515625, -0.1103515625, -0.0950927734375, -0.079833984375, -0.0645751953125, -0.04931640625, -0.0340576171875, -0.018798828125, -0.0035400390625, 0.01171875, 0.0269775390625, 0.042236328125, 0.0574951171875, 0.07275390625, 0.0880126953125, 0.103271484375, 0.1185302734375, 0.1337890625, 0.1490478515625, 0.164306640625, 0.1795654296875, 0.19482421875, 0.2100830078125, 0.225341796875, 0.2406005859375, 0.255859375, 0.2711181640625, 0.286376953125, 0.3016357421875, 0.31689453125, 0.3321533203125, 0.347412109375, 0.3626708984375, 0.3779296875, 0.3931884765625, 0.408447265625, 0.4237060546875, 0.43896484375, 0.4542236328125, 0.469482421875, 0.4847412109375, 0.5]}, "parameters/_orig_mod.model.layers.5._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 2.0, 4.0, 4.0, 8.0, 21.0, 53.0, 117.0, 220.0, 363.0, 727.0, 1252.0, 2253.0, 3484.0, 5825.0, 9489.0, 14512.0, 21044.0, 29475.0, 43365.0, 52528.0, 67165.0, 70052.0, 77687.0, 79497.0, 74619.0, 69926.0, 58173.0, 46272.0, 38022.0, 25165.0, 17921.0, 11910.0, 7399.0, 4628.0, 2802.0, 1613.0, 981.0, 537.0, 310.0, 123.0, 57.0, 45.0, 17.0, 6.0, 2.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.04296875, -0.041332244873046875, -0.03969573974609375, -0.038059234619140625, -0.0364227294921875, -0.034786224365234375, -0.03314971923828125, -0.031513214111328125, -0.029876708984375, -0.028240203857421875, -0.02660369873046875, -0.024967193603515625, -0.0233306884765625, -0.021694183349609375, -0.02005767822265625, -0.018421173095703125, -0.01678466796875, -0.015148162841796875, -0.01351165771484375, -0.011875152587890625, -0.0102386474609375, -0.008602142333984375, -0.00696563720703125, -0.005329132080078125, -0.003692626953125, -0.002056121826171875, -0.00041961669921875, 0.001216888427734375, 0.0028533935546875, 0.004489898681640625, 0.00612640380859375, 0.007762908935546875, 0.0093994140625, 0.011035919189453125, 0.01267242431640625, 0.014308929443359375, 0.0159454345703125, 0.017581939697265625, 0.01921844482421875, 0.020854949951171875, 0.022491455078125, 0.024127960205078125, 0.02576446533203125, 0.027400970458984375, 0.0290374755859375, 0.030673980712890625, 0.03231048583984375, 0.033946990966796875, 0.03558349609375, 0.037220001220703125, 0.03885650634765625, 0.040493011474609375, 0.0421295166015625, 0.043766021728515625, 0.04540252685546875, 0.047039031982421875, 0.048675537109375, 0.050312042236328125, 0.05194854736328125, 0.053585052490234375, 0.0552215576171875, 0.056858062744140625, 0.05849456787109375, 0.060131072998046875, 0.061767578125]}, "parameters/_orig_mod.model.layers.6._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 0.0, 0.0, 2.0, 7.0, 4.0, 3.0, 9.0, 5.0, 16.0, 29.0, 52.0, 77.0, 142.0, 312.0, 492.0, 967.0, 1994.0, 4258.0, 8028.0, 17843.0, 38288.0, 87338.0, 177152.0, 369644.0, 761132.0, 1401174.0, 2426021.0, 3170642.0, 3134479.0, 2379100.0, 1362360.0, 752611.0, 352725.0, 177627.0, 81955.0, 37493.0, 16895.0, 8337.0, 4086.0, 1783.0, 1030.0, 513.0, 253.0, 140.0, 75.0, 37.0, 27.0, 18.0, 8.0, 9.0, 9.0, 4.0, 2.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.236328125, -0.22869873046875, -0.2210693359375, -0.21343994140625, -0.205810546875, -0.19818115234375, -0.1905517578125, -0.18292236328125, -0.17529296875, -0.16766357421875, -0.1600341796875, -0.15240478515625, -0.144775390625, -0.13714599609375, -0.1295166015625, -0.12188720703125, -0.1142578125, -0.10662841796875, -0.0989990234375, -0.09136962890625, -0.083740234375, -0.07611083984375, -0.0684814453125, -0.06085205078125, -0.05322265625, -0.04559326171875, -0.0379638671875, -0.03033447265625, -0.022705078125, -0.01507568359375, -0.0074462890625, 0.00018310546875, 0.0078125, 0.01544189453125, 0.0230712890625, 0.03070068359375, 0.038330078125, 0.04595947265625, 0.0535888671875, 0.06121826171875, 0.06884765625, 0.07647705078125, 0.0841064453125, 0.09173583984375, 0.099365234375, 0.10699462890625, 0.1146240234375, 0.12225341796875, 0.1298828125, 0.13751220703125, 0.1451416015625, 0.15277099609375, 0.160400390625, 0.16802978515625, 0.1756591796875, 0.18328857421875, 0.19091796875, 0.19854736328125, 0.2061767578125, 0.21380615234375, 0.221435546875, 0.22906494140625, 0.2366943359375, 0.24432373046875, 0.251953125]}, "parameters/_orig_mod.model.layers.6._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 1.0, 1.0, 1.0, 1.0, 6.0, 12.0, 7.0, 23.0, 27.0, 31.0, 42.0, 77.0, 99.0, 152.0, 291.0, 486.0, 868.0, 1617.0, 2949.0, 4954.0, 9510.0, 17033.0, 28698.0, 54588.0, 88660.0, 140411.0, 226763.0, 334426.0, 448477.0, 556556.0, 569572.0, 525378.0, 402219.0, 296848.0, 191467.0, 119195.0, 74888.0, 44743.0, 22429.0, 13761.0, 7668.0, 4101.0, 2264.0, 1271.0, 684.0, 416.0, 238.0, 141.0, 82.0, 66.0, 33.0, 28.0, 10.0, 12.0, 7.0, 4.0, 4.0, 3.0, 1.0, 0.0, 1.0], "bins": [-0.271484375, -0.263336181640625, -0.25518798828125, -0.247039794921875, -0.2388916015625, -0.230743408203125, -0.22259521484375, -0.214447021484375, -0.206298828125, -0.198150634765625, -0.19000244140625, -0.181854248046875, -0.1737060546875, -0.165557861328125, -0.15740966796875, -0.149261474609375, -0.14111328125, -0.132965087890625, -0.12481689453125, -0.116668701171875, -0.1085205078125, -0.100372314453125, -0.09222412109375, -0.084075927734375, -0.075927734375, -0.067779541015625, -0.05963134765625, -0.051483154296875, -0.0433349609375, -0.035186767578125, -0.02703857421875, -0.018890380859375, -0.0107421875, -0.002593994140625, 0.00555419921875, 0.013702392578125, 0.0218505859375, 0.029998779296875, 0.03814697265625, 0.046295166015625, 0.054443359375, 0.062591552734375, 0.07073974609375, 0.078887939453125, 0.0870361328125, 0.095184326171875, 0.10333251953125, 0.111480712890625, 0.11962890625, 0.127777099609375, 0.13592529296875, 0.144073486328125, 0.1522216796875, 0.160369873046875, 0.16851806640625, 0.176666259765625, 0.184814453125, 0.192962646484375, 0.20111083984375, 0.209259033203125, 0.2174072265625, 0.225555419921875, 0.23370361328125, 0.241851806640625, 0.25]}, "parameters/_orig_mod.model.layers.6._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 5.0, 1.0, 7.0, 12.0, 22.0, 32.0, 74.0, 114.0, 147.0, 266.0, 477.0, 646.0, 970.0, 1713.0, 2151.0, 2874.0, 4534.0, 5785.0, 7594.0, 10724.0, 13567.0, 16644.0, 21413.0, 25738.0, 29174.0, 39006.0, 39912.0, 46763.0, 48477.0, 51204.0, 54077.0, 53464.0, 50638.0, 49303.0, 45201.0, 38949.0, 38525.0, 28283.0, 25113.0, 20808.0, 16087.0, 13031.0, 10285.0, 7107.0, 5390.0, 4371.0, 2818.0, 2035.0, 1538.0, 906.0, 636.0, 416.0, 239.0, 154.0, 119.0, 72.0, 25.0, 14.0, 14.0, 6.0, 6.0, 0.0, 2.0], "bins": [-0.044189453125, -0.042804718017578125, -0.04141998291015625, -0.040035247802734375, -0.0386505126953125, -0.037265777587890625, -0.03588104248046875, -0.034496307373046875, -0.033111572265625, -0.031726837158203125, -0.03034210205078125, -0.028957366943359375, -0.0275726318359375, -0.026187896728515625, -0.02480316162109375, -0.023418426513671875, -0.02203369140625, -0.020648956298828125, -0.01926422119140625, -0.017879486083984375, -0.0164947509765625, -0.015110015869140625, -0.01372528076171875, -0.012340545654296875, -0.010955810546875, -0.009571075439453125, -0.00818634033203125, -0.006801605224609375, -0.0054168701171875, -0.004032135009765625, -0.00264739990234375, -0.001262664794921875, 0.0001220703125, 0.001506805419921875, 0.00289154052734375, 0.004276275634765625, 0.0056610107421875, 0.007045745849609375, 0.00843048095703125, 0.009815216064453125, 0.011199951171875, 0.012584686279296875, 0.01396942138671875, 0.015354156494140625, 0.0167388916015625, 0.018123626708984375, 0.01950836181640625, 0.020893096923828125, 0.02227783203125, 0.023662567138671875, 0.02504730224609375, 0.026432037353515625, 0.0278167724609375, 0.029201507568359375, 0.03058624267578125, 0.031970977783203125, 0.033355712890625, 0.034740447998046875, 0.03612518310546875, 0.037509918212890625, 0.0388946533203125, 0.040279388427734375, 0.04166412353515625, 0.043048858642578125, 0.04443359375]}, "parameters/_orig_mod.model.layers.7._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 1.0, 4.0, 5.0, 10.0, 18.0, 34.0, 71.0, 137.0, 259.0, 561.0, 1175.0, 2535.0, 5369.0, 11609.0, 25247.0, 53979.0, 110362.0, 217119.0, 428895.0, 804135.0, 1398361.0, 2326315.0, 3032542.0, 3009460.0, 2304840.0, 1400052.0, 797118.0, 417531.0, 218049.0, 113123.0, 52395.0, 24868.0, 10760.0, 5340.0, 2520.0, 1192.0, 648.0, 302.0, 135.0, 57.0, 31.0, 20.0, 6.0, 11.0, 4.0, 2.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.2333984375, -0.226531982421875, -0.21966552734375, -0.212799072265625, -0.2059326171875, -0.199066162109375, -0.19219970703125, -0.185333251953125, -0.178466796875, -0.171600341796875, -0.16473388671875, -0.157867431640625, -0.1510009765625, -0.144134521484375, -0.13726806640625, -0.130401611328125, -0.12353515625, -0.116668701171875, -0.10980224609375, -0.102935791015625, -0.0960693359375, -0.089202880859375, -0.08233642578125, -0.075469970703125, -0.068603515625, -0.061737060546875, -0.05487060546875, -0.048004150390625, -0.0411376953125, -0.034271240234375, -0.02740478515625, -0.020538330078125, -0.013671875, -0.006805419921875, 6.103515625e-05, 0.006927490234375, 0.0137939453125, 0.020660400390625, 0.02752685546875, 0.034393310546875, 0.041259765625, 0.048126220703125, 0.05499267578125, 0.061859130859375, 0.0687255859375, 0.075592041015625, 0.08245849609375, 0.089324951171875, 0.09619140625, 0.103057861328125, 0.10992431640625, 0.116790771484375, 0.1236572265625, 0.130523681640625, 0.13739013671875, 0.144256591796875, 0.151123046875, 0.157989501953125, 0.16485595703125, 0.171722412109375, 0.1785888671875, 0.185455322265625, 0.19232177734375, 0.199188232421875, 0.2060546875]}, "parameters/_orig_mod.model.layers.7._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 5.0, 4.0, 3.0, 5.0, 18.0, 33.0, 43.0, 76.0, 129.0, 182.0, 332.0, 609.0, 1172.0, 1972.0, 4017.0, 6108.0, 11399.0, 17708.0, 30467.0, 49198.0, 76786.0, 113613.0, 165646.0, 241129.0, 319179.0, 416399.0, 488427.0, 491543.0, 464634.0, 377950.0, 291230.0, 209862.0, 144484.0, 99466.0, 66455.0, 41774.0, 25743.0, 14694.0, 9551.0, 5236.0, 3219.0, 1619.0, 992.0, 509.0, 300.0, 132.0, 95.0, 54.0, 35.0, 31.0, 10.0, 8.0, 9.0, 2.0, 0.0, 3.0], "bins": [-0.25, -0.242919921875, -0.23583984375, -0.228759765625, -0.2216796875, -0.214599609375, -0.20751953125, -0.200439453125, -0.193359375, -0.186279296875, -0.17919921875, -0.172119140625, -0.1650390625, -0.157958984375, -0.15087890625, -0.143798828125, -0.13671875, -0.129638671875, -0.12255859375, -0.115478515625, -0.1083984375, -0.101318359375, -0.09423828125, -0.087158203125, -0.080078125, -0.072998046875, -0.06591796875, -0.058837890625, -0.0517578125, -0.044677734375, -0.03759765625, -0.030517578125, -0.0234375, -0.016357421875, -0.00927734375, -0.002197265625, 0.0048828125, 0.011962890625, 0.01904296875, 0.026123046875, 0.033203125, 0.040283203125, 0.04736328125, 0.054443359375, 0.0615234375, 0.068603515625, 0.07568359375, 0.082763671875, 0.08984375, 0.096923828125, 0.10400390625, 0.111083984375, 0.1181640625, 0.125244140625, 0.13232421875, 0.139404296875, 0.146484375, 0.153564453125, 0.16064453125, 0.167724609375, 0.1748046875, 0.181884765625, 0.18896484375, 0.196044921875, 0.203125]}, "parameters/_orig_mod.model.layers.7._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 3.0, 5.0, 4.0, 8.0, 14.0, 29.0, 57.0, 106.0, 199.0, 377.0, 632.0, 991.0, 1768.0, 2609.0, 4573.0, 6893.0, 10882.0, 16020.0, 22327.0, 30594.0, 43707.0, 51084.0, 63182.0, 64291.0, 70080.0, 72349.0, 69854.0, 63097.0, 61190.0, 48946.0, 42035.0, 28954.0, 21377.0, 14812.0, 10086.0, 6336.0, 3971.0, 2610.0, 1496.0, 957.0, 486.0, 289.0, 189.0, 92.0, 56.0, 27.0, 18.0, 4.0, 4.0, 3.0, 0.0, 0.0, 1.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.04931640625, -0.047794342041015625, -0.04627227783203125, -0.044750213623046875, -0.0432281494140625, -0.041706085205078125, -0.04018402099609375, -0.038661956787109375, -0.037139892578125, -0.035617828369140625, -0.03409576416015625, -0.032573699951171875, -0.0310516357421875, -0.029529571533203125, -0.02800750732421875, -0.026485443115234375, -0.02496337890625, -0.023441314697265625, -0.02191925048828125, -0.020397186279296875, -0.0188751220703125, -0.017353057861328125, -0.01583099365234375, -0.014308929443359375, -0.012786865234375, -0.011264801025390625, -0.00974273681640625, -0.008220672607421875, -0.0066986083984375, -0.005176544189453125, -0.00365447998046875, -0.002132415771484375, -0.0006103515625, 0.000911712646484375, 0.00243377685546875, 0.003955841064453125, 0.0054779052734375, 0.006999969482421875, 0.00852203369140625, 0.010044097900390625, 0.011566162109375, 0.013088226318359375, 0.01461029052734375, 0.016132354736328125, 0.0176544189453125, 0.019176483154296875, 0.02069854736328125, 0.022220611572265625, 0.02374267578125, 0.025264739990234375, 0.02678680419921875, 0.028308868408203125, 0.0298309326171875, 0.031352996826171875, 0.03287506103515625, 0.034397125244140625, 0.035919189453125, 0.037441253662109375, 0.03896331787109375, 0.040485382080078125, 0.0420074462890625, 0.043529510498046875, 0.04505157470703125, 0.046573638916015625, 0.048095703125]}, "parameters/_orig_mod.model.layers.8._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 4.0, 1.0, 0.0, 5.0, 2.0, 7.0, 5.0, 3.0, 4.0, 18.0, 13.0, 16.0, 28.0, 37.0, 40.0, 42.0, 66.0, 82.0, 121.0, 176.0, 309.0, 680.0, 1784.0, 5477.0, 16968.0, 65438.0, 225994.0, 791543.0, 2511358.0, 5299047.0, 4904912.0, 2062913.0, 636346.0, 183027.0, 49845.0, 13671.0, 4185.0, 1441.0, 688.0, 293.0, 147.0, 121.0, 87.0, 59.0, 45.0, 20.0, 40.0, 25.0, 20.0, 13.0, 10.0, 8.0, 6.0, 6.0, 8.0, 2.0, 4.0, 1.0, 0.0, 3.0], "bins": [-0.43359375, -0.420806884765625, -0.40802001953125, -0.395233154296875, -0.3824462890625, -0.369659423828125, -0.35687255859375, -0.344085693359375, -0.331298828125, -0.318511962890625, -0.30572509765625, -0.292938232421875, -0.2801513671875, -0.267364501953125, -0.25457763671875, -0.241790771484375, -0.22900390625, -0.216217041015625, -0.20343017578125, -0.190643310546875, -0.1778564453125, -0.165069580078125, -0.15228271484375, -0.139495849609375, -0.126708984375, -0.113922119140625, -0.10113525390625, -0.088348388671875, -0.0755615234375, -0.062774658203125, -0.04998779296875, -0.037200927734375, -0.0244140625, -0.011627197265625, 0.00115966796875, 0.013946533203125, 0.0267333984375, 0.039520263671875, 0.05230712890625, 0.065093994140625, 0.077880859375, 0.090667724609375, 0.10345458984375, 0.116241455078125, 0.1290283203125, 0.141815185546875, 0.15460205078125, 0.167388916015625, 0.18017578125, 0.192962646484375, 0.20574951171875, 0.218536376953125, 0.2313232421875, 0.244110107421875, 0.25689697265625, 0.269683837890625, 0.282470703125, 0.295257568359375, 0.30804443359375, 0.320831298828125, 0.3336181640625, 0.346405029296875, 0.35919189453125, 0.371978759765625, 0.384765625]}, "parameters/_orig_mod.model.layers.8._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 3.0, 3.0, 5.0, 0.0, 3.0, 4.0, 5.0, 8.0, 12.0, 25.0, 26.0, 43.0, 49.0, 79.0, 140.0, 245.0, 402.0, 838.0, 2097.0, 4493.0, 10899.0, 26442.0, 62315.0, 148671.0, 319590.0, 625383.0, 956274.0, 930084.0, 585712.0, 289139.0, 133089.0, 57520.0, 22920.0, 10162.0, 4102.0, 1674.0, 851.0, 410.0, 197.0, 127.0, 87.0, 43.0, 45.0, 35.0, 8.0, 14.0, 10.0, 8.0, 0.0, 3.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.4453125, -0.431793212890625, -0.41827392578125, -0.404754638671875, -0.3912353515625, -0.377716064453125, -0.36419677734375, -0.350677490234375, -0.337158203125, -0.323638916015625, -0.31011962890625, -0.296600341796875, -0.2830810546875, -0.269561767578125, -0.25604248046875, -0.242523193359375, -0.22900390625, -0.215484619140625, -0.20196533203125, -0.188446044921875, -0.1749267578125, -0.161407470703125, -0.14788818359375, -0.134368896484375, -0.120849609375, -0.107330322265625, -0.09381103515625, -0.080291748046875, -0.0667724609375, -0.053253173828125, -0.03973388671875, -0.026214599609375, -0.0126953125, 0.000823974609375, 0.01434326171875, 0.027862548828125, 0.0413818359375, 0.054901123046875, 0.06842041015625, 0.081939697265625, 0.095458984375, 0.108978271484375, 0.12249755859375, 0.136016845703125, 0.1495361328125, 0.163055419921875, 0.17657470703125, 0.190093994140625, 0.20361328125, 0.217132568359375, 0.23065185546875, 0.244171142578125, 0.2576904296875, 0.271209716796875, 0.28472900390625, 0.298248291015625, 0.311767578125, 0.325286865234375, 0.33880615234375, 0.352325439453125, 0.3658447265625, 0.379364013671875, 0.39288330078125, 0.406402587890625, 0.419921875]}, "parameters/_orig_mod.model.layers.8._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0, 2.0, 8.0, 19.0, 30.0, 38.0, 84.0, 148.0, 189.0, 351.0, 533.0, 786.0, 1173.0, 1762.0, 2605.0, 3706.0, 5545.0, 8050.0, 11425.0, 16047.0, 22114.0, 29063.0, 42438.0, 47955.0, 60865.0, 61399.0, 68410.0, 70627.0, 67672.0, 61356.0, 60059.0, 48423.0, 41897.0, 29619.0, 21911.0, 16507.0, 11349.0, 8088.0, 5508.0, 3804.0, 2665.0, 1857.0, 1219.0, 854.0, 547.0, 396.0, 215.0, 135.0, 80.0, 58.0, 29.0, 26.0, 11.0, 8.0, 3.0, 3.0, 4.0], "bins": [-0.050537109375, -0.049072265625, -0.047607421875, -0.046142578125, -0.044677734375, -0.043212890625, -0.041748046875, -0.040283203125, -0.038818359375, -0.037353515625, -0.035888671875, -0.034423828125, -0.032958984375, -0.031494140625, -0.030029296875, -0.028564453125, -0.027099609375, -0.025634765625, -0.024169921875, -0.022705078125, -0.021240234375, -0.019775390625, -0.018310546875, -0.016845703125, -0.015380859375, -0.013916015625, -0.012451171875, -0.010986328125, -0.009521484375, -0.008056640625, -0.006591796875, -0.005126953125, -0.003662109375, -0.002197265625, -0.000732421875, 0.000732421875, 0.002197265625, 0.003662109375, 0.005126953125, 0.006591796875, 0.008056640625, 0.009521484375, 0.010986328125, 0.012451171875, 0.013916015625, 0.015380859375, 0.016845703125, 0.018310546875, 0.019775390625, 0.021240234375, 0.022705078125, 0.024169921875, 0.025634765625, 0.027099609375, 0.028564453125, 0.030029296875, 0.031494140625, 0.032958984375, 0.034423828125, 0.035888671875, 0.037353515625, 0.038818359375, 0.040283203125, 0.041748046875, 0.043212890625]}, "parameters/_orig_mod.model.layers.9._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [4.0, 3.0, 1.0, 1.0, 2.0, 4.0, 5.0, 7.0, 22.0, 37.0, 50.0, 96.0, 202.0, 362.0, 558.0, 1098.0, 2052.0, 3653.0, 7386.0, 14160.0, 29917.0, 62253.0, 124200.0, 235062.0, 494957.0, 938935.0, 1668612.0, 2657859.0, 3181738.0, 2956790.0, 2012691.0, 1132257.0, 629928.0, 308869.0, 158149.0, 79595.0, 38721.0, 17867.0, 9087.0, 4526.0, 2407.0, 1425.0, 702.0, 418.0, 231.0, 145.0, 70.0, 33.0, 26.0, 11.0, 13.0, 7.0, 4.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 1.0], "bins": [-0.2021484375, -0.195098876953125, -0.18804931640625, -0.180999755859375, -0.1739501953125, -0.166900634765625, -0.15985107421875, -0.152801513671875, -0.145751953125, -0.138702392578125, -0.13165283203125, -0.124603271484375, -0.1175537109375, -0.110504150390625, -0.10345458984375, -0.096405029296875, -0.08935546875, -0.082305908203125, -0.07525634765625, -0.068206787109375, -0.0611572265625, -0.054107666015625, -0.04705810546875, -0.040008544921875, -0.032958984375, -0.025909423828125, -0.01885986328125, -0.011810302734375, -0.0047607421875, 0.002288818359375, 0.00933837890625, 0.016387939453125, 0.0234375, 0.030487060546875, 0.03753662109375, 0.044586181640625, 0.0516357421875, 0.058685302734375, 0.06573486328125, 0.072784423828125, 0.079833984375, 0.086883544921875, 0.09393310546875, 0.100982666015625, 0.1080322265625, 0.115081787109375, 0.12213134765625, 0.129180908203125, 0.13623046875, 0.143280029296875, 0.15032958984375, 0.157379150390625, 0.1644287109375, 0.171478271484375, 0.17852783203125, 0.185577392578125, 0.192626953125, 0.199676513671875, 0.20672607421875, 0.213775634765625, 0.2208251953125, 0.227874755859375, 0.23492431640625, 0.241973876953125, 0.2490234375]}, "parameters/_orig_mod.model.layers.9._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 5.0, 6.0, 1.0, 9.0, 9.0, 10.0, 18.0, 35.0, 69.0, 107.0, 193.0, 405.0, 782.0, 1688.0, 3424.0, 6946.0, 13442.0, 24866.0, 48930.0, 89670.0, 153907.0, 267186.0, 424707.0, 630664.0, 706885.0, 667992.0, 464926.0, 296862.0, 172770.0, 101334.0, 56031.0, 28783.0, 15682.0, 7976.0, 4157.0, 1951.0, 935.0, 422.0, 222.0, 128.0, 59.0, 33.0, 21.0, 15.0, 8.0, 7.0, 9.0, 3.0, 2.0, 4.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.26953125, -0.259765625, -0.25, -0.240234375, -0.23046875, -0.220703125, -0.2109375, -0.201171875, -0.19140625, -0.181640625, -0.171875, -0.162109375, -0.15234375, -0.142578125, -0.1328125, -0.123046875, -0.11328125, -0.103515625, -0.09375, -0.083984375, -0.07421875, -0.064453125, -0.0546875, -0.044921875, -0.03515625, -0.025390625, -0.015625, -0.005859375, 0.00390625, 0.013671875, 0.0234375, 0.033203125, 0.04296875, 0.052734375, 0.0625, 0.072265625, 0.08203125, 0.091796875, 0.1015625, 0.111328125, 0.12109375, 0.130859375, 0.140625, 0.150390625, 0.16015625, 0.169921875, 0.1796875, 0.189453125, 0.19921875, 0.208984375, 0.21875, 0.228515625, 0.23828125, 0.248046875, 0.2578125, 0.267578125, 0.27734375, 0.287109375, 0.296875, 0.306640625, 0.31640625, 0.326171875, 0.3359375, 0.345703125, 0.35546875]}, "parameters/_orig_mod.model.layers.9._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 4.0, 2.0, 1.0, 5.0, 17.0, 13.0, 51.0, 82.0, 93.0, 164.0, 285.0, 432.0, 662.0, 1110.0, 1386.0, 1932.0, 2745.0, 3939.0, 5765.0, 7559.0, 10428.0, 15196.0, 22288.0, 31629.0, 44779.0, 61113.0, 77815.0, 86221.0, 94658.0, 84749.0, 78219.0, 59263.0, 43768.0, 31083.0, 22056.0, 14264.0, 10128.0, 7513.0, 5440.0, 3945.0, 2901.0, 1863.0, 1329.0, 907.0, 683.0, 469.0, 281.0, 182.0, 87.0, 59.0, 62.0, 19.0, 21.0, 8.0, 1.0, 3.0], "bins": [-0.08984375, -0.08737945556640625, -0.0849151611328125, -0.08245086669921875, -0.079986572265625, -0.07752227783203125, -0.0750579833984375, -0.07259368896484375, -0.07012939453125, -0.06766510009765625, -0.0652008056640625, -0.06273651123046875, -0.060272216796875, -0.05780792236328125, -0.0553436279296875, -0.05287933349609375, -0.0504150390625, -0.04795074462890625, -0.0454864501953125, -0.04302215576171875, -0.040557861328125, -0.03809356689453125, -0.0356292724609375, -0.03316497802734375, -0.03070068359375, -0.02823638916015625, -0.0257720947265625, -0.02330780029296875, -0.020843505859375, -0.01837921142578125, -0.0159149169921875, -0.01345062255859375, -0.010986328125, -0.00852203369140625, -0.0060577392578125, -0.00359344482421875, -0.001129150390625, 0.00133514404296875, 0.0037994384765625, 0.00626373291015625, 0.00872802734375, 0.01119232177734375, 0.0136566162109375, 0.01612091064453125, 0.018585205078125, 0.02104949951171875, 0.0235137939453125, 0.02597808837890625, 0.0284423828125, 0.03090667724609375, 0.0333709716796875, 0.03583526611328125, 0.038299560546875, 0.04076385498046875, 0.0432281494140625, 0.04569244384765625, 0.04815673828125, 0.05062103271484375, 0.0530853271484375, 0.05554962158203125, 0.058013916015625, 0.06047821044921875, 0.0629425048828125, 0.06540679931640625, 0.06787109375]}, "parameters/_orig_mod.model.layers.10._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 1.0, 2.0, 5.0, 4.0, 13.0, 22.0, 23.0, 29.0, 28.0, 46.0, 70.0, 72.0, 134.0, 199.0, 328.0, 793.0, 1921.0, 5074.0, 14964.0, 42869.0, 124225.0, 342355.0, 947860.0, 2362985.0, 4698632.0, 4577208.0, 2271303.0, 886817.0, 316819.0, 120334.0, 39540.0, 14082.0, 4985.0, 1719.0, 774.0, 351.0, 182.0, 121.0, 86.0, 54.0, 52.0, 35.0, 20.0, 19.0, 17.0, 11.0, 7.0, 2.0, 4.0, 5.0, 2.0, 2.0, 1.0, 2.0, 1.0, 1.0], "bins": [-0.365234375, -0.354156494140625, -0.34307861328125, -0.332000732421875, -0.3209228515625, -0.309844970703125, -0.29876708984375, -0.287689208984375, -0.276611328125, -0.265533447265625, -0.25445556640625, -0.243377685546875, -0.2322998046875, -0.221221923828125, -0.21014404296875, -0.199066162109375, -0.18798828125, -0.176910400390625, -0.16583251953125, -0.154754638671875, -0.1436767578125, -0.132598876953125, -0.12152099609375, -0.110443115234375, -0.099365234375, -0.088287353515625, -0.07720947265625, -0.066131591796875, -0.0550537109375, -0.043975830078125, -0.03289794921875, -0.021820068359375, -0.0107421875, 0.000335693359375, 0.01141357421875, 0.022491455078125, 0.0335693359375, 0.044647216796875, 0.05572509765625, 0.066802978515625, 0.077880859375, 0.088958740234375, 0.10003662109375, 0.111114501953125, 0.1221923828125, 0.133270263671875, 0.14434814453125, 0.155426025390625, 0.16650390625, 0.177581787109375, 0.18865966796875, 0.199737548828125, 0.2108154296875, 0.221893310546875, 0.23297119140625, 0.244049072265625, 0.255126953125, 0.266204833984375, 0.27728271484375, 0.288360595703125, 0.2994384765625, 0.310516357421875, 0.32159423828125, 0.332672119140625, 0.34375]}, "parameters/_orig_mod.model.layers.10._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 2.0, 2.0, 5.0, 12.0, 7.0, 11.0, 12.0, 29.0, 37.0, 38.0, 65.0, 78.0, 102.0, 185.0, 247.0, 435.0, 1028.0, 2095.0, 5107.0, 10371.0, 24548.0, 52843.0, 114934.0, 219355.0, 431182.0, 720336.0, 880967.0, 761053.0, 474411.0, 251610.0, 130403.0, 60657.0, 28345.0, 12930.0, 5771.0, 2442.0, 1168.0, 554.0, 278.0, 174.0, 134.0, 78.0, 65.0, 57.0, 35.0, 31.0, 24.0, 11.0, 11.0, 4.0, 9.0, 2.0, 4.0, 1.0, 1.0, 2.0], "bins": [-0.41796875, -0.4058837890625, -0.393798828125, -0.3817138671875, -0.36962890625, -0.3575439453125, -0.345458984375, -0.3333740234375, -0.3212890625, -0.3092041015625, -0.297119140625, -0.2850341796875, -0.27294921875, -0.2608642578125, -0.248779296875, -0.2366943359375, -0.224609375, -0.2125244140625, -0.200439453125, -0.1883544921875, -0.17626953125, -0.1641845703125, -0.152099609375, -0.1400146484375, -0.1279296875, -0.1158447265625, -0.103759765625, -0.0916748046875, -0.07958984375, -0.0675048828125, -0.055419921875, -0.0433349609375, -0.03125, -0.0191650390625, -0.007080078125, 0.0050048828125, 0.01708984375, 0.0291748046875, 0.041259765625, 0.0533447265625, 0.0654296875, 0.0775146484375, 0.089599609375, 0.1016845703125, 0.11376953125, 0.1258544921875, 0.137939453125, 0.1500244140625, 0.162109375, 0.1741943359375, 0.186279296875, 0.1983642578125, 0.21044921875, 0.2225341796875, 0.234619140625, 0.2467041015625, 0.2587890625, 0.2708740234375, 0.282958984375, 0.2950439453125, 0.30712890625, 0.3192138671875, 0.331298828125, 0.3433837890625, 0.35546875]}, "parameters/_orig_mod.model.layers.10._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 3.0, 2.0, 6.0, 9.0, 25.0, 22.0, 54.0, 74.0, 108.0, 166.0, 267.0, 344.0, 544.0, 739.0, 1107.0, 1602.0, 2126.0, 2738.0, 3790.0, 5172.0, 6947.0, 9344.0, 12993.0, 18354.0, 23395.0, 30053.0, 41962.0, 49072.0, 59723.0, 60868.0, 67017.0, 68753.0, 65106.0, 61270.0, 54575.0, 45035.0, 38544.0, 27365.0, 20979.0, 16394.0, 11543.0, 8286.0, 6101.0, 4658.0, 3473.0, 2557.0, 1879.0, 1475.0, 958.0, 648.0, 481.0, 332.0, 212.0, 169.0, 81.0, 71.0, 39.0, 25.0, 17.0, 10.0, 5.0, 9.0, 2.0], "bins": [-0.051513671875, -0.04991912841796875, -0.0483245849609375, -0.04673004150390625, -0.045135498046875, -0.04354095458984375, -0.0419464111328125, -0.04035186767578125, -0.03875732421875, -0.03716278076171875, -0.0355682373046875, -0.03397369384765625, -0.032379150390625, -0.03078460693359375, -0.0291900634765625, -0.02759552001953125, -0.0260009765625, -0.02440643310546875, -0.0228118896484375, -0.02121734619140625, -0.019622802734375, -0.01802825927734375, -0.0164337158203125, -0.01483917236328125, -0.01324462890625, -0.01165008544921875, -0.0100555419921875, -0.00846099853515625, -0.006866455078125, -0.00527191162109375, -0.0036773681640625, -0.00208282470703125, -0.00048828125, 0.00110626220703125, 0.0027008056640625, 0.00429534912109375, 0.005889892578125, 0.00748443603515625, 0.0090789794921875, 0.01067352294921875, 0.01226806640625, 0.01386260986328125, 0.0154571533203125, 0.01705169677734375, 0.018646240234375, 0.02024078369140625, 0.0218353271484375, 0.02342987060546875, 0.0250244140625, 0.02661895751953125, 0.0282135009765625, 0.02980804443359375, 0.031402587890625, 0.03299713134765625, 0.0345916748046875, 0.03618621826171875, 0.03778076171875, 0.03937530517578125, 0.0409698486328125, 0.04256439208984375, 0.044158935546875, 0.04575347900390625, 0.0473480224609375, 0.04894256591796875, 0.050537109375]}, "parameters/_orig_mod.model.layers.11._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [5.0, 5.0, 6.0, 13.0, 19.0, 45.0, 79.0, 147.0, 246.0, 404.0, 743.0, 1434.0, 2744.0, 4157.0, 7683.0, 13956.0, 26402.0, 46432.0, 76123.0, 135737.0, 217834.0, 382567.0, 623730.0, 982554.0, 1529843.0, 2132339.0, 2458956.0, 2420442.0, 2024085.0, 1390650.0, 896690.0, 572893.0, 347919.0, 201686.0, 115621.0, 71401.0, 39474.0, 23435.0, 12750.0, 6908.0, 4082.0, 2059.0, 1336.0, 674.0, 421.0, 196.0, 133.0, 67.0, 29.0, 18.0, 12.0, 14.0, 9.0, 4.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.1416015625, -0.1363372802734375, -0.131072998046875, -0.1258087158203125, -0.12054443359375, -0.1152801513671875, -0.110015869140625, -0.1047515869140625, -0.0994873046875, -0.0942230224609375, -0.088958740234375, -0.0836944580078125, -0.07843017578125, -0.0731658935546875, -0.067901611328125, -0.0626373291015625, -0.057373046875, -0.0521087646484375, -0.046844482421875, -0.0415802001953125, -0.03631591796875, -0.0310516357421875, -0.025787353515625, -0.0205230712890625, -0.0152587890625, -0.0099945068359375, -0.004730224609375, 0.0005340576171875, 0.00579833984375, 0.0110626220703125, 0.016326904296875, 0.0215911865234375, 0.02685546875, 0.0321197509765625, 0.037384033203125, 0.0426483154296875, 0.04791259765625, 0.0531768798828125, 0.058441162109375, 0.0637054443359375, 0.0689697265625, 0.0742340087890625, 0.079498291015625, 0.0847625732421875, 0.09002685546875, 0.0952911376953125, 0.100555419921875, 0.1058197021484375, 0.111083984375, 0.1163482666015625, 0.121612548828125, 0.1268768310546875, 0.13214111328125, 0.1374053955078125, 0.142669677734375, 0.1479339599609375, 0.1531982421875, 0.1584625244140625, 0.163726806640625, 0.1689910888671875, 0.17425537109375, 0.1795196533203125, 0.184783935546875, 0.1900482177734375, 0.1953125]}, "parameters/_orig_mod.model.layers.11._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 4.0, 5.0, 3.0, 3.0, 8.0, 18.0, 17.0, 32.0, 51.0, 67.0, 108.0, 168.0, 277.0, 543.0, 1078.0, 2307.0, 4760.0, 9295.0, 18348.0, 34079.0, 64552.0, 111221.0, 184008.0, 303955.0, 456386.0, 636427.0, 662874.0, 601278.0, 424882.0, 281706.0, 170595.0, 103352.0, 58159.0, 30379.0, 16427.0, 8400.0, 4246.0, 2061.0, 1030.0, 517.0, 239.0, 159.0, 107.0, 59.0, 32.0, 28.0, 13.0, 10.0, 8.0, 6.0, 4.0, 6.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.28125, -0.272003173828125, -0.26275634765625, -0.253509521484375, -0.2442626953125, -0.235015869140625, -0.22576904296875, -0.216522216796875, -0.207275390625, -0.198028564453125, -0.18878173828125, -0.179534912109375, -0.1702880859375, -0.161041259765625, -0.15179443359375, -0.142547607421875, -0.13330078125, -0.124053955078125, -0.11480712890625, -0.105560302734375, -0.0963134765625, -0.087066650390625, -0.07781982421875, -0.068572998046875, -0.059326171875, -0.050079345703125, -0.04083251953125, -0.031585693359375, -0.0223388671875, -0.013092041015625, -0.00384521484375, 0.005401611328125, 0.0146484375, 0.023895263671875, 0.03314208984375, 0.042388916015625, 0.0516357421875, 0.060882568359375, 0.07012939453125, 0.079376220703125, 0.088623046875, 0.097869873046875, 0.10711669921875, 0.116363525390625, 0.1256103515625, 0.134857177734375, 0.14410400390625, 0.153350830078125, 0.16259765625, 0.171844482421875, 0.18109130859375, 0.190338134765625, 0.1995849609375, 0.208831787109375, 0.21807861328125, 0.227325439453125, 0.236572265625, 0.245819091796875, 0.25506591796875, 0.264312744140625, 0.2735595703125, 0.282806396484375, 0.29205322265625, 0.301300048828125, 0.310546875]}, "parameters/_orig_mod.model.layers.11._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 4.0, 1.0, 1.0, 7.0, 19.0, 34.0, 65.0, 161.0, 319.0, 628.0, 1029.0, 1990.0, 3113.0, 5418.0, 8546.0, 12755.0, 18448.0, 25543.0, 34214.0, 47164.0, 54412.0, 65485.0, 67992.0, 73129.0, 72671.0, 68335.0, 65079.0, 54217.0, 46386.0, 34998.0, 25654.0, 18424.0, 12645.0, 7919.0, 5334.0, 3184.0, 1990.0, 1149.0, 573.0, 343.0, 143.0, 87.0, 26.0, 26.0, 5.0, 6.0, 2.0, 1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.044677734375, -0.04302215576171875, -0.0413665771484375, -0.03971099853515625, -0.038055419921875, -0.03639984130859375, -0.0347442626953125, -0.03308868408203125, -0.03143310546875, -0.02977752685546875, -0.0281219482421875, -0.02646636962890625, -0.024810791015625, -0.02315521240234375, -0.0214996337890625, -0.01984405517578125, -0.0181884765625, -0.01653289794921875, -0.0148773193359375, -0.01322174072265625, -0.011566162109375, -0.00991058349609375, -0.0082550048828125, -0.00659942626953125, -0.00494384765625, -0.00328826904296875, -0.0016326904296875, 2.288818359375e-05, 0.001678466796875, 0.00333404541015625, 0.0049896240234375, 0.00664520263671875, 0.00830078125, 0.00995635986328125, 0.0116119384765625, 0.01326751708984375, 0.014923095703125, 0.01657867431640625, 0.0182342529296875, 0.01988983154296875, 0.02154541015625, 0.02320098876953125, 0.0248565673828125, 0.02651214599609375, 0.028167724609375, 0.02982330322265625, 0.0314788818359375, 0.03313446044921875, 0.0347900390625, 0.03644561767578125, 0.0381011962890625, 0.03975677490234375, 0.041412353515625, 0.04306793212890625, 0.0447235107421875, 0.04637908935546875, 0.04803466796875, 0.04969024658203125, 0.0513458251953125, 0.05300140380859375, 0.054656982421875, 0.05631256103515625, 0.0579681396484375, 0.05962371826171875, 0.061279296875]}, "parameters/_orig_mod.model.layers.12._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 4.0, 0.0, 6.0, 3.0, 8.0, 7.0, 23.0, 31.0, 67.0, 113.0, 263.0, 503.0, 925.0, 2274.0, 4786.0, 11263.0, 27350.0, 60112.0, 134654.0, 265555.0, 563638.0, 1019376.0, 1741509.0, 2545042.0, 2928725.0, 2793059.0, 2034438.0, 1241379.0, 719147.0, 361664.0, 176469.0, 79877.0, 37423.0, 15385.0, 6642.0, 2819.0, 1407.0, 647.0, 329.0, 132.0, 57.0, 36.0, 14.0, 17.0, 11.0, 3.0, 9.0, 6.0, 1.0, 0.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.2333984375, -0.2266693115234375, -0.219940185546875, -0.2132110595703125, -0.20648193359375, -0.1997528076171875, -0.193023681640625, -0.1862945556640625, -0.1795654296875, -0.1728363037109375, -0.166107177734375, -0.1593780517578125, -0.15264892578125, -0.1459197998046875, -0.139190673828125, -0.1324615478515625, -0.125732421875, -0.1190032958984375, -0.112274169921875, -0.1055450439453125, -0.09881591796875, -0.0920867919921875, -0.085357666015625, -0.0786285400390625, -0.0718994140625, -0.0651702880859375, -0.058441162109375, -0.0517120361328125, -0.04498291015625, -0.0382537841796875, -0.031524658203125, -0.0247955322265625, -0.01806640625, -0.0113372802734375, -0.004608154296875, 0.0021209716796875, 0.00885009765625, 0.0155792236328125, 0.022308349609375, 0.0290374755859375, 0.0357666015625, 0.0424957275390625, 0.049224853515625, 0.0559539794921875, 0.06268310546875, 0.0694122314453125, 0.076141357421875, 0.0828704833984375, 0.089599609375, 0.0963287353515625, 0.103057861328125, 0.1097869873046875, 0.11651611328125, 0.1232452392578125, 0.129974365234375, 0.1367034912109375, 0.1434326171875, 0.1501617431640625, 0.156890869140625, 0.1636199951171875, 0.17034912109375, 0.1770782470703125, 0.183807373046875, 0.1905364990234375, 0.197265625]}, "parameters/_orig_mod.model.layers.12._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 3.0, 0.0, 2.0, 8.0, 10.0, 14.0, 15.0, 30.0, 61.0, 127.0, 203.0, 370.0, 716.0, 1320.0, 2546.0, 4375.0, 7892.0, 14449.0, 23101.0, 39764.0, 67699.0, 98776.0, 145885.0, 224089.0, 294638.0, 384924.0, 463637.0, 480723.0, 478905.0, 412206.0, 325791.0, 247876.0, 166511.0, 118245.0, 74534.0, 47984.0, 29600.0, 15697.0, 9584.0, 5399.0, 2995.0, 1649.0, 968.0, 458.0, 211.0, 121.0, 77.0, 28.0, 23.0, 23.0, 16.0, 12.0, 3.0, 4.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.205078125, -0.1981658935546875, -0.191253662109375, -0.1843414306640625, -0.17742919921875, -0.1705169677734375, -0.163604736328125, -0.1566925048828125, -0.1497802734375, -0.1428680419921875, -0.135955810546875, -0.1290435791015625, -0.12213134765625, -0.1152191162109375, -0.108306884765625, -0.1013946533203125, -0.094482421875, -0.0875701904296875, -0.080657958984375, -0.0737457275390625, -0.06683349609375, -0.0599212646484375, -0.053009033203125, -0.0460968017578125, -0.0391845703125, -0.0322723388671875, -0.025360107421875, -0.0184478759765625, -0.01153564453125, -0.0046234130859375, 0.002288818359375, 0.0092010498046875, 0.01611328125, 0.0230255126953125, 0.029937744140625, 0.0368499755859375, 0.04376220703125, 0.0506744384765625, 0.057586669921875, 0.0644989013671875, 0.0714111328125, 0.0783233642578125, 0.085235595703125, 0.0921478271484375, 0.09906005859375, 0.1059722900390625, 0.112884521484375, 0.1197967529296875, 0.126708984375, 0.1336212158203125, 0.140533447265625, 0.1474456787109375, 0.15435791015625, 0.1612701416015625, 0.168182373046875, 0.1750946044921875, 0.1820068359375, 0.1889190673828125, 0.195831298828125, 0.2027435302734375, 0.20965576171875, 0.2165679931640625, 0.223480224609375, 0.2303924560546875, 0.2373046875]}, "parameters/_orig_mod.model.layers.12._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 2.0, 7.0, 4.0, 17.0, 16.0, 34.0, 71.0, 79.0, 146.0, 210.0, 344.0, 572.0, 846.0, 1311.0, 1722.0, 2795.0, 4077.0, 5621.0, 7844.0, 10319.0, 13154.0, 18098.0, 22280.0, 28218.0, 33684.0, 40961.0, 44761.0, 53809.0, 51157.0, 56485.0, 56307.0, 55829.0, 52419.0, 50522.0, 43291.0, 41240.0, 30991.0, 26785.0, 21004.0, 16935.0, 13087.0, 8994.0, 7248.0, 5241.0, 3725.0, 2493.0, 1575.0, 1151.0, 761.0, 534.0, 360.0, 201.0, 117.0, 87.0, 54.0, 40.0, 12.0, 6.0, 12.0, 5.0, 3.0, 1.0, 3.0], "bins": [-0.049072265625, -0.04750823974609375, -0.0459442138671875, -0.04438018798828125, -0.042816162109375, -0.04125213623046875, -0.0396881103515625, -0.03812408447265625, -0.03656005859375, -0.03499603271484375, -0.0334320068359375, -0.03186798095703125, -0.030303955078125, -0.02873992919921875, -0.0271759033203125, -0.02561187744140625, -0.0240478515625, -0.02248382568359375, -0.0209197998046875, -0.01935577392578125, -0.017791748046875, -0.01622772216796875, -0.0146636962890625, -0.01309967041015625, -0.01153564453125, -0.00997161865234375, -0.0084075927734375, -0.00684356689453125, -0.005279541015625, -0.00371551513671875, -0.0021514892578125, -0.00058746337890625, 0.0009765625, 0.00254058837890625, 0.0041046142578125, 0.00566864013671875, 0.007232666015625, 0.00879669189453125, 0.0103607177734375, 0.01192474365234375, 0.01348876953125, 0.01505279541015625, 0.0166168212890625, 0.01818084716796875, 0.019744873046875, 0.02130889892578125, 0.0228729248046875, 0.02443695068359375, 0.0260009765625, 0.02756500244140625, 0.0291290283203125, 0.03069305419921875, 0.032257080078125, 0.03382110595703125, 0.0353851318359375, 0.03694915771484375, 0.03851318359375, 0.04007720947265625, 0.0416412353515625, 0.04320526123046875, 0.044769287109375, 0.04633331298828125, 0.0478973388671875, 0.04946136474609375, 0.051025390625]}, "parameters/_orig_mod.model.layers.13._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 4.0, 0.0, 7.0, 11.0, 18.0, 39.0, 70.0, 146.0, 258.0, 482.0, 830.0, 1565.0, 2923.0, 5081.0, 8719.0, 15948.0, 30784.0, 52094.0, 88206.0, 146645.0, 243893.0, 413243.0, 664663.0, 1049823.0, 1620415.0, 2322253.0, 2443077.0, 2389305.0, 1931732.0, 1275619.0, 807339.0, 511626.0, 312187.0, 179986.0, 107404.0, 65153.0, 38091.0, 21083.0, 11630.0, 6588.0, 3957.0, 2001.0, 1082.0, 568.0, 292.0, 174.0, 79.0, 51.0, 29.0, 19.0, 9.0, 4.0, 2.0, 0.0, 1.0, 1.0], "bins": [-0.1875, -0.182098388671875, -0.17669677734375, -0.171295166015625, -0.1658935546875, -0.160491943359375, -0.15509033203125, -0.149688720703125, -0.144287109375, -0.138885498046875, -0.13348388671875, -0.128082275390625, -0.1226806640625, -0.117279052734375, -0.11187744140625, -0.106475830078125, -0.10107421875, -0.095672607421875, -0.09027099609375, -0.084869384765625, -0.0794677734375, -0.074066162109375, -0.06866455078125, -0.063262939453125, -0.057861328125, -0.052459716796875, -0.04705810546875, -0.041656494140625, -0.0362548828125, -0.030853271484375, -0.02545166015625, -0.020050048828125, -0.0146484375, -0.009246826171875, -0.00384521484375, 0.001556396484375, 0.0069580078125, 0.012359619140625, 0.01776123046875, 0.023162841796875, 0.028564453125, 0.033966064453125, 0.03936767578125, 0.044769287109375, 0.0501708984375, 0.055572509765625, 0.06097412109375, 0.066375732421875, 0.07177734375, 0.077178955078125, 0.08258056640625, 0.087982177734375, 0.0933837890625, 0.098785400390625, 0.10418701171875, 0.109588623046875, 0.114990234375, 0.120391845703125, 0.12579345703125, 0.131195068359375, 0.1365966796875, 0.141998291015625, 0.14739990234375, 0.152801513671875, 0.158203125]}, "parameters/_orig_mod.model.layers.13._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 3.0, 2.0, 1.0, 2.0, 4.0, 9.0, 8.0, 12.0, 16.0, 21.0, 32.0, 48.0, 77.0, 112.0, 191.0, 323.0, 767.0, 1696.0, 3499.0, 7663.0, 15164.0, 31183.0, 61676.0, 112224.0, 198737.0, 350900.0, 551987.0, 751041.0, 752536.0, 558458.0, 355754.0, 197828.0, 117094.0, 63891.0, 30381.0, 16335.0, 7652.0, 3714.0, 1598.0, 758.0, 354.0, 203.0, 114.0, 67.0, 43.0, 32.0, 27.0, 20.0, 10.0, 8.0, 8.0, 10.0, 2.0, 2.0, 0.0, 5.0], "bins": [-0.376953125, -0.366485595703125, -0.35601806640625, -0.345550537109375, -0.3350830078125, -0.324615478515625, -0.31414794921875, -0.303680419921875, -0.293212890625, -0.282745361328125, -0.27227783203125, -0.261810302734375, -0.2513427734375, -0.240875244140625, -0.23040771484375, -0.219940185546875, -0.20947265625, -0.199005126953125, -0.18853759765625, -0.178070068359375, -0.1676025390625, -0.157135009765625, -0.14666748046875, -0.136199951171875, -0.125732421875, -0.115264892578125, -0.10479736328125, -0.094329833984375, -0.0838623046875, -0.073394775390625, -0.06292724609375, -0.052459716796875, -0.0419921875, -0.031524658203125, -0.02105712890625, -0.010589599609375, -0.0001220703125, 0.010345458984375, 0.02081298828125, 0.031280517578125, 0.041748046875, 0.052215576171875, 0.06268310546875, 0.073150634765625, 0.0836181640625, 0.094085693359375, 0.10455322265625, 0.115020751953125, 0.12548828125, 0.135955810546875, 0.14642333984375, 0.156890869140625, 0.1673583984375, 0.177825927734375, 0.18829345703125, 0.198760986328125, 0.209228515625, 0.219696044921875, 0.23016357421875, 0.240631103515625, 0.2510986328125, 0.261566162109375, 0.27203369140625, 0.282501220703125, 0.29296875]}, "parameters/_orig_mod.model.layers.13._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 1.0, 1.0, 17.0, 13.0, 31.0, 83.0, 202.0, 430.0, 961.0, 1918.0, 3679.0, 6047.0, 11476.0, 17910.0, 27265.0, 41259.0, 56544.0, 70661.0, 79346.0, 88195.0, 88980.0, 81088.0, 75249.0, 59633.0, 47770.0, 30760.0, 21084.0, 12530.0, 7794.0, 4032.0, 2385.0, 1241.0, 571.0, 294.0, 132.0, 50.0, 22.0, 11.0, 6.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0712890625, -0.06937026977539062, -0.06745147705078125, -0.06553268432617188, -0.0636138916015625, -0.061695098876953125, -0.05977630615234375, -0.057857513427734375, -0.055938720703125, -0.054019927978515625, -0.05210113525390625, -0.050182342529296875, -0.0482635498046875, -0.046344757080078125, -0.04442596435546875, -0.042507171630859375, -0.04058837890625, -0.038669586181640625, -0.03675079345703125, -0.034832000732421875, -0.0329132080078125, -0.030994415283203125, -0.02907562255859375, -0.027156829833984375, -0.025238037109375, -0.023319244384765625, -0.02140045166015625, -0.019481658935546875, -0.0175628662109375, -0.015644073486328125, -0.01372528076171875, -0.011806488037109375, -0.0098876953125, -0.007968902587890625, -0.00605010986328125, -0.004131317138671875, -0.0022125244140625, -0.000293731689453125, 0.00162506103515625, 0.003543853759765625, 0.005462646484375, 0.007381439208984375, 0.00930023193359375, 0.011219024658203125, 0.0131378173828125, 0.015056610107421875, 0.01697540283203125, 0.018894195556640625, 0.02081298828125, 0.022731781005859375, 0.02465057373046875, 0.026569366455078125, 0.0284881591796875, 0.030406951904296875, 0.03232574462890625, 0.034244537353515625, 0.036163330078125, 0.038082122802734375, 0.04000091552734375, 0.041919708251953125, 0.0438385009765625, 0.045757293701171875, 0.04767608642578125, 0.049594879150390625, 0.051513671875]}, "parameters/_orig_mod.model.layers.14._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 2.0, 0.0, 2.0, 5.0, 2.0, 4.0, 7.0, 12.0, 11.0, 19.0, 28.0, 34.0, 48.0, 56.0, 87.0, 83.0, 136.0, 172.0, 271.0, 446.0, 913.0, 2602.0, 8272.0, 25042.0, 77767.0, 217612.0, 613843.0, 1490538.0, 3224531.0, 4486527.0, 3676066.0, 1784537.0, 749263.0, 271020.0, 99359.0, 31813.0, 9950.0, 3442.0, 1202.0, 500.0, 239.0, 171.0, 132.0, 86.0, 95.0, 49.0, 64.0, 34.0, 23.0, 22.0, 20.0, 16.0, 12.0, 9.0, 12.0, 0.0, 0.0, 2.0, 1.0, 1.0, 1.0, 1.0], "bins": [-0.310546875, -0.30072021484375, -0.2908935546875, -0.28106689453125, -0.271240234375, -0.26141357421875, -0.2515869140625, -0.24176025390625, -0.23193359375, -0.22210693359375, -0.2122802734375, -0.20245361328125, -0.192626953125, -0.18280029296875, -0.1729736328125, -0.16314697265625, -0.1533203125, -0.14349365234375, -0.1336669921875, -0.12384033203125, -0.114013671875, -0.10418701171875, -0.0943603515625, -0.08453369140625, -0.07470703125, -0.06488037109375, -0.0550537109375, -0.04522705078125, -0.035400390625, -0.02557373046875, -0.0157470703125, -0.00592041015625, 0.00390625, 0.01373291015625, 0.0235595703125, 0.03338623046875, 0.043212890625, 0.05303955078125, 0.0628662109375, 0.07269287109375, 0.08251953125, 0.09234619140625, 0.1021728515625, 0.11199951171875, 0.121826171875, 0.13165283203125, 0.1414794921875, 0.15130615234375, 0.1611328125, 0.17095947265625, 0.1807861328125, 0.19061279296875, 0.200439453125, 0.21026611328125, 0.2200927734375, 0.22991943359375, 0.23974609375, 0.24957275390625, 0.2593994140625, 0.26922607421875, 0.279052734375, 0.28887939453125, 0.2987060546875, 0.30853271484375, 0.318359375]}, "parameters/_orig_mod.model.layers.14._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 3.0, 5.0, 6.0, 12.0, 6.0, 12.0, 24.0, 22.0, 34.0, 38.0, 50.0, 73.0, 76.0, 120.0, 196.0, 253.0, 448.0, 789.0, 1481.0, 3277.0, 6796.0, 13039.0, 26618.0, 55964.0, 110011.0, 208941.0, 390988.0, 629169.0, 818197.0, 770721.0, 527496.0, 302141.0, 160364.0, 85754.0, 40296.0, 21027.0, 9932.0, 4583.0, 2450.0, 1179.0, 561.0, 334.0, 225.0, 145.0, 103.0, 78.0, 59.0, 48.0, 37.0, 28.0, 23.0, 20.0, 11.0, 7.0, 11.0, 9.0, 4.0, 3.0, 0.0, 4.0], "bins": [-0.36328125, -0.352203369140625, -0.34112548828125, -0.330047607421875, -0.3189697265625, -0.307891845703125, -0.29681396484375, -0.285736083984375, -0.274658203125, -0.263580322265625, -0.25250244140625, -0.241424560546875, -0.2303466796875, -0.219268798828125, -0.20819091796875, -0.197113037109375, -0.18603515625, -0.174957275390625, -0.16387939453125, -0.152801513671875, -0.1417236328125, -0.130645751953125, -0.11956787109375, -0.108489990234375, -0.097412109375, -0.086334228515625, -0.07525634765625, -0.064178466796875, -0.0531005859375, -0.042022705078125, -0.03094482421875, -0.019866943359375, -0.0087890625, 0.002288818359375, 0.01336669921875, 0.024444580078125, 0.0355224609375, 0.046600341796875, 0.05767822265625, 0.068756103515625, 0.079833984375, 0.090911865234375, 0.10198974609375, 0.113067626953125, 0.1241455078125, 0.135223388671875, 0.14630126953125, 0.157379150390625, 0.16845703125, 0.179534912109375, 0.19061279296875, 0.201690673828125, 0.2127685546875, 0.223846435546875, 0.23492431640625, 0.246002197265625, 0.257080078125, 0.268157958984375, 0.27923583984375, 0.290313720703125, 0.3013916015625, 0.312469482421875, 0.32354736328125, 0.334625244140625, 0.345703125]}, "parameters/_orig_mod.model.layers.14._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 6.0, 3.0, 7.0, 23.0, 25.0, 51.0, 97.0, 147.0, 200.0, 357.0, 535.0, 867.0, 1129.0, 1821.0, 2640.0, 3692.0, 5072.0, 7058.0, 9373.0, 12307.0, 15767.0, 20348.0, 25050.0, 30081.0, 38132.0, 39868.0, 47379.0, 51798.0, 53871.0, 57011.0, 56470.0, 53331.0, 50807.0, 46227.0, 41950.0, 34638.0, 28200.0, 24435.0, 19717.0, 15718.0, 11798.0, 9085.0, 6654.0, 5034.0, 3266.0, 2503.0, 1777.0, 1188.0, 811.0, 488.0, 351.0, 183.0, 137.0, 88.0, 52.0, 25.0, 16.0, 4.0, 4.0, 1.0, 5.0], "bins": [-0.052001953125, -0.05042266845703125, -0.0488433837890625, -0.04726409912109375, -0.045684814453125, -0.04410552978515625, -0.0425262451171875, -0.04094696044921875, -0.03936767578125, -0.03778839111328125, -0.0362091064453125, -0.03462982177734375, -0.033050537109375, -0.03147125244140625, -0.0298919677734375, -0.02831268310546875, -0.0267333984375, -0.02515411376953125, -0.0235748291015625, -0.02199554443359375, -0.020416259765625, -0.01883697509765625, -0.0172576904296875, -0.01567840576171875, -0.01409912109375, -0.01251983642578125, -0.0109405517578125, -0.00936126708984375, -0.007781982421875, -0.00620269775390625, -0.0046234130859375, -0.00304412841796875, -0.00146484375, 0.00011444091796875, 0.0016937255859375, 0.00327301025390625, 0.004852294921875, 0.00643157958984375, 0.0080108642578125, 0.00959014892578125, 0.01116943359375, 0.01274871826171875, 0.0143280029296875, 0.01590728759765625, 0.017486572265625, 0.01906585693359375, 0.0206451416015625, 0.02222442626953125, 0.0238037109375, 0.02538299560546875, 0.0269622802734375, 0.02854156494140625, 0.030120849609375, 0.03170013427734375, 0.0332794189453125, 0.03485870361328125, 0.03643798828125, 0.03801727294921875, 0.0395965576171875, 0.04117584228515625, 0.042755126953125, 0.04433441162109375, 0.0459136962890625, 0.04749298095703125, 0.049072265625]}, "parameters/_orig_mod.model.layers.15._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 2.0, 6.0, 5.0, 5.0, 7.0, 7.0, 14.0, 23.0, 23.0, 44.0, 51.0, 109.0, 167.0, 334.0, 810.0, 1907.0, 5840.0, 15509.0, 45712.0, 140350.0, 396404.0, 1171147.0, 3093209.0, 5534664.0, 3968399.0, 1566856.0, 548064.0, 190767.0, 63105.0, 21814.0, 7370.0, 2574.0, 1010.0, 397.0, 190.0, 91.0, 62.0, 37.0, 31.0, 21.0, 14.0, 13.0, 12.0, 10.0, 5.0, 2.0, 6.0, 1.0, 0.0, 4.0, 2.0, 2.0, 0.0, 2.0], "bins": [-0.419921875, -0.407440185546875, -0.39495849609375, -0.382476806640625, -0.3699951171875, -0.357513427734375, -0.34503173828125, -0.332550048828125, -0.320068359375, -0.307586669921875, -0.29510498046875, -0.282623291015625, -0.2701416015625, -0.257659912109375, -0.24517822265625, -0.232696533203125, -0.22021484375, -0.207733154296875, -0.19525146484375, -0.182769775390625, -0.1702880859375, -0.157806396484375, -0.14532470703125, -0.132843017578125, -0.120361328125, -0.107879638671875, -0.09539794921875, -0.082916259765625, -0.0704345703125, -0.057952880859375, -0.04547119140625, -0.032989501953125, -0.0205078125, -0.008026123046875, 0.00445556640625, 0.016937255859375, 0.0294189453125, 0.041900634765625, 0.05438232421875, 0.066864013671875, 0.079345703125, 0.091827392578125, 0.10430908203125, 0.116790771484375, 0.1292724609375, 0.141754150390625, 0.15423583984375, 0.166717529296875, 0.17919921875, 0.191680908203125, 0.20416259765625, 0.216644287109375, 0.2291259765625, 0.241607666015625, 0.25408935546875, 0.266571044921875, 0.279052734375, 0.291534423828125, 0.30401611328125, 0.316497802734375, 0.3289794921875, 0.341461181640625, 0.35394287109375, 0.366424560546875, 0.37890625]}, "parameters/_orig_mod.model.layers.15._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 8.0, 3.0, 3.0, 11.0, 7.0, 19.0, 19.0, 33.0, 61.0, 94.0, 177.0, 231.0, 502.0, 1106.0, 2699.0, 6381.0, 14133.0, 32964.0, 71010.0, 147516.0, 273298.0, 477389.0, 708680.0, 783909.0, 686035.0, 461596.0, 263070.0, 140673.0, 67958.0, 30570.0, 13630.0, 5895.0, 2442.0, 1080.0, 490.0, 244.0, 141.0, 79.0, 45.0, 32.0, 21.0, 17.0, 9.0, 8.0, 6.0, 0.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.3828125, -0.3717041015625, -0.360595703125, -0.3494873046875, -0.33837890625, -0.3272705078125, -0.316162109375, -0.3050537109375, -0.2939453125, -0.2828369140625, -0.271728515625, -0.2606201171875, -0.24951171875, -0.2384033203125, -0.227294921875, -0.2161865234375, -0.205078125, -0.1939697265625, -0.182861328125, -0.1717529296875, -0.16064453125, -0.1495361328125, -0.138427734375, -0.1273193359375, -0.1162109375, -0.1051025390625, -0.093994140625, -0.0828857421875, -0.07177734375, -0.0606689453125, -0.049560546875, -0.0384521484375, -0.02734375, -0.0162353515625, -0.005126953125, 0.0059814453125, 0.01708984375, 0.0281982421875, 0.039306640625, 0.0504150390625, 0.0615234375, 0.0726318359375, 0.083740234375, 0.0948486328125, 0.10595703125, 0.1170654296875, 0.128173828125, 0.1392822265625, 0.150390625, 0.1614990234375, 0.172607421875, 0.1837158203125, 0.19482421875, 0.2059326171875, 0.217041015625, 0.2281494140625, 0.2392578125, 0.2503662109375, 0.261474609375, 0.2725830078125, 0.28369140625, 0.2947998046875, 0.305908203125, 0.3170166015625, 0.328125]}, "parameters/_orig_mod.model.layers.15._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 1.0, 0.0, 3.0, 1.0, 5.0, 4.0, 10.0, 14.0, 24.0, 46.0, 41.0, 94.0, 126.0, 226.0, 415.0, 628.0, 1239.0, 1994.0, 3196.0, 5550.0, 8673.0, 13824.0, 20691.0, 27950.0, 42216.0, 50770.0, 63702.0, 67854.0, 75054.0, 77992.0, 74321.0, 68855.0, 61949.0, 49210.0, 40848.0, 27864.0, 19381.0, 13439.0, 8509.0, 5318.0, 3004.0, 1973.0, 1084.0, 641.0, 344.0, 242.0, 119.0, 61.0, 62.0, 33.0, 26.0, 23.0, 10.0, 9.0, 2.0, 1.0, 1.0, 1.0, 2.0, 1.0, 2.0], "bins": [-0.052734375, -0.05110931396484375, -0.0494842529296875, -0.04785919189453125, -0.046234130859375, -0.04460906982421875, -0.0429840087890625, -0.04135894775390625, -0.03973388671875, -0.03810882568359375, -0.0364837646484375, -0.03485870361328125, -0.033233642578125, -0.03160858154296875, -0.0299835205078125, -0.02835845947265625, -0.0267333984375, -0.02510833740234375, -0.0234832763671875, -0.02185821533203125, -0.020233154296875, -0.01860809326171875, -0.0169830322265625, -0.01535797119140625, -0.01373291015625, -0.01210784912109375, -0.0104827880859375, -0.00885772705078125, -0.007232666015625, -0.00560760498046875, -0.0039825439453125, -0.00235748291015625, -0.000732421875, 0.00089263916015625, 0.0025177001953125, 0.00414276123046875, 0.005767822265625, 0.00739288330078125, 0.0090179443359375, 0.01064300537109375, 0.01226806640625, 0.01389312744140625, 0.0155181884765625, 0.01714324951171875, 0.018768310546875, 0.02039337158203125, 0.0220184326171875, 0.02364349365234375, 0.0252685546875, 0.02689361572265625, 0.0285186767578125, 0.03014373779296875, 0.031768798828125, 0.03339385986328125, 0.0350189208984375, 0.03664398193359375, 0.03826904296875, 0.03989410400390625, 0.0415191650390625, 0.04314422607421875, 0.044769287109375, 0.04639434814453125, 0.0480194091796875, 0.04964447021484375, 0.05126953125]}, "parameters/_orig_mod.model.layers.16._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 1.0, 2.0, 1.0, 6.0, 6.0, 7.0, 3.0, 15.0, 18.0, 12.0, 21.0, 24.0, 28.0, 50.0, 47.0, 73.0, 95.0, 124.0, 280.0, 608.0, 1845.0, 6233.0, 20414.0, 73090.0, 237961.0, 750117.0, 2028471.0, 4340242.0, 4963261.0, 2728844.0, 1101447.0, 364387.0, 112832.0, 32675.0, 9337.0, 2752.0, 898.0, 383.0, 169.0, 106.0, 85.0, 54.0, 55.0, 27.0, 28.0, 23.0, 10.0, 10.0, 6.0, 8.0, 8.0, 2.0, 2.0, 5.0, 0.0, 0.0, 1.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.369140625, -0.3572998046875, -0.345458984375, -0.3336181640625, -0.32177734375, -0.3099365234375, -0.298095703125, -0.2862548828125, -0.2744140625, -0.2625732421875, -0.250732421875, -0.2388916015625, -0.22705078125, -0.2152099609375, -0.203369140625, -0.1915283203125, -0.1796875, -0.1678466796875, -0.156005859375, -0.1441650390625, -0.13232421875, -0.1204833984375, -0.108642578125, -0.0968017578125, -0.0849609375, -0.0731201171875, -0.061279296875, -0.0494384765625, -0.03759765625, -0.0257568359375, -0.013916015625, -0.0020751953125, 0.009765625, 0.0216064453125, 0.033447265625, 0.0452880859375, 0.05712890625, 0.0689697265625, 0.080810546875, 0.0926513671875, 0.1044921875, 0.1163330078125, 0.128173828125, 0.1400146484375, 0.15185546875, 0.1636962890625, 0.175537109375, 0.1873779296875, 0.19921875, 0.2110595703125, 0.222900390625, 0.2347412109375, 0.24658203125, 0.2584228515625, 0.270263671875, 0.2821044921875, 0.2939453125, 0.3057861328125, 0.317626953125, 0.3294677734375, 0.34130859375, 0.3531494140625, 0.364990234375, 0.3768310546875, 0.388671875]}, "parameters/_orig_mod.model.layers.16._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 1.0, 2.0, 4.0, 2.0, 2.0, 4.0, 9.0, 9.0, 9.0, 17.0, 23.0, 31.0, 38.0, 66.0, 118.0, 197.0, 358.0, 645.0, 1460.0, 2860.0, 6025.0, 11784.0, 24649.0, 52174.0, 103844.0, 191537.0, 357078.0, 553732.0, 734299.0, 741110.0, 588225.0, 381730.0, 213773.0, 114887.0, 60227.0, 26471.0, 13697.0, 6535.0, 3300.0, 1610.0, 757.0, 375.0, 220.0, 144.0, 76.0, 46.0, 40.0, 32.0, 21.0, 8.0, 9.0, 8.0, 4.0, 7.0, 3.0, 4.0, 0.0, 0.0, 2.0, 0.0, 1.0, 2.0], "bins": [-0.328125, -0.31756591796875, -0.3070068359375, -0.29644775390625, -0.285888671875, -0.27532958984375, -0.2647705078125, -0.25421142578125, -0.24365234375, -0.23309326171875, -0.2225341796875, -0.21197509765625, -0.201416015625, -0.19085693359375, -0.1802978515625, -0.16973876953125, -0.1591796875, -0.14862060546875, -0.1380615234375, -0.12750244140625, -0.116943359375, -0.10638427734375, -0.0958251953125, -0.08526611328125, -0.07470703125, -0.06414794921875, -0.0535888671875, -0.04302978515625, -0.032470703125, -0.02191162109375, -0.0113525390625, -0.00079345703125, 0.009765625, 0.02032470703125, 0.0308837890625, 0.04144287109375, 0.052001953125, 0.06256103515625, 0.0731201171875, 0.08367919921875, 0.09423828125, 0.10479736328125, 0.1153564453125, 0.12591552734375, 0.136474609375, 0.14703369140625, 0.1575927734375, 0.16815185546875, 0.1787109375, 0.18927001953125, 0.1998291015625, 0.21038818359375, 0.220947265625, 0.23150634765625, 0.2420654296875, 0.25262451171875, 0.26318359375, 0.27374267578125, 0.2843017578125, 0.29486083984375, 0.305419921875, 0.31597900390625, 0.3265380859375, 0.33709716796875, 0.34765625]}, "parameters/_orig_mod.model.layers.16._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 2.0, 1.0, 3.0, 6.0, 10.0, 12.0, 12.0, 20.0, 30.0, 41.0, 48.0, 46.0, 80.0, 113.0, 191.0, 306.0, 458.0, 839.0, 1471.0, 2979.0, 5349.0, 9986.0, 17511.0, 29440.0, 45931.0, 69861.0, 91966.0, 98280.0, 105511.0, 98903.0, 87520.0, 66749.0, 43050.0, 27275.0, 15806.0, 9067.0, 4692.0, 2687.0, 1399.0, 730.0, 451.0, 254.0, 177.0, 101.0, 106.0, 53.0, 40.0, 30.0, 21.0, 24.0, 5.0, 10.0, 5.0, 2.0, 6.0, 5.0, 1.0, 1.0, 0.0, 1.0, 1.0, 2.0], "bins": [-0.080078125, -0.07744598388671875, -0.0748138427734375, -0.07218170166015625, -0.069549560546875, -0.06691741943359375, -0.0642852783203125, -0.06165313720703125, -0.05902099609375, -0.05638885498046875, -0.0537567138671875, -0.05112457275390625, -0.048492431640625, -0.04586029052734375, -0.0432281494140625, -0.04059600830078125, -0.0379638671875, -0.03533172607421875, -0.0326995849609375, -0.03006744384765625, -0.027435302734375, -0.02480316162109375, -0.0221710205078125, -0.01953887939453125, -0.01690673828125, -0.01427459716796875, -0.0116424560546875, -0.00901031494140625, -0.006378173828125, -0.00374603271484375, -0.0011138916015625, 0.00151824951171875, 0.004150390625, 0.00678253173828125, 0.0094146728515625, 0.01204681396484375, 0.014678955078125, 0.01731109619140625, 0.0199432373046875, 0.02257537841796875, 0.02520751953125, 0.02783966064453125, 0.0304718017578125, 0.03310394287109375, 0.035736083984375, 0.03836822509765625, 0.0410003662109375, 0.04363250732421875, 0.0462646484375, 0.04889678955078125, 0.0515289306640625, 0.05416107177734375, 0.056793212890625, 0.05942535400390625, 0.0620574951171875, 0.06468963623046875, 0.06732177734375, 0.06995391845703125, 0.0725860595703125, 0.07521820068359375, 0.077850341796875, 0.08048248291015625, 0.0831146240234375, 0.08574676513671875, 0.08837890625]}, "parameters/_orig_mod.model.layers.17._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 3.0, 1.0, 8.0, 3.0, 19.0, 21.0, 37.0, 49.0, 72.0, 122.0, 200.0, 418.0, 662.0, 1279.0, 2469.0, 4538.0, 7638.0, 15546.0, 28730.0, 56735.0, 100368.0, 184910.0, 312398.0, 559856.0, 915082.0, 1472068.0, 2234206.0, 2763724.0, 2680159.0, 2099258.0, 1338254.0, 841116.0, 506483.0, 290766.0, 160732.0, 93703.0, 49905.0, 26430.0, 13444.0, 7344.0, 3769.0, 2144.0, 1101.0, 580.0, 342.0, 190.0, 142.0, 73.0, 33.0, 21.0, 23.0, 11.0, 6.0, 11.0, 2.0, 2.0, 0.0, 1.0, 3.0, 1.0, 1.0], "bins": [-0.1923828125, -0.1861572265625, -0.179931640625, -0.1737060546875, -0.16748046875, -0.1612548828125, -0.155029296875, -0.1488037109375, -0.142578125, -0.1363525390625, -0.130126953125, -0.1239013671875, -0.11767578125, -0.1114501953125, -0.105224609375, -0.0989990234375, -0.0927734375, -0.0865478515625, -0.080322265625, -0.0740966796875, -0.06787109375, -0.0616455078125, -0.055419921875, -0.0491943359375, -0.04296875, -0.0367431640625, -0.030517578125, -0.0242919921875, -0.01806640625, -0.0118408203125, -0.005615234375, 0.0006103515625, 0.0068359375, 0.0130615234375, 0.019287109375, 0.0255126953125, 0.03173828125, 0.0379638671875, 0.044189453125, 0.0504150390625, 0.056640625, 0.0628662109375, 0.069091796875, 0.0753173828125, 0.08154296875, 0.0877685546875, 0.093994140625, 0.1002197265625, 0.1064453125, 0.1126708984375, 0.118896484375, 0.1251220703125, 0.13134765625, 0.1375732421875, 0.143798828125, 0.1500244140625, 0.15625, 0.1624755859375, 0.168701171875, 0.1749267578125, 0.18115234375, 0.1873779296875, 0.193603515625, 0.1998291015625, 0.2060546875]}, "parameters/_orig_mod.model.layers.17._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 0.0, 3.0, 1.0, 6.0, 10.0, 3.0, 10.0, 17.0, 13.0, 15.0, 20.0, 40.0, 66.0, 68.0, 97.0, 180.0, 288.0, 508.0, 1082.0, 2311.0, 5040.0, 11803.0, 23735.0, 52696.0, 109253.0, 197832.0, 366159.0, 575774.0, 759900.0, 755009.0, 570814.0, 362832.0, 197846.0, 104386.0, 52543.0, 23054.0, 11165.0, 5223.0, 2286.0, 983.0, 524.0, 239.0, 154.0, 83.0, 59.0, 40.0, 41.0, 24.0, 13.0, 9.0, 11.0, 14.0, 4.0, 6.0, 2.0, 3.0, 0.0, 2.0, 0.0, 2.0], "bins": [-0.359375, -0.348480224609375, -0.33758544921875, -0.326690673828125, -0.3157958984375, -0.304901123046875, -0.29400634765625, -0.283111572265625, -0.272216796875, -0.261322021484375, -0.25042724609375, -0.239532470703125, -0.2286376953125, -0.217742919921875, -0.20684814453125, -0.195953369140625, -0.18505859375, -0.174163818359375, -0.16326904296875, -0.152374267578125, -0.1414794921875, -0.130584716796875, -0.11968994140625, -0.108795166015625, -0.097900390625, -0.087005615234375, -0.07611083984375, -0.065216064453125, -0.0543212890625, -0.043426513671875, -0.03253173828125, -0.021636962890625, -0.0107421875, 0.000152587890625, 0.01104736328125, 0.021942138671875, 0.0328369140625, 0.043731689453125, 0.05462646484375, 0.065521240234375, 0.076416015625, 0.087310791015625, 0.09820556640625, 0.109100341796875, 0.1199951171875, 0.130889892578125, 0.14178466796875, 0.152679443359375, 0.16357421875, 0.174468994140625, 0.18536376953125, 0.196258544921875, 0.2071533203125, 0.218048095703125, 0.22894287109375, 0.239837646484375, 0.250732421875, 0.261627197265625, 0.27252197265625, 0.283416748046875, 0.2943115234375, 0.305206298828125, 0.31610107421875, 0.326995849609375, 0.337890625]}, "parameters/_orig_mod.model.layers.17._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 2.0, 6.0, 7.0, 29.0, 34.0, 62.0, 75.0, 177.0, 243.0, 444.0, 601.0, 1029.0, 1405.0, 2318.0, 2860.0, 4460.0, 5601.0, 8102.0, 10176.0, 13445.0, 17005.0, 20796.0, 25176.0, 29734.0, 38004.0, 40106.0, 45974.0, 48888.0, 51105.0, 54268.0, 53336.0, 51167.0, 48317.0, 45030.0, 39366.0, 37567.0, 29593.0, 25071.0, 20906.0, 17132.0, 13057.0, 10409.0, 7404.0, 5908.0, 3936.0, 3068.0, 2017.0, 1561.0, 941.0, 706.0, 386.0, 285.0, 147.0, 87.0, 60.0, 38.0, 28.0, 10.0, 5.0, 4.0, 0.0, 1.0, 2.0], "bins": [-0.04345703125, -0.04205322265625, -0.0406494140625, -0.03924560546875, -0.037841796875, -0.03643798828125, -0.0350341796875, -0.03363037109375, -0.0322265625, -0.03082275390625, -0.0294189453125, -0.02801513671875, -0.026611328125, -0.02520751953125, -0.0238037109375, -0.02239990234375, -0.02099609375, -0.01959228515625, -0.0181884765625, -0.01678466796875, -0.015380859375, -0.01397705078125, -0.0125732421875, -0.01116943359375, -0.009765625, -0.00836181640625, -0.0069580078125, -0.00555419921875, -0.004150390625, -0.00274658203125, -0.0013427734375, 6.103515625e-05, 0.00146484375, 0.00286865234375, 0.0042724609375, 0.00567626953125, 0.007080078125, 0.00848388671875, 0.0098876953125, 0.01129150390625, 0.0126953125, 0.01409912109375, 0.0155029296875, 0.01690673828125, 0.018310546875, 0.01971435546875, 0.0211181640625, 0.02252197265625, 0.02392578125, 0.02532958984375, 0.0267333984375, 0.02813720703125, 0.029541015625, 0.03094482421875, 0.0323486328125, 0.03375244140625, 0.03515625, 0.03656005859375, 0.0379638671875, 0.03936767578125, 0.040771484375, 0.04217529296875, 0.0435791015625, 0.04498291015625, 0.04638671875]}, "parameters/_orig_mod.model.layers.18._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 1.0, 1.0, 2.0, 1.0, 4.0, 8.0, 9.0, 12.0, 14.0, 7.0, 29.0, 44.0, 46.0, 81.0, 145.0, 230.0, 407.0, 863.0, 1807.0, 3332.0, 7310.0, 15288.0, 30667.0, 64726.0, 118175.0, 218328.0, 409942.0, 729720.0, 1138160.0, 1808879.0, 2482428.0, 2597873.0, 2496166.0, 1832741.0, 1183309.0, 742777.0, 408762.0, 235629.0, 121139.0, 66327.0, 31517.0, 15127.0, 7680.0, 3824.0, 1694.0, 963.0, 451.0, 222.0, 128.0, 66.0, 46.0, 32.0, 24.0, 12.0, 12.0, 5.0, 5.0, 8.0, 5.0, 1.0, 3.0], "bins": [-0.2138671875, -0.207672119140625, -0.20147705078125, -0.195281982421875, -0.1890869140625, -0.182891845703125, -0.17669677734375, -0.170501708984375, -0.164306640625, -0.158111572265625, -0.15191650390625, -0.145721435546875, -0.1395263671875, -0.133331298828125, -0.12713623046875, -0.120941162109375, -0.11474609375, -0.108551025390625, -0.10235595703125, -0.096160888671875, -0.0899658203125, -0.083770751953125, -0.07757568359375, -0.071380615234375, -0.065185546875, -0.058990478515625, -0.05279541015625, -0.046600341796875, -0.0404052734375, -0.034210205078125, -0.02801513671875, -0.021820068359375, -0.015625, -0.009429931640625, -0.00323486328125, 0.002960205078125, 0.0091552734375, 0.015350341796875, 0.02154541015625, 0.027740478515625, 0.033935546875, 0.040130615234375, 0.04632568359375, 0.052520751953125, 0.0587158203125, 0.064910888671875, 0.07110595703125, 0.077301025390625, 0.08349609375, 0.089691162109375, 0.09588623046875, 0.102081298828125, 0.1082763671875, 0.114471435546875, 0.12066650390625, 0.126861572265625, 0.133056640625, 0.139251708984375, 0.14544677734375, 0.151641845703125, 0.1578369140625, 0.164031982421875, 0.17022705078125, 0.176422119140625, 0.1826171875]}, "parameters/_orig_mod.model.layers.18._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [3.0, 5.0, 2.0, 3.0, 4.0, 7.0, 7.0, 10.0, 15.0, 13.0, 30.0, 42.0, 46.0, 49.0, 97.0, 148.0, 221.0, 355.0, 624.0, 1120.0, 2112.0, 4219.0, 7779.0, 15086.0, 27006.0, 52409.0, 90723.0, 146767.0, 235243.0, 354763.0, 491475.0, 601461.0, 602669.0, 528810.0, 379925.0, 260631.0, 162165.0, 100831.0, 58779.0, 32237.0, 17202.0, 9101.0, 4778.0, 2308.0, 1286.0, 614.0, 392.0, 232.0, 155.0, 109.0, 75.0, 41.0, 29.0, 27.0, 26.0, 14.0, 7.0, 5.0, 5.0, 2.0, 2.0, 1.0, 1.0, 1.0], "bins": [-0.283203125, -0.274383544921875, -0.26556396484375, -0.256744384765625, -0.2479248046875, -0.239105224609375, -0.23028564453125, -0.221466064453125, -0.212646484375, -0.203826904296875, -0.19500732421875, -0.186187744140625, -0.1773681640625, -0.168548583984375, -0.15972900390625, -0.150909423828125, -0.14208984375, -0.133270263671875, -0.12445068359375, -0.115631103515625, -0.1068115234375, -0.097991943359375, -0.08917236328125, -0.080352783203125, -0.071533203125, -0.062713623046875, -0.05389404296875, -0.045074462890625, -0.0362548828125, -0.027435302734375, -0.01861572265625, -0.009796142578125, -0.0009765625, 0.007843017578125, 0.01666259765625, 0.025482177734375, 0.0343017578125, 0.043121337890625, 0.05194091796875, 0.060760498046875, 0.069580078125, 0.078399658203125, 0.08721923828125, 0.096038818359375, 0.1048583984375, 0.113677978515625, 0.12249755859375, 0.131317138671875, 0.14013671875, 0.148956298828125, 0.15777587890625, 0.166595458984375, 0.1754150390625, 0.184234619140625, 0.19305419921875, 0.201873779296875, 0.210693359375, 0.219512939453125, 0.22833251953125, 0.237152099609375, 0.2459716796875, 0.254791259765625, 0.26361083984375, 0.272430419921875, 0.28125]}, "parameters/_orig_mod.model.layers.18._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 3.0, 0.0, 4.0, 4.0, 5.0, 6.0, 13.0, 27.0, 44.0, 82.0, 115.0, 204.0, 411.0, 669.0, 1396.0, 2412.0, 3743.0, 6389.0, 10351.0, 15920.0, 23510.0, 34476.0, 46412.0, 61520.0, 75928.0, 79197.0, 85457.0, 82491.0, 78557.0, 64047.0, 52674.0, 37958.0, 26852.0, 18807.0, 11444.0, 7616.0, 4677.0, 2740.0, 1508.0, 909.0, 489.0, 241.0, 146.0, 77.0, 55.0, 34.0, 17.0, 15.0, 6.0, 6.0, 3.0, 3.0, 1.0, 3.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.059814453125, -0.057651519775390625, -0.05548858642578125, -0.053325653076171875, -0.0511627197265625, -0.048999786376953125, -0.04683685302734375, -0.044673919677734375, -0.042510986328125, -0.040348052978515625, -0.03818511962890625, -0.036022186279296875, -0.0338592529296875, -0.031696319580078125, -0.02953338623046875, -0.027370452880859375, -0.02520751953125, -0.023044586181640625, -0.02088165283203125, -0.018718719482421875, -0.0165557861328125, -0.014392852783203125, -0.01222991943359375, -0.010066986083984375, -0.007904052734375, -0.005741119384765625, -0.00357818603515625, -0.001415252685546875, 0.0007476806640625, 0.002910614013671875, 0.00507354736328125, 0.007236480712890625, 0.0093994140625, 0.011562347412109375, 0.01372528076171875, 0.015888214111328125, 0.0180511474609375, 0.020214080810546875, 0.02237701416015625, 0.024539947509765625, 0.026702880859375, 0.028865814208984375, 0.03102874755859375, 0.033191680908203125, 0.0353546142578125, 0.037517547607421875, 0.03968048095703125, 0.041843414306640625, 0.04400634765625, 0.046169281005859375, 0.04833221435546875, 0.050495147705078125, 0.0526580810546875, 0.054821014404296875, 0.05698394775390625, 0.059146881103515625, 0.061309814453125, 0.06347274780273438, 0.06563568115234375, 0.06779861450195312, 0.0699615478515625, 0.07212448120117188, 0.07428741455078125, 0.07645034790039062, 0.07861328125]}, "parameters/_orig_mod.model.layers.19._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 3.0, 2.0, 2.0, 7.0, 7.0, 11.0, 13.0, 18.0, 26.0, 55.0, 70.0, 136.0, 298.0, 639.0, 1110.0, 2251.0, 5087.0, 9540.0, 22802.0, 50424.0, 106590.0, 218728.0, 420673.0, 830895.0, 1419234.0, 2324726.0, 2925977.0, 2932168.0, 2338601.0, 1458516.0, 837514.0, 441914.0, 223785.0, 110096.0, 52067.0, 22907.0, 10480.0, 4887.0, 2358.0, 1303.0, 601.0, 326.0, 159.0, 85.0, 47.0, 35.0, 9.0, 12.0, 6.0, 1.0, 3.0, 4.0, 1.0, 2.0, 2.0], "bins": [-0.25390625, -0.246856689453125, -0.23980712890625, -0.232757568359375, -0.2257080078125, -0.218658447265625, -0.21160888671875, -0.204559326171875, -0.197509765625, -0.190460205078125, -0.18341064453125, -0.176361083984375, -0.1693115234375, -0.162261962890625, -0.15521240234375, -0.148162841796875, -0.14111328125, -0.134063720703125, -0.12701416015625, -0.119964599609375, -0.1129150390625, -0.105865478515625, -0.09881591796875, -0.091766357421875, -0.084716796875, -0.077667236328125, -0.07061767578125, -0.063568115234375, -0.0565185546875, -0.049468994140625, -0.04241943359375, -0.035369873046875, -0.0283203125, -0.021270751953125, -0.01422119140625, -0.007171630859375, -0.0001220703125, 0.006927490234375, 0.01397705078125, 0.021026611328125, 0.028076171875, 0.035125732421875, 0.04217529296875, 0.049224853515625, 0.0562744140625, 0.063323974609375, 0.07037353515625, 0.077423095703125, 0.08447265625, 0.091522216796875, 0.09857177734375, 0.105621337890625, 0.1126708984375, 0.119720458984375, 0.12677001953125, 0.133819580078125, 0.140869140625, 0.147918701171875, 0.15496826171875, 0.162017822265625, 0.1690673828125, 0.176116943359375, 0.18316650390625, 0.190216064453125, 0.197265625]}, "parameters/_orig_mod.model.layers.19._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 4.0, 0.0, 4.0, 6.0, 4.0, 14.0, 14.0, 17.0, 22.0, 28.0, 55.0, 66.0, 108.0, 204.0, 336.0, 606.0, 1232.0, 2721.0, 5161.0, 10738.0, 22257.0, 44376.0, 86318.0, 147753.0, 260258.0, 406112.0, 587443.0, 654188.0, 650095.0, 506648.0, 349015.0, 205344.0, 120928.0, 66215.0, 32869.0, 16633.0, 8402.0, 4044.0, 1977.0, 926.0, 446.0, 256.0, 152.0, 106.0, 68.0, 33.0, 36.0, 18.0, 8.0, 13.0, 8.0, 6.0, 1.0, 2.0, 0.0, 3.0, 0.0, 0.0, 1.0, 2.0], "bins": [-0.30078125, -0.29132080078125, -0.2818603515625, -0.27239990234375, -0.262939453125, -0.25347900390625, -0.2440185546875, -0.23455810546875, -0.22509765625, -0.21563720703125, -0.2061767578125, -0.19671630859375, -0.187255859375, -0.17779541015625, -0.1683349609375, -0.15887451171875, -0.1494140625, -0.13995361328125, -0.1304931640625, -0.12103271484375, -0.111572265625, -0.10211181640625, -0.0926513671875, -0.08319091796875, -0.07373046875, -0.06427001953125, -0.0548095703125, -0.04534912109375, -0.035888671875, -0.02642822265625, -0.0169677734375, -0.00750732421875, 0.001953125, 0.01141357421875, 0.0208740234375, 0.03033447265625, 0.039794921875, 0.04925537109375, 0.0587158203125, 0.06817626953125, 0.07763671875, 0.08709716796875, 0.0965576171875, 0.10601806640625, 0.115478515625, 0.12493896484375, 0.1343994140625, 0.14385986328125, 0.1533203125, 0.16278076171875, 0.1722412109375, 0.18170166015625, 0.191162109375, 0.20062255859375, 0.2100830078125, 0.21954345703125, 0.22900390625, 0.23846435546875, 0.2479248046875, 0.25738525390625, 0.266845703125, 0.27630615234375, 0.2857666015625, 0.29522705078125, 0.3046875]}, "parameters/_orig_mod.model.layers.19._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 3.0, 7.0, 5.0, 7.0, 16.0, 26.0, 55.0, 75.0, 108.0, 194.0, 326.0, 502.0, 728.0, 1199.0, 1654.0, 2610.0, 3470.0, 5564.0, 7270.0, 11049.0, 14688.0, 21277.0, 28222.0, 37431.0, 50316.0, 57663.0, 65751.0, 68053.0, 70781.0, 68914.0, 65915.0, 60173.0, 47966.0, 42799.0, 29647.0, 22464.0, 15665.0, 11360.0, 8267.0, 5500.0, 4082.0, 2572.0, 1846.0, 1195.0, 797.0, 543.0, 322.0, 245.0, 144.0, 79.0, 60.0, 33.0, 9.0, 11.0, 6.0, 5.0, 5.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.0556640625, -0.05390167236328125, -0.0521392822265625, -0.05037689208984375, -0.048614501953125, -0.04685211181640625, -0.0450897216796875, -0.04332733154296875, -0.04156494140625, -0.03980255126953125, -0.0380401611328125, -0.03627777099609375, -0.034515380859375, -0.03275299072265625, -0.0309906005859375, -0.02922821044921875, -0.0274658203125, -0.02570343017578125, -0.0239410400390625, -0.02217864990234375, -0.020416259765625, -0.01865386962890625, -0.0168914794921875, -0.01512908935546875, -0.01336669921875, -0.01160430908203125, -0.0098419189453125, -0.00807952880859375, -0.006317138671875, -0.00455474853515625, -0.0027923583984375, -0.00102996826171875, 0.000732421875, 0.00249481201171875, 0.0042572021484375, 0.00601959228515625, 0.007781982421875, 0.00954437255859375, 0.0113067626953125, 0.01306915283203125, 0.01483154296875, 0.01659393310546875, 0.0183563232421875, 0.02011871337890625, 0.021881103515625, 0.02364349365234375, 0.0254058837890625, 0.02716827392578125, 0.0289306640625, 0.03069305419921875, 0.0324554443359375, 0.03421783447265625, 0.035980224609375, 0.03774261474609375, 0.0395050048828125, 0.04126739501953125, 0.04302978515625, 0.04479217529296875, 0.0465545654296875, 0.04831695556640625, 0.050079345703125, 0.05184173583984375, 0.0536041259765625, 0.05536651611328125, 0.05712890625]}, "parameters/_orig_mod.model.layers.20._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 0.0, 0.0, 2.0, 2.0, 3.0, 2.0, 2.0, 6.0, 16.0, 16.0, 15.0, 35.0, 47.0, 70.0, 149.0, 294.0, 606.0, 1309.0, 2730.0, 5785.0, 12295.0, 26646.0, 57537.0, 116739.0, 229359.0, 447547.0, 831723.0, 1419985.0, 2307633.0, 2927556.0, 2920318.0, 2300353.0, 1417840.0, 839759.0, 447755.0, 233689.0, 118969.0, 58686.0, 27640.0, 12617.0, 5916.0, 2956.0, 1337.0, 634.0, 289.0, 164.0, 82.0, 41.0, 23.0, 15.0, 6.0, 5.0, 3.0, 2.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.21875, -0.2119140625, -0.205078125, -0.1982421875, -0.19140625, -0.1845703125, -0.177734375, -0.1708984375, -0.1640625, -0.1572265625, -0.150390625, -0.1435546875, -0.13671875, -0.1298828125, -0.123046875, -0.1162109375, -0.109375, -0.1025390625, -0.095703125, -0.0888671875, -0.08203125, -0.0751953125, -0.068359375, -0.0615234375, -0.0546875, -0.0478515625, -0.041015625, -0.0341796875, -0.02734375, -0.0205078125, -0.013671875, -0.0068359375, 0.0, 0.0068359375, 0.013671875, 0.0205078125, 0.02734375, 0.0341796875, 0.041015625, 0.0478515625, 0.0546875, 0.0615234375, 0.068359375, 0.0751953125, 0.08203125, 0.0888671875, 0.095703125, 0.1025390625, 0.109375, 0.1162109375, 0.123046875, 0.1298828125, 0.13671875, 0.1435546875, 0.150390625, 0.1572265625, 0.1640625, 0.1708984375, 0.177734375, 0.1845703125, 0.19140625, 0.1982421875, 0.205078125, 0.2119140625, 0.21875]}, "parameters/_orig_mod.model.layers.20._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 4.0, 1.0, 0.0, 2.0, 6.0, 4.0, 11.0, 10.0, 12.0, 30.0, 33.0, 51.0, 73.0, 96.0, 119.0, 224.0, 315.0, 571.0, 1017.0, 2123.0, 4120.0, 8071.0, 16900.0, 32038.0, 65860.0, 122281.0, 207630.0, 351851.0, 520510.0, 668974.0, 678884.0, 567998.0, 397415.0, 248364.0, 138767.0, 79692.0, 40145.0, 19335.0, 10306.0, 4758.0, 2527.0, 1379.0, 696.0, 362.0, 225.0, 167.0, 92.0, 69.0, 56.0, 42.0, 21.0, 21.0, 14.0, 8.0, 6.0, 5.0, 1.0, 2.0, 2.0, 3.0, 2.0], "bins": [-0.3125, -0.303070068359375, -0.29364013671875, -0.284210205078125, -0.2747802734375, -0.265350341796875, -0.25592041015625, -0.246490478515625, -0.237060546875, -0.227630615234375, -0.21820068359375, -0.208770751953125, -0.1993408203125, -0.189910888671875, -0.18048095703125, -0.171051025390625, -0.16162109375, -0.152191162109375, -0.14276123046875, -0.133331298828125, -0.1239013671875, -0.114471435546875, -0.10504150390625, -0.095611572265625, -0.086181640625, -0.076751708984375, -0.06732177734375, -0.057891845703125, -0.0484619140625, -0.039031982421875, -0.02960205078125, -0.020172119140625, -0.0107421875, -0.001312255859375, 0.00811767578125, 0.017547607421875, 0.0269775390625, 0.036407470703125, 0.04583740234375, 0.055267333984375, 0.064697265625, 0.074127197265625, 0.08355712890625, 0.092987060546875, 0.1024169921875, 0.111846923828125, 0.12127685546875, 0.130706787109375, 0.14013671875, 0.149566650390625, 0.15899658203125, 0.168426513671875, 0.1778564453125, 0.187286376953125, 0.19671630859375, 0.206146240234375, 0.215576171875, 0.225006103515625, 0.23443603515625, 0.243865966796875, 0.2532958984375, 0.262725830078125, 0.27215576171875, 0.281585693359375, 0.291015625]}, "parameters/_orig_mod.model.layers.20._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 0.0, 1.0, 1.0, 2.0, 4.0, 7.0, 3.0, 8.0, 15.0, 16.0, 25.0, 34.0, 59.0, 75.0, 118.0, 218.0, 357.0, 517.0, 971.0, 1656.0, 2778.0, 4259.0, 7414.0, 11728.0, 18776.0, 27387.0, 39740.0, 58017.0, 69644.0, 81832.0, 88413.0, 88601.0, 82594.0, 70994.0, 60149.0, 41923.0, 29482.0, 18853.0, 12606.0, 8072.0, 5081.0, 3037.0, 1731.0, 996.0, 539.0, 350.0, 236.0, 128.0, 81.0, 40.0, 44.0, 14.0, 17.0, 8.0, 8.0, 9.0, 2.0, 2.0, 2.0, 2.0, 0.0, 1.0], "bins": [-0.076171875, -0.0738677978515625, -0.071563720703125, -0.0692596435546875, -0.06695556640625, -0.0646514892578125, -0.062347412109375, -0.0600433349609375, -0.0577392578125, -0.0554351806640625, -0.053131103515625, -0.0508270263671875, -0.04852294921875, -0.0462188720703125, -0.043914794921875, -0.0416107177734375, -0.039306640625, -0.0370025634765625, -0.034698486328125, -0.0323944091796875, -0.03009033203125, -0.0277862548828125, -0.025482177734375, -0.0231781005859375, -0.0208740234375, -0.0185699462890625, -0.016265869140625, -0.0139617919921875, -0.01165771484375, -0.0093536376953125, -0.007049560546875, -0.0047454833984375, -0.00244140625, -0.0001373291015625, 0.002166748046875, 0.0044708251953125, 0.00677490234375, 0.0090789794921875, 0.011383056640625, 0.0136871337890625, 0.0159912109375, 0.0182952880859375, 0.020599365234375, 0.0229034423828125, 0.02520751953125, 0.0275115966796875, 0.029815673828125, 0.0321197509765625, 0.034423828125, 0.0367279052734375, 0.039031982421875, 0.0413360595703125, 0.04364013671875, 0.0459442138671875, 0.048248291015625, 0.0505523681640625, 0.0528564453125, 0.0551605224609375, 0.057464599609375, 0.0597686767578125, 0.06207275390625, 0.0643768310546875, 0.066680908203125, 0.0689849853515625, 0.0712890625]}, "parameters/_orig_mod.model.layers.21._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 2.0, 2.0, 7.0, 1.0, 8.0, 1.0, 10.0, 10.0, 24.0, 30.0, 41.0, 66.0, 77.0, 151.0, 270.0, 433.0, 956.0, 1892.0, 4401.0, 9437.0, 22817.0, 50728.0, 114803.0, 243225.0, 512405.0, 952795.0, 1674623.0, 2538979.0, 2994869.0, 2896963.0, 2097951.0, 1279534.0, 718294.0, 349939.0, 170820.0, 78845.0, 35150.0, 14484.0, 6177.0, 3126.0, 1298.0, 693.0, 328.0, 177.0, 92.0, 85.0, 54.0, 37.0, 26.0, 20.0, 13.0, 12.0, 11.0, 2.0, 8.0, 0.0, 3.0, 1.0, 2.0, 1.0, 1.0, 3.0], "bins": [-0.216796875, -0.209747314453125, -0.20269775390625, -0.195648193359375, -0.1885986328125, -0.181549072265625, -0.17449951171875, -0.167449951171875, -0.160400390625, -0.153350830078125, -0.14630126953125, -0.139251708984375, -0.1322021484375, -0.125152587890625, -0.11810302734375, -0.111053466796875, -0.10400390625, -0.096954345703125, -0.08990478515625, -0.082855224609375, -0.0758056640625, -0.068756103515625, -0.06170654296875, -0.054656982421875, -0.047607421875, -0.040557861328125, -0.03350830078125, -0.026458740234375, -0.0194091796875, -0.012359619140625, -0.00531005859375, 0.001739501953125, 0.0087890625, 0.015838623046875, 0.02288818359375, 0.029937744140625, 0.0369873046875, 0.044036865234375, 0.05108642578125, 0.058135986328125, 0.065185546875, 0.072235107421875, 0.07928466796875, 0.086334228515625, 0.0933837890625, 0.100433349609375, 0.10748291015625, 0.114532470703125, 0.12158203125, 0.128631591796875, 0.13568115234375, 0.142730712890625, 0.1497802734375, 0.156829833984375, 0.16387939453125, 0.170928955078125, 0.177978515625, 0.185028076171875, 0.19207763671875, 0.199127197265625, 0.2061767578125, 0.213226318359375, 0.22027587890625, 0.227325439453125, 0.234375]}, "parameters/_orig_mod.model.layers.21._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 0.0, 5.0, 4.0, 6.0, 15.0, 17.0, 34.0, 40.0, 63.0, 96.0, 129.0, 207.0, 374.0, 594.0, 1351.0, 3092.0, 6793.0, 14046.0, 32458.0, 66979.0, 134500.0, 241351.0, 408599.0, 588205.0, 707961.0, 687691.0, 530868.0, 355745.0, 203345.0, 108209.0, 55351.0, 24559.0, 11660.0, 5116.0, 2407.0, 1097.0, 532.0, 298.0, 160.0, 134.0, 59.0, 46.0, 40.0, 18.0, 15.0, 8.0, 5.0, 4.0, 4.0, 1.0, 3.0, 1.0, 0.0, 1.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.3125, -0.3023681640625, -0.292236328125, -0.2821044921875, -0.27197265625, -0.2618408203125, -0.251708984375, -0.2415771484375, -0.2314453125, -0.2213134765625, -0.211181640625, -0.2010498046875, -0.19091796875, -0.1807861328125, -0.170654296875, -0.1605224609375, -0.150390625, -0.1402587890625, -0.130126953125, -0.1199951171875, -0.10986328125, -0.0997314453125, -0.089599609375, -0.0794677734375, -0.0693359375, -0.0592041015625, -0.049072265625, -0.0389404296875, -0.02880859375, -0.0186767578125, -0.008544921875, 0.0015869140625, 0.01171875, 0.0218505859375, 0.031982421875, 0.0421142578125, 0.05224609375, 0.0623779296875, 0.072509765625, 0.0826416015625, 0.0927734375, 0.1029052734375, 0.113037109375, 0.1231689453125, 0.13330078125, 0.1434326171875, 0.153564453125, 0.1636962890625, 0.173828125, 0.1839599609375, 0.194091796875, 0.2042236328125, 0.21435546875, 0.2244873046875, 0.234619140625, 0.2447509765625, 0.2548828125, 0.2650146484375, 0.275146484375, 0.2852783203125, 0.29541015625, 0.3055419921875, 0.315673828125, 0.3258056640625, 0.3359375]}, "parameters/_orig_mod.model.layers.21._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 2.0, 3.0, 3.0, 6.0, 5.0, 15.0, 25.0, 60.0, 124.0, 195.0, 370.0, 615.0, 1003.0, 1783.0, 2764.0, 4413.0, 6543.0, 9443.0, 13212.0, 18334.0, 24364.0, 31317.0, 42157.0, 44786.0, 54966.0, 58466.0, 61308.0, 63469.0, 62515.0, 57677.0, 58959.0, 47960.0, 45518.0, 33824.0, 25822.0, 20951.0, 15165.0, 11004.0, 7315.0, 4650.0, 3374.0, 2145.0, 1297.0, 769.0, 462.0, 237.0, 128.0, 72.0, 38.0, 22.0, 6.0, 6.0, 4.0, 0.0, 0.0, 2.0, 2.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.04833984375, -0.046764373779296875, -0.04518890380859375, -0.043613433837890625, -0.0420379638671875, -0.040462493896484375, -0.03888702392578125, -0.037311553955078125, -0.035736083984375, -0.034160614013671875, -0.03258514404296875, -0.031009674072265625, -0.0294342041015625, -0.027858734130859375, -0.02628326416015625, -0.024707794189453125, -0.02313232421875, -0.021556854248046875, -0.01998138427734375, -0.018405914306640625, -0.0168304443359375, -0.015254974365234375, -0.01367950439453125, -0.012104034423828125, -0.010528564453125, -0.008953094482421875, -0.00737762451171875, -0.005802154541015625, -0.0042266845703125, -0.002651214599609375, -0.00107574462890625, 0.000499725341796875, 0.0020751953125, 0.003650665283203125, 0.00522613525390625, 0.006801605224609375, 0.0083770751953125, 0.009952545166015625, 0.01152801513671875, 0.013103485107421875, 0.014678955078125, 0.016254425048828125, 0.01782989501953125, 0.019405364990234375, 0.0209808349609375, 0.022556304931640625, 0.02413177490234375, 0.025707244873046875, 0.02728271484375, 0.028858184814453125, 0.03043365478515625, 0.032009124755859375, 0.0335845947265625, 0.035160064697265625, 0.03673553466796875, 0.038311004638671875, 0.039886474609375, 0.041461944580078125, 0.04303741455078125, 0.044612884521484375, 0.0461883544921875, 0.047763824462890625, 0.04933929443359375, 0.050914764404296875, 0.052490234375]}, "parameters/_orig_mod.model.layers.22._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 4.0, 2.0, 2.0, 5.0, 4.0, 6.0, 21.0, 8.0, 11.0, 21.0, 33.0, 56.0, 64.0, 126.0, 208.0, 356.0, 664.0, 1620.0, 3470.0, 8298.0, 21105.0, 49699.0, 109205.0, 226735.0, 452581.0, 848591.0, 1401459.0, 2249987.0, 2779160.0, 2827319.0, 2340822.0, 1542517.0, 926720.0, 515762.0, 244731.0, 125746.0, 58043.0, 24622.0, 9978.0, 3904.0, 1875.0, 737.0, 373.0, 215.0, 123.0, 75.0, 47.0, 32.0, 26.0, 17.0, 12.0, 5.0, 3.0, 5.0, 0.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.2236328125, -0.217071533203125, -0.21051025390625, -0.203948974609375, -0.1973876953125, -0.190826416015625, -0.18426513671875, -0.177703857421875, -0.171142578125, -0.164581298828125, -0.15802001953125, -0.151458740234375, -0.1448974609375, -0.138336181640625, -0.13177490234375, -0.125213623046875, -0.11865234375, -0.112091064453125, -0.10552978515625, -0.098968505859375, -0.0924072265625, -0.085845947265625, -0.07928466796875, -0.072723388671875, -0.066162109375, -0.059600830078125, -0.05303955078125, -0.046478271484375, -0.0399169921875, -0.033355712890625, -0.02679443359375, -0.020233154296875, -0.013671875, -0.007110595703125, -0.00054931640625, 0.006011962890625, 0.0125732421875, 0.019134521484375, 0.02569580078125, 0.032257080078125, 0.038818359375, 0.045379638671875, 0.05194091796875, 0.058502197265625, 0.0650634765625, 0.071624755859375, 0.07818603515625, 0.084747314453125, 0.09130859375, 0.097869873046875, 0.10443115234375, 0.110992431640625, 0.1175537109375, 0.124114990234375, 0.13067626953125, 0.137237548828125, 0.143798828125, 0.150360107421875, 0.15692138671875, 0.163482666015625, 0.1700439453125, 0.176605224609375, 0.18316650390625, 0.189727783203125, 0.1962890625]}, "parameters/_orig_mod.model.layers.22._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 1.0, 2.0, 3.0, 3.0, 3.0, 5.0, 12.0, 4.0, 14.0, 18.0, 29.0, 34.0, 56.0, 77.0, 93.0, 157.0, 301.0, 544.0, 1084.0, 2485.0, 5471.0, 12651.0, 29242.0, 64354.0, 130276.0, 241782.0, 427044.0, 645603.0, 743957.0, 719441.0, 515560.0, 317501.0, 171839.0, 89654.0, 41407.0, 18688.0, 8136.0, 3404.0, 1613.0, 709.0, 373.0, 238.0, 125.0, 99.0, 76.0, 43.0, 31.0, 16.0, 13.0, 12.0, 3.0, 8.0, 2.0, 1.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.33203125, -0.321563720703125, -0.31109619140625, -0.300628662109375, -0.2901611328125, -0.279693603515625, -0.26922607421875, -0.258758544921875, -0.248291015625, -0.237823486328125, -0.22735595703125, -0.216888427734375, -0.2064208984375, -0.195953369140625, -0.18548583984375, -0.175018310546875, -0.16455078125, -0.154083251953125, -0.14361572265625, -0.133148193359375, -0.1226806640625, -0.112213134765625, -0.10174560546875, -0.091278076171875, -0.080810546875, -0.070343017578125, -0.05987548828125, -0.049407958984375, -0.0389404296875, -0.028472900390625, -0.01800537109375, -0.007537841796875, 0.0029296875, 0.013397216796875, 0.02386474609375, 0.034332275390625, 0.0447998046875, 0.055267333984375, 0.06573486328125, 0.076202392578125, 0.086669921875, 0.097137451171875, 0.10760498046875, 0.118072509765625, 0.1285400390625, 0.139007568359375, 0.14947509765625, 0.159942626953125, 0.17041015625, 0.180877685546875, 0.19134521484375, 0.201812744140625, 0.2122802734375, 0.222747802734375, 0.23321533203125, 0.243682861328125, 0.254150390625, 0.264617919921875, 0.27508544921875, 0.285552978515625, 0.2960205078125, 0.306488037109375, 0.31695556640625, 0.327423095703125, 0.337890625]}, "parameters/_orig_mod.model.layers.22._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 5.0, 1.0, 3.0, 10.0, 7.0, 12.0, 23.0, 31.0, 53.0, 95.0, 133.0, 252.0, 415.0, 748.0, 1554.0, 3055.0, 5440.0, 9423.0, 15677.0, 25286.0, 35798.0, 50660.0, 67183.0, 78482.0, 85415.0, 88972.0, 85204.0, 76938.0, 65977.0, 49278.0, 34686.0, 24149.0, 14665.0, 9027.0, 5119.0, 2824.0, 1421.0, 699.0, 401.0, 218.0, 133.0, 79.0, 44.0, 18.0, 14.0, 18.0, 11.0, 7.0, 5.0, 4.0, 3.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.08642578125, -0.083587646484375, -0.08074951171875, -0.077911376953125, -0.0750732421875, -0.072235107421875, -0.06939697265625, -0.066558837890625, -0.063720703125, -0.060882568359375, -0.05804443359375, -0.055206298828125, -0.0523681640625, -0.049530029296875, -0.04669189453125, -0.043853759765625, -0.041015625, -0.038177490234375, -0.03533935546875, -0.032501220703125, -0.0296630859375, -0.026824951171875, -0.02398681640625, -0.021148681640625, -0.018310546875, -0.015472412109375, -0.01263427734375, -0.009796142578125, -0.0069580078125, -0.004119873046875, -0.00128173828125, 0.001556396484375, 0.00439453125, 0.007232666015625, 0.01007080078125, 0.012908935546875, 0.0157470703125, 0.018585205078125, 0.02142333984375, 0.024261474609375, 0.027099609375, 0.029937744140625, 0.03277587890625, 0.035614013671875, 0.0384521484375, 0.041290283203125, 0.04412841796875, 0.046966552734375, 0.0498046875, 0.052642822265625, 0.05548095703125, 0.058319091796875, 0.0611572265625, 0.063995361328125, 0.06683349609375, 0.069671630859375, 0.072509765625, 0.075347900390625, 0.07818603515625, 0.081024169921875, 0.0838623046875, 0.086700439453125, 0.08953857421875, 0.092376708984375, 0.09521484375]}, "parameters/_orig_mod.model.layers.23._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 5.0, 7.0, 4.0, 7.0, 12.0, 13.0, 23.0, 18.0, 38.0, 26.0, 71.0, 101.0, 113.0, 217.0, 288.0, 672.0, 1353.0, 3010.0, 7995.0, 20860.0, 53574.0, 131980.0, 282365.0, 593772.0, 1121066.0, 1873100.0, 2825684.0, 3043467.0, 2818911.0, 1849791.0, 1078206.0, 579342.0, 282777.0, 122832.0, 52056.0, 20088.0, 7507.0, 3129.0, 1290.0, 561.0, 324.0, 157.0, 116.0, 91.0, 50.0, 41.0, 33.0, 18.0, 20.0, 15.0, 6.0, 4.0, 3.0, 1.0, 1.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.2314453125, -0.2243194580078125, -0.217193603515625, -0.2100677490234375, -0.20294189453125, -0.1958160400390625, -0.188690185546875, -0.1815643310546875, -0.1744384765625, -0.1673126220703125, -0.160186767578125, -0.1530609130859375, -0.14593505859375, -0.1388092041015625, -0.131683349609375, -0.1245574951171875, -0.117431640625, -0.1103057861328125, -0.103179931640625, -0.0960540771484375, -0.08892822265625, -0.0818023681640625, -0.074676513671875, -0.0675506591796875, -0.0604248046875, -0.0532989501953125, -0.046173095703125, -0.0390472412109375, -0.03192138671875, -0.0247955322265625, -0.017669677734375, -0.0105438232421875, -0.00341796875, 0.0037078857421875, 0.010833740234375, 0.0179595947265625, 0.02508544921875, 0.0322113037109375, 0.039337158203125, 0.0464630126953125, 0.0535888671875, 0.0607147216796875, 0.067840576171875, 0.0749664306640625, 0.08209228515625, 0.0892181396484375, 0.096343994140625, 0.1034698486328125, 0.110595703125, 0.1177215576171875, 0.124847412109375, 0.1319732666015625, 0.13909912109375, 0.1462249755859375, 0.153350830078125, 0.1604766845703125, 0.1676025390625, 0.1747283935546875, 0.181854248046875, 0.1889801025390625, 0.19610595703125, 0.2032318115234375, 0.210357666015625, 0.2174835205078125, 0.224609375]}, "parameters/_orig_mod.model.layers.23._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 2.0, 4.0, 7.0, 4.0, 12.0, 13.0, 26.0, 31.0, 43.0, 84.0, 102.0, 162.0, 274.0, 548.0, 1229.0, 3207.0, 8789.0, 21899.0, 59836.0, 139983.0, 289959.0, 532034.0, 805493.0, 856584.0, 685817.0, 419103.0, 207513.0, 98837.0, 38460.0, 14641.0, 5601.0, 2101.0, 850.0, 399.0, 230.0, 157.0, 86.0, 64.0, 38.0, 21.0, 18.0, 20.0, 8.0, 4.0, 2.0, 1.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.3046875, -0.292633056640625, -0.28057861328125, -0.268524169921875, -0.2564697265625, -0.244415283203125, -0.23236083984375, -0.220306396484375, -0.208251953125, -0.196197509765625, -0.18414306640625, -0.172088623046875, -0.1600341796875, -0.147979736328125, -0.13592529296875, -0.123870849609375, -0.11181640625, -0.099761962890625, -0.08770751953125, -0.075653076171875, -0.0635986328125, -0.051544189453125, -0.03948974609375, -0.027435302734375, -0.015380859375, -0.003326416015625, 0.00872802734375, 0.020782470703125, 0.0328369140625, 0.044891357421875, 0.05694580078125, 0.069000244140625, 0.0810546875, 0.093109130859375, 0.10516357421875, 0.117218017578125, 0.1292724609375, 0.141326904296875, 0.15338134765625, 0.165435791015625, 0.177490234375, 0.189544677734375, 0.20159912109375, 0.213653564453125, 0.2257080078125, 0.237762451171875, 0.24981689453125, 0.261871337890625, 0.27392578125, 0.285980224609375, 0.29803466796875, 0.310089111328125, 0.3221435546875, 0.334197998046875, 0.34625244140625, 0.358306884765625, 0.370361328125, 0.382415771484375, 0.39447021484375, 0.406524658203125, 0.4185791015625, 0.430633544921875, 0.44268798828125, 0.454742431640625, 0.466796875]}, "parameters/_orig_mod.model.layers.23._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 2.0, 0.0, 0.0, 2.0, 5.0, 4.0, 5.0, 3.0, 5.0, 13.0, 11.0, 8.0, 27.0, 39.0, 76.0, 116.0, 212.0, 364.0, 766.0, 1387.0, 2550.0, 4652.0, 8163.0, 13157.0, 20169.0, 29720.0, 41423.0, 53975.0, 68226.0, 79637.0, 81237.0, 83108.0, 80374.0, 71811.0, 58103.0, 45613.0, 33602.0, 23195.0, 15453.0, 9752.0, 5655.0, 3197.0, 1788.0, 922.0, 489.0, 279.0, 142.0, 81.0, 53.0, 33.0, 24.0, 17.0, 9.0, 13.0, 4.0, 3.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.10302734375, -0.10009765625, -0.09716796875, -0.09423828125, -0.09130859375, -0.08837890625, -0.08544921875, -0.08251953125, -0.07958984375, -0.07666015625, -0.07373046875, -0.07080078125, -0.06787109375, -0.06494140625, -0.06201171875, -0.05908203125, -0.05615234375, -0.05322265625, -0.05029296875, -0.04736328125, -0.04443359375, -0.04150390625, -0.03857421875, -0.03564453125, -0.03271484375, -0.02978515625, -0.02685546875, -0.02392578125, -0.02099609375, -0.01806640625, -0.01513671875, -0.01220703125, -0.00927734375, -0.00634765625, -0.00341796875, -0.00048828125, 0.00244140625, 0.00537109375, 0.00830078125, 0.01123046875, 0.01416015625, 0.01708984375, 0.02001953125, 0.02294921875, 0.02587890625, 0.02880859375, 0.03173828125, 0.03466796875, 0.03759765625, 0.04052734375, 0.04345703125, 0.04638671875, 0.04931640625, 0.05224609375, 0.05517578125, 0.05810546875, 0.06103515625, 0.06396484375, 0.06689453125, 0.06982421875, 0.07275390625, 0.07568359375, 0.07861328125, 0.08154296875, 0.08447265625]}, "parameters/_orig_mod.model.layers.24._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 3.0, 1.0, 3.0, 3.0, 6.0, 10.0, 10.0, 9.0, 24.0, 34.0, 48.0, 76.0, 99.0, 114.0, 172.0, 354.0, 570.0, 1075.0, 2244.0, 5145.0, 13502.0, 33269.0, 77864.0, 180709.0, 363509.0, 751505.0, 1301455.0, 2149202.0, 3010015.0, 3077734.0, 2467933.0, 1565112.0, 896041.0, 473918.0, 224209.0, 105735.0, 44262.0, 18015.0, 6956.0, 3217.0, 1376.0, 610.0, 344.0, 230.0, 157.0, 87.0, 68.0, 56.0, 39.0, 24.0, 15.0, 9.0, 12.0, 7.0, 5.0, 3.0, 1.0, 2.0, 3.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.2080078125, -0.20111083984375, -0.1942138671875, -0.18731689453125, -0.180419921875, -0.17352294921875, -0.1666259765625, -0.15972900390625, -0.15283203125, -0.14593505859375, -0.1390380859375, -0.13214111328125, -0.125244140625, -0.11834716796875, -0.1114501953125, -0.10455322265625, -0.09765625, -0.09075927734375, -0.0838623046875, -0.07696533203125, -0.070068359375, -0.06317138671875, -0.0562744140625, -0.04937744140625, -0.04248046875, -0.03558349609375, -0.0286865234375, -0.02178955078125, -0.014892578125, -0.00799560546875, -0.0010986328125, 0.00579833984375, 0.0126953125, 0.01959228515625, 0.0264892578125, 0.03338623046875, 0.040283203125, 0.04718017578125, 0.0540771484375, 0.06097412109375, 0.06787109375, 0.07476806640625, 0.0816650390625, 0.08856201171875, 0.095458984375, 0.10235595703125, 0.1092529296875, 0.11614990234375, 0.123046875, 0.12994384765625, 0.1368408203125, 0.14373779296875, 0.150634765625, 0.15753173828125, 0.1644287109375, 0.17132568359375, 0.17822265625, 0.18511962890625, 0.1920166015625, 0.19891357421875, 0.205810546875, 0.21270751953125, 0.2196044921875, 0.22650146484375, 0.2333984375]}, "parameters/_orig_mod.model.layers.24._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 5.0, 5.0, 3.0, 2.0, 5.0, 7.0, 13.0, 5.0, 16.0, 31.0, 41.0, 51.0, 55.0, 89.0, 129.0, 153.0, 281.0, 439.0, 889.0, 1994.0, 4321.0, 9415.0, 20986.0, 46423.0, 98424.0, 186597.0, 360019.0, 592328.0, 805965.0, 794323.0, 573093.0, 344346.0, 181930.0, 91431.0, 44011.0, 19901.0, 8773.0, 3871.0, 1721.0, 860.0, 513.0, 261.0, 155.0, 120.0, 84.0, 57.0, 37.0, 33.0, 25.0, 14.0, 17.0, 9.0, 7.0, 7.0, 1.0, 2.0, 0.0, 2.0, 2.0, 1.0, 2.0, 1.0], "bins": [-0.3203125, -0.309967041015625, -0.29962158203125, -0.289276123046875, -0.2789306640625, -0.268585205078125, -0.25823974609375, -0.247894287109375, -0.237548828125, -0.227203369140625, -0.21685791015625, -0.206512451171875, -0.1961669921875, -0.185821533203125, -0.17547607421875, -0.165130615234375, -0.15478515625, -0.144439697265625, -0.13409423828125, -0.123748779296875, -0.1134033203125, -0.103057861328125, -0.09271240234375, -0.082366943359375, -0.072021484375, -0.061676025390625, -0.05133056640625, -0.040985107421875, -0.0306396484375, -0.020294189453125, -0.00994873046875, 0.000396728515625, 0.0107421875, 0.021087646484375, 0.03143310546875, 0.041778564453125, 0.0521240234375, 0.062469482421875, 0.07281494140625, 0.083160400390625, 0.093505859375, 0.103851318359375, 0.11419677734375, 0.124542236328125, 0.1348876953125, 0.145233154296875, 0.15557861328125, 0.165924072265625, 0.17626953125, 0.186614990234375, 0.19696044921875, 0.207305908203125, 0.2176513671875, 0.227996826171875, 0.23834228515625, 0.248687744140625, 0.259033203125, 0.269378662109375, 0.27972412109375, 0.290069580078125, 0.3004150390625, 0.310760498046875, 0.32110595703125, 0.331451416015625, 0.341796875]}, "parameters/_orig_mod.model.layers.24._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 5.0, 5.0, 10.0, 27.0, 59.0, 85.0, 153.0, 282.0, 487.0, 796.0, 1337.0, 1955.0, 3078.0, 4572.0, 6798.0, 9610.0, 13015.0, 16111.0, 21977.0, 27266.0, 34041.0, 40957.0, 45884.0, 53919.0, 59538.0, 60332.0, 62122.0, 59160.0, 57167.0, 51798.0, 42826.0, 37886.0, 31605.0, 23697.0, 19859.0, 15245.0, 11475.0, 8305.0, 5487.0, 3895.0, 2528.0, 1699.0, 1053.0, 669.0, 375.0, 238.0, 132.0, 71.0, 39.0, 24.0, 7.0, 6.0, 5.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0732421875, -0.07082366943359375, -0.0684051513671875, -0.06598663330078125, -0.063568115234375, -0.06114959716796875, -0.0587310791015625, -0.05631256103515625, -0.05389404296875, -0.05147552490234375, -0.0490570068359375, -0.04663848876953125, -0.044219970703125, -0.04180145263671875, -0.0393829345703125, -0.03696441650390625, -0.0345458984375, -0.03212738037109375, -0.0297088623046875, -0.02729034423828125, -0.024871826171875, -0.02245330810546875, -0.0200347900390625, -0.01761627197265625, -0.01519775390625, -0.01277923583984375, -0.0103607177734375, -0.00794219970703125, -0.005523681640625, -0.00310516357421875, -0.0006866455078125, 0.00173187255859375, 0.004150390625, 0.00656890869140625, 0.0089874267578125, 0.01140594482421875, 0.013824462890625, 0.01624298095703125, 0.0186614990234375, 0.02108001708984375, 0.02349853515625, 0.02591705322265625, 0.0283355712890625, 0.03075408935546875, 0.033172607421875, 0.03559112548828125, 0.0380096435546875, 0.04042816162109375, 0.0428466796875, 0.04526519775390625, 0.0476837158203125, 0.05010223388671875, 0.052520751953125, 0.05493927001953125, 0.0573577880859375, 0.05977630615234375, 0.06219482421875, 0.06461334228515625, 0.0670318603515625, 0.06945037841796875, 0.071868896484375, 0.07428741455078125, 0.0767059326171875, 0.07912445068359375, 0.08154296875]}, "parameters/_orig_mod.model.layers.25._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 1.0, 3.0, 0.0, 4.0, 10.0, 6.0, 10.0, 14.0, 22.0, 34.0, 47.0, 74.0, 90.0, 105.0, 215.0, 268.0, 487.0, 754.0, 1472.0, 3219.0, 7973.0, 21690.0, 55649.0, 125890.0, 293822.0, 638187.0, 1169223.0, 2029569.0, 2956399.0, 3128665.0, 2706803.0, 1737913.0, 996748.0, 508413.0, 221506.0, 102264.0, 41727.0, 15995.0, 6178.0, 2509.0, 1265.0, 682.0, 432.0, 283.0, 156.0, 134.0, 81.0, 58.0, 50.0, 30.0, 23.0, 24.0, 4.0, 8.0, 7.0, 5.0, 6.0, 3.0, 2.0, 0.0, 2.0], "bins": [-0.2275390625, -0.2205047607421875, -0.213470458984375, -0.2064361572265625, -0.19940185546875, -0.1923675537109375, -0.185333251953125, -0.1782989501953125, -0.1712646484375, -0.1642303466796875, -0.157196044921875, -0.1501617431640625, -0.14312744140625, -0.1360931396484375, -0.129058837890625, -0.1220245361328125, -0.114990234375, -0.1079559326171875, -0.100921630859375, -0.0938873291015625, -0.08685302734375, -0.0798187255859375, -0.072784423828125, -0.0657501220703125, -0.0587158203125, -0.0516815185546875, -0.044647216796875, -0.0376129150390625, -0.03057861328125, -0.0235443115234375, -0.016510009765625, -0.0094757080078125, -0.00244140625, 0.0045928955078125, 0.011627197265625, 0.0186614990234375, 0.02569580078125, 0.0327301025390625, 0.039764404296875, 0.0467987060546875, 0.0538330078125, 0.0608673095703125, 0.067901611328125, 0.0749359130859375, 0.08197021484375, 0.0890045166015625, 0.096038818359375, 0.1030731201171875, 0.110107421875, 0.1171417236328125, 0.124176025390625, 0.1312103271484375, 0.13824462890625, 0.1452789306640625, 0.152313232421875, 0.1593475341796875, 0.1663818359375, 0.1734161376953125, 0.180450439453125, 0.1874847412109375, 0.19451904296875, 0.2015533447265625, 0.208587646484375, 0.2156219482421875, 0.22265625]}, "parameters/_orig_mod.model.layers.25._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 5.0, 1.0, 3.0, 3.0, 4.0, 8.0, 6.0, 6.0, 20.0, 23.0, 26.0, 51.0, 55.0, 79.0, 149.0, 196.0, 368.0, 870.0, 2269.0, 6503.0, 17729.0, 50626.0, 120144.0, 263687.0, 516677.0, 887595.0, 987716.0, 671331.0, 368221.0, 176455.0, 76830.0, 29457.0, 10639.0, 3881.0, 1244.0, 580.0, 290.0, 169.0, 116.0, 74.0, 56.0, 39.0, 19.0, 19.0, 13.0, 16.0, 15.0, 5.0, 5.0, 4.0, 2.0, 2.0, 1.0], "bins": [-0.458984375, -0.4466552734375, -0.434326171875, -0.4219970703125, -0.40966796875, -0.3973388671875, -0.385009765625, -0.3726806640625, -0.3603515625, -0.3480224609375, -0.335693359375, -0.3233642578125, -0.31103515625, -0.2987060546875, -0.286376953125, -0.2740478515625, -0.26171875, -0.2493896484375, -0.237060546875, -0.2247314453125, -0.21240234375, -0.2000732421875, -0.187744140625, -0.1754150390625, -0.1630859375, -0.1507568359375, -0.138427734375, -0.1260986328125, -0.11376953125, -0.1014404296875, -0.089111328125, -0.0767822265625, -0.064453125, -0.0521240234375, -0.039794921875, -0.0274658203125, -0.01513671875, -0.0028076171875, 0.009521484375, 0.0218505859375, 0.0341796875, 0.0465087890625, 0.058837890625, 0.0711669921875, 0.08349609375, 0.0958251953125, 0.108154296875, 0.1204833984375, 0.1328125, 0.1451416015625, 0.157470703125, 0.1697998046875, 0.18212890625, 0.1944580078125, 0.206787109375, 0.2191162109375, 0.2314453125, 0.2437744140625, 0.256103515625, 0.2684326171875, 0.28076171875, 0.2930908203125, 0.305419921875, 0.3177490234375, 0.330078125]}, "parameters/_orig_mod.model.layers.25._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 5.0, 3.0, 6.0, 5.0, 5.0, 6.0, 12.0, 17.0, 19.0, 34.0, 50.0, 60.0, 93.0, 140.0, 226.0, 429.0, 699.0, 1235.0, 2355.0, 4624.0, 8059.0, 14252.0, 23065.0, 33396.0, 49146.0, 62195.0, 79167.0, 89577.0, 89912.0, 88596.0, 81302.0, 65290.0, 51148.0, 35338.0, 24146.0, 15155.0, 8815.0, 5062.0, 2718.0, 1335.0, 770.0, 462.0, 228.0, 168.0, 104.0, 66.0, 50.0, 36.0, 19.0, 19.0, 17.0, 9.0, 7.0, 3.0, 7.0, 5.0, 2.0, 5.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.099609375, -0.09635162353515625, -0.0930938720703125, -0.08983612060546875, -0.086578369140625, -0.08332061767578125, -0.0800628662109375, -0.07680511474609375, -0.07354736328125, -0.07028961181640625, -0.0670318603515625, -0.06377410888671875, -0.060516357421875, -0.05725860595703125, -0.0540008544921875, -0.05074310302734375, -0.0474853515625, -0.04422760009765625, -0.0409698486328125, -0.03771209716796875, -0.034454345703125, -0.03119659423828125, -0.0279388427734375, -0.02468109130859375, -0.02142333984375, -0.01816558837890625, -0.0149078369140625, -0.01165008544921875, -0.008392333984375, -0.00513458251953125, -0.0018768310546875, 0.00138092041015625, 0.004638671875, 0.00789642333984375, 0.0111541748046875, 0.01441192626953125, 0.017669677734375, 0.02092742919921875, 0.0241851806640625, 0.02744293212890625, 0.03070068359375, 0.03395843505859375, 0.0372161865234375, 0.04047393798828125, 0.043731689453125, 0.04698944091796875, 0.0502471923828125, 0.05350494384765625, 0.0567626953125, 0.06002044677734375, 0.0632781982421875, 0.06653594970703125, 0.069793701171875, 0.07305145263671875, 0.0763092041015625, 0.07956695556640625, 0.08282470703125, 0.08608245849609375, 0.0893402099609375, 0.09259796142578125, 0.095855712890625, 0.09911346435546875, 0.1023712158203125, 0.10562896728515625, 0.10888671875]}, "parameters/_orig_mod.model.layers.26._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 1.0, 3.0, 2.0, 3.0, 6.0, 6.0, 10.0, 11.0, 20.0, 33.0, 49.0, 88.0, 101.0, 150.0, 214.0, 352.0, 506.0, 1004.0, 1857.0, 4236.0, 10169.0, 25415.0, 59901.0, 128244.0, 276378.0, 584621.0, 1053885.0, 1801606.0, 2733516.0, 2959215.0, 2814926.0, 1941920.0, 1149261.0, 651246.0, 308811.0, 151840.0, 66234.0, 30270.0, 11846.0, 4390.0, 2246.0, 1102.0, 607.0, 288.0, 202.0, 140.0, 100.0, 52.0, 35.0, 31.0, 16.0, 11.0, 10.0, 5.0, 6.0, 5.0, 3.0, 4.0, 2.0, 0.0, 2.0], "bins": [-0.2197265625, -0.212982177734375, -0.20623779296875, -0.199493408203125, -0.1927490234375, -0.186004638671875, -0.17926025390625, -0.172515869140625, -0.165771484375, -0.159027099609375, -0.15228271484375, -0.145538330078125, -0.1387939453125, -0.132049560546875, -0.12530517578125, -0.118560791015625, -0.11181640625, -0.105072021484375, -0.09832763671875, -0.091583251953125, -0.0848388671875, -0.078094482421875, -0.07135009765625, -0.064605712890625, -0.057861328125, -0.051116943359375, -0.04437255859375, -0.037628173828125, -0.0308837890625, -0.024139404296875, -0.01739501953125, -0.010650634765625, -0.00390625, 0.002838134765625, 0.00958251953125, 0.016326904296875, 0.0230712890625, 0.029815673828125, 0.03656005859375, 0.043304443359375, 0.050048828125, 0.056793212890625, 0.06353759765625, 0.070281982421875, 0.0770263671875, 0.083770751953125, 0.09051513671875, 0.097259521484375, 0.10400390625, 0.110748291015625, 0.11749267578125, 0.124237060546875, 0.1309814453125, 0.137725830078125, 0.14447021484375, 0.151214599609375, 0.157958984375, 0.164703369140625, 0.17144775390625, 0.178192138671875, 0.1849365234375, 0.191680908203125, 0.19842529296875, 0.205169677734375, 0.2119140625]}, "parameters/_orig_mod.model.layers.26._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 4.0, 3.0, 5.0, 3.0, 6.0, 16.0, 13.0, 30.0, 25.0, 43.0, 58.0, 90.0, 109.0, 195.0, 260.0, 460.0, 880.0, 1918.0, 4306.0, 10323.0, 21180.0, 49114.0, 103855.0, 200535.0, 383754.0, 635749.0, 847532.0, 797179.0, 531668.0, 305441.0, 156455.0, 77149.0, 36573.0, 15795.0, 7192.0, 3268.0, 1448.0, 628.0, 350.0, 196.0, 133.0, 114.0, 74.0, 48.0, 33.0, 33.0, 10.0, 12.0, 10.0, 10.0, 3.0, 4.0, 5.0, 1.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.357421875, -0.346527099609375, -0.33563232421875, -0.324737548828125, -0.3138427734375, -0.302947998046875, -0.29205322265625, -0.281158447265625, -0.270263671875, -0.259368896484375, -0.24847412109375, -0.237579345703125, -0.2266845703125, -0.215789794921875, -0.20489501953125, -0.194000244140625, -0.18310546875, -0.172210693359375, -0.16131591796875, -0.150421142578125, -0.1395263671875, -0.128631591796875, -0.11773681640625, -0.106842041015625, -0.095947265625, -0.085052490234375, -0.07415771484375, -0.063262939453125, -0.0523681640625, -0.041473388671875, -0.03057861328125, -0.019683837890625, -0.0087890625, 0.002105712890625, 0.01300048828125, 0.023895263671875, 0.0347900390625, 0.045684814453125, 0.05657958984375, 0.067474365234375, 0.078369140625, 0.089263916015625, 0.10015869140625, 0.111053466796875, 0.1219482421875, 0.132843017578125, 0.14373779296875, 0.154632568359375, 0.16552734375, 0.176422119140625, 0.18731689453125, 0.198211669921875, 0.2091064453125, 0.220001220703125, 0.23089599609375, 0.241790771484375, 0.252685546875, 0.263580322265625, 0.27447509765625, 0.285369873046875, 0.2962646484375, 0.307159423828125, 0.31805419921875, 0.328948974609375, 0.33984375]}, "parameters/_orig_mod.model.layers.26._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 7.0, 4.0, 10.0, 17.0, 30.0, 41.0, 95.0, 141.0, 272.0, 436.0, 670.0, 1182.0, 1886.0, 2822.0, 4551.0, 6245.0, 9493.0, 12072.0, 17349.0, 22089.0, 28980.0, 35909.0, 43189.0, 53035.0, 59396.0, 62262.0, 64345.0, 63622.0, 61181.0, 56850.0, 50771.0, 42000.0, 35081.0, 27837.0, 21928.0, 15824.0, 12444.0, 8248.0, 6190.0, 3982.0, 2770.0, 1570.0, 1191.0, 675.0, 394.0, 251.0, 148.0, 86.0, 51.0, 25.0, 15.0, 6.0, 3.0, 3.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.0703125, -0.06817626953125, -0.0660400390625, -0.06390380859375, -0.061767578125, -0.05963134765625, -0.0574951171875, -0.05535888671875, -0.05322265625, -0.05108642578125, -0.0489501953125, -0.04681396484375, -0.044677734375, -0.04254150390625, -0.0404052734375, -0.03826904296875, -0.0361328125, -0.03399658203125, -0.0318603515625, -0.02972412109375, -0.027587890625, -0.02545166015625, -0.0233154296875, -0.02117919921875, -0.01904296875, -0.01690673828125, -0.0147705078125, -0.01263427734375, -0.010498046875, -0.00836181640625, -0.0062255859375, -0.00408935546875, -0.001953125, 0.00018310546875, 0.0023193359375, 0.00445556640625, 0.006591796875, 0.00872802734375, 0.0108642578125, 0.01300048828125, 0.01513671875, 0.01727294921875, 0.0194091796875, 0.02154541015625, 0.023681640625, 0.02581787109375, 0.0279541015625, 0.03009033203125, 0.0322265625, 0.03436279296875, 0.0364990234375, 0.03863525390625, 0.040771484375, 0.04290771484375, 0.0450439453125, 0.04718017578125, 0.04931640625, 0.05145263671875, 0.0535888671875, 0.05572509765625, 0.057861328125, 0.05999755859375, 0.0621337890625, 0.06427001953125, 0.06640625]}, "parameters/_orig_mod.model.layers.27._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [3.0, 0.0, 2.0, 0.0, 3.0, 5.0, 7.0, 7.0, 11.0, 17.0, 15.0, 22.0, 35.0, 50.0, 49.0, 123.0, 202.0, 260.0, 372.0, 621.0, 1148.0, 2525.0, 5504.0, 14263.0, 35655.0, 84796.0, 199117.0, 413952.0, 826611.0, 1488049.0, 2433082.0, 3106947.0, 3035777.0, 2322454.0, 1372750.0, 756110.0, 370405.0, 178199.0, 75098.0, 31055.0, 12075.0, 4769.0, 2346.0, 1040.0, 587.0, 350.0, 237.0, 157.0, 121.0, 65.0, 46.0, 37.0, 24.0, 22.0, 19.0, 3.0, 4.0, 4.0, 3.0, 2.0, 2.0, 1.0, 0.0, 1.0], "bins": [-0.2197265625, -0.2128448486328125, -0.205963134765625, -0.1990814208984375, -0.19219970703125, -0.1853179931640625, -0.178436279296875, -0.1715545654296875, -0.1646728515625, -0.1577911376953125, -0.150909423828125, -0.1440277099609375, -0.13714599609375, -0.1302642822265625, -0.123382568359375, -0.1165008544921875, -0.109619140625, -0.1027374267578125, -0.095855712890625, -0.0889739990234375, -0.08209228515625, -0.0752105712890625, -0.068328857421875, -0.0614471435546875, -0.0545654296875, -0.0476837158203125, -0.040802001953125, -0.0339202880859375, -0.02703857421875, -0.0201568603515625, -0.013275146484375, -0.0063934326171875, 0.00048828125, 0.0073699951171875, 0.014251708984375, 0.0211334228515625, 0.02801513671875, 0.0348968505859375, 0.041778564453125, 0.0486602783203125, 0.0555419921875, 0.0624237060546875, 0.069305419921875, 0.0761871337890625, 0.08306884765625, 0.0899505615234375, 0.096832275390625, 0.1037139892578125, 0.110595703125, 0.1174774169921875, 0.124359130859375, 0.1312408447265625, 0.13812255859375, 0.1450042724609375, 0.151885986328125, 0.1587677001953125, 0.1656494140625, 0.1725311279296875, 0.179412841796875, 0.1862945556640625, 0.19317626953125, 0.2000579833984375, 0.206939697265625, 0.2138214111328125, 0.220703125]}, "parameters/_orig_mod.model.layers.27._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 8.0, 6.0, 11.0, 9.0, 24.0, 22.0, 31.0, 51.0, 74.0, 110.0, 139.0, 212.0, 335.0, 569.0, 1075.0, 2621.0, 6093.0, 16307.0, 39361.0, 99511.0, 213746.0, 443503.0, 794432.0, 1008105.0, 772388.0, 431754.0, 203075.0, 96241.0, 38717.0, 14650.0, 5870.0, 2639.0, 1115.0, 554.0, 310.0, 191.0, 133.0, 93.0, 59.0, 44.0, 32.0, 19.0, 19.0, 10.0, 9.0, 6.0, 3.0, 2.0, 3.0, 1.0, 2.0, 0.0, 1.0, 1.0, 2.0], "bins": [-0.40234375, -0.38995361328125, -0.3775634765625, -0.36517333984375, -0.352783203125, -0.34039306640625, -0.3280029296875, -0.31561279296875, -0.30322265625, -0.29083251953125, -0.2784423828125, -0.26605224609375, -0.253662109375, -0.24127197265625, -0.2288818359375, -0.21649169921875, -0.2041015625, -0.19171142578125, -0.1793212890625, -0.16693115234375, -0.154541015625, -0.14215087890625, -0.1297607421875, -0.11737060546875, -0.10498046875, -0.09259033203125, -0.0802001953125, -0.06781005859375, -0.055419921875, -0.04302978515625, -0.0306396484375, -0.01824951171875, -0.005859375, 0.00653076171875, 0.0189208984375, 0.03131103515625, 0.043701171875, 0.05609130859375, 0.0684814453125, 0.08087158203125, 0.09326171875, 0.10565185546875, 0.1180419921875, 0.13043212890625, 0.142822265625, 0.15521240234375, 0.1676025390625, 0.17999267578125, 0.1923828125, 0.20477294921875, 0.2171630859375, 0.22955322265625, 0.241943359375, 0.25433349609375, 0.2667236328125, 0.27911376953125, 0.29150390625, 0.30389404296875, 0.3162841796875, 0.32867431640625, 0.341064453125, 0.35345458984375, 0.3658447265625, 0.37823486328125, 0.390625]}, "parameters/_orig_mod.model.layers.27._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 3.0, 7.0, 7.0, 11.0, 14.0, 24.0, 41.0, 72.0, 89.0, 163.0, 273.0, 587.0, 1258.0, 3619.0, 10006.0, 28470.0, 61340.0, 110299.0, 158275.0, 167017.0, 139414.0, 87075.0, 43190.0, 18177.0, 6217.0, 2168.0, 857.0, 411.0, 227.0, 130.0, 97.0, 51.0, 26.0, 19.0, 14.0, 9.0, 7.0, 4.0, 2.0, 2.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.1884765625, -0.1818084716796875, -0.175140380859375, -0.1684722900390625, -0.16180419921875, -0.1551361083984375, -0.148468017578125, -0.1417999267578125, -0.1351318359375, -0.1284637451171875, -0.121795654296875, -0.1151275634765625, -0.10845947265625, -0.1017913818359375, -0.095123291015625, -0.0884552001953125, -0.081787109375, -0.0751190185546875, -0.068450927734375, -0.0617828369140625, -0.05511474609375, -0.0484466552734375, -0.041778564453125, -0.0351104736328125, -0.0284423828125, -0.0217742919921875, -0.015106201171875, -0.0084381103515625, -0.00177001953125, 0.0048980712890625, 0.011566162109375, 0.0182342529296875, 0.02490234375, 0.0315704345703125, 0.038238525390625, 0.0449066162109375, 0.05157470703125, 0.0582427978515625, 0.064910888671875, 0.0715789794921875, 0.0782470703125, 0.0849151611328125, 0.091583251953125, 0.0982513427734375, 0.10491943359375, 0.1115875244140625, 0.118255615234375, 0.1249237060546875, 0.131591796875, 0.1382598876953125, 0.144927978515625, 0.1515960693359375, 0.15826416015625, 0.1649322509765625, 0.171600341796875, 0.1782684326171875, 0.1849365234375, 0.1916046142578125, 0.198272705078125, 0.2049407958984375, 0.21160888671875, 0.2182769775390625, 0.224945068359375, 0.2316131591796875, 0.23828125]}, "parameters/_orig_mod.model.layers.28._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 0.0, 6.0, 8.0, 10.0, 15.0, 16.0, 26.0, 24.0, 46.0, 75.0, 111.0, 192.0, 365.0, 668.0, 1606.0, 4154.0, 13914.0, 42531.0, 131441.0, 347645.0, 872846.0, 1790812.0, 3134250.0, 3758152.0, 3273305.0, 1876595.0, 935423.0, 378377.0, 143535.0, 47674.0, 15188.0, 4653.0, 1851.0, 678.0, 405.0, 193.0, 140.0, 82.0, 58.0, 49.0, 39.0, 11.0, 7.0, 11.0, 4.0, 6.0, 5.0, 3.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0], "bins": [-0.265625, -0.2572021484375, -0.248779296875, -0.2403564453125, -0.23193359375, -0.2235107421875, -0.215087890625, -0.2066650390625, -0.1982421875, -0.1898193359375, -0.181396484375, -0.1729736328125, -0.16455078125, -0.1561279296875, -0.147705078125, -0.1392822265625, -0.130859375, -0.1224365234375, -0.114013671875, -0.1055908203125, -0.09716796875, -0.0887451171875, -0.080322265625, -0.0718994140625, -0.0634765625, -0.0550537109375, -0.046630859375, -0.0382080078125, -0.02978515625, -0.0213623046875, -0.012939453125, -0.0045166015625, 0.00390625, 0.0123291015625, 0.020751953125, 0.0291748046875, 0.03759765625, 0.0460205078125, 0.054443359375, 0.0628662109375, 0.0712890625, 0.0797119140625, 0.088134765625, 0.0965576171875, 0.10498046875, 0.1134033203125, 0.121826171875, 0.1302490234375, 0.138671875, 0.1470947265625, 0.155517578125, 0.1639404296875, 0.17236328125, 0.1807861328125, 0.189208984375, 0.1976318359375, 0.2060546875, 0.2144775390625, 0.222900390625, 0.2313232421875, 0.23974609375, 0.2481689453125, 0.256591796875, 0.2650146484375, 0.2734375]}, "parameters/_orig_mod.model.layers.28._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 1.0, 3.0, 2.0, 2.0, 9.0, 7.0, 12.0, 24.0, 14.0, 31.0, 37.0, 69.0, 91.0, 115.0, 171.0, 230.0, 313.0, 532.0, 889.0, 1867.0, 3728.0, 8309.0, 18999.0, 41329.0, 83336.0, 154990.0, 281756.0, 431161.0, 610948.0, 680938.0, 656039.0, 494506.0, 329768.0, 191660.0, 103710.0, 52824.0, 25026.0, 10662.0, 4913.0, 2203.0, 1141.0, 658.0, 361.0, 283.0, 173.0, 125.0, 87.0, 67.0, 52.0, 44.0, 25.0, 20.0, 10.0, 9.0, 8.0, 4.0, 2.0, 1.0, 2.0, 1.0, 3.0, 1.0], "bins": [-0.279296875, -0.270477294921875, -0.26165771484375, -0.252838134765625, -0.2440185546875, -0.235198974609375, -0.22637939453125, -0.217559814453125, -0.208740234375, -0.199920654296875, -0.19110107421875, -0.182281494140625, -0.1734619140625, -0.164642333984375, -0.15582275390625, -0.147003173828125, -0.13818359375, -0.129364013671875, -0.12054443359375, -0.111724853515625, -0.1029052734375, -0.094085693359375, -0.08526611328125, -0.076446533203125, -0.067626953125, -0.058807373046875, -0.04998779296875, -0.041168212890625, -0.0323486328125, -0.023529052734375, -0.01470947265625, -0.005889892578125, 0.0029296875, 0.011749267578125, 0.02056884765625, 0.029388427734375, 0.0382080078125, 0.047027587890625, 0.05584716796875, 0.064666748046875, 0.073486328125, 0.082305908203125, 0.09112548828125, 0.099945068359375, 0.1087646484375, 0.117584228515625, 0.12640380859375, 0.135223388671875, 0.14404296875, 0.152862548828125, 0.16168212890625, 0.170501708984375, 0.1793212890625, 0.188140869140625, 0.19696044921875, 0.205780029296875, 0.214599609375, 0.223419189453125, 0.23223876953125, 0.241058349609375, 0.2498779296875, 0.258697509765625, 0.26751708984375, 0.276336669921875, 0.28515625]}, "parameters/_orig_mod.model.layers.28._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 4.0, 0.0, 7.0, 7.0, 19.0, 41.0, 62.0, 104.0, 132.0, 279.0, 410.0, 579.0, 967.0, 1329.0, 2074.0, 2975.0, 3666.0, 5442.0, 7245.0, 9774.0, 13251.0, 15886.0, 19699.0, 23972.0, 29082.0, 34848.0, 39297.0, 46654.0, 48329.0, 53716.0, 54700.0, 55394.0, 53135.0, 49872.0, 46721.0, 39650.0, 36314.0, 30579.0, 24977.0, 20573.0, 16340.0, 14189.0, 10220.0, 7700.0, 5903.0, 3906.0, 3142.0, 2272.0, 1362.0, 1021.0, 631.0, 483.0, 282.0, 171.0, 122.0, 56.0, 50.0, 26.0, 15.0, 11.0, 6.0, 1.0, 4.0], "bins": [-0.0751953125, -0.07285308837890625, -0.0705108642578125, -0.06816864013671875, -0.065826416015625, -0.06348419189453125, -0.0611419677734375, -0.05879974365234375, -0.05645751953125, -0.05411529541015625, -0.0517730712890625, -0.04943084716796875, -0.047088623046875, -0.04474639892578125, -0.0424041748046875, -0.04006195068359375, -0.0377197265625, -0.03537750244140625, -0.0330352783203125, -0.03069305419921875, -0.028350830078125, -0.02600860595703125, -0.0236663818359375, -0.02132415771484375, -0.01898193359375, -0.01663970947265625, -0.0142974853515625, -0.01195526123046875, -0.009613037109375, -0.00727081298828125, -0.0049285888671875, -0.00258636474609375, -0.000244140625, 0.00209808349609375, 0.0044403076171875, 0.00678253173828125, 0.009124755859375, 0.01146697998046875, 0.0138092041015625, 0.01615142822265625, 0.01849365234375, 0.02083587646484375, 0.0231781005859375, 0.02552032470703125, 0.027862548828125, 0.03020477294921875, 0.0325469970703125, 0.03488922119140625, 0.0372314453125, 0.03957366943359375, 0.0419158935546875, 0.04425811767578125, 0.046600341796875, 0.04894256591796875, 0.0512847900390625, 0.05362701416015625, 0.05596923828125, 0.05831146240234375, 0.0606536865234375, 0.06299591064453125, 0.065338134765625, 0.06768035888671875, 0.0700225830078125, 0.07236480712890625, 0.07470703125]}, "parameters/_orig_mod.model.layers.29._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 4.0, 4.0, 10.0, 6.0, 6.0, 17.0, 21.0, 40.0, 74.0, 84.0, 146.0, 268.0, 381.0, 647.0, 1246.0, 2476.0, 6504.0, 20480.0, 64968.0, 189888.0, 523140.0, 1228650.0, 2434630.0, 3698263.0, 3736194.0, 2598222.0, 1350650.0, 588797.0, 218252.0, 75488.0, 24150.0, 7518.0, 2865.0, 1312.0, 732.0, 364.0, 278.0, 158.0, 87.0, 62.0, 38.0, 31.0, 14.0, 14.0, 4.0, 13.0, 7.0, 1.0, 2.0, 2.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.291015625, -0.282470703125, -0.27392578125, -0.265380859375, -0.2568359375, -0.248291015625, -0.23974609375, -0.231201171875, -0.22265625, -0.214111328125, -0.20556640625, -0.197021484375, -0.1884765625, -0.179931640625, -0.17138671875, -0.162841796875, -0.154296875, -0.145751953125, -0.13720703125, -0.128662109375, -0.1201171875, -0.111572265625, -0.10302734375, -0.094482421875, -0.0859375, -0.077392578125, -0.06884765625, -0.060302734375, -0.0517578125, -0.043212890625, -0.03466796875, -0.026123046875, -0.017578125, -0.009033203125, -0.00048828125, 0.008056640625, 0.0166015625, 0.025146484375, 0.03369140625, 0.042236328125, 0.05078125, 0.059326171875, 0.06787109375, 0.076416015625, 0.0849609375, 0.093505859375, 0.10205078125, 0.110595703125, 0.119140625, 0.127685546875, 0.13623046875, 0.144775390625, 0.1533203125, 0.161865234375, 0.17041015625, 0.178955078125, 0.1875, 0.196044921875, 0.20458984375, 0.213134765625, 0.2216796875, 0.230224609375, 0.23876953125, 0.247314453125, 0.255859375]}, "parameters/_orig_mod.model.layers.29._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 4.0, 5.0, 4.0, 6.0, 8.0, 16.0, 23.0, 34.0, 41.0, 77.0, 77.0, 157.0, 222.0, 400.0, 996.0, 2354.0, 6023.0, 15577.0, 40300.0, 98757.0, 209001.0, 428759.0, 696699.0, 868503.0, 804314.0, 521813.0, 277074.0, 130974.0, 55250.0, 22016.0, 8693.0, 3424.0, 1289.0, 608.0, 287.0, 139.0, 108.0, 61.0, 66.0, 33.0, 28.0, 27.0, 12.0, 14.0, 9.0, 8.0, 2.0, 3.0, 2.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.404296875, -0.39263916015625, -0.3809814453125, -0.36932373046875, -0.357666015625, -0.34600830078125, -0.3343505859375, -0.32269287109375, -0.31103515625, -0.29937744140625, -0.2877197265625, -0.27606201171875, -0.264404296875, -0.25274658203125, -0.2410888671875, -0.22943115234375, -0.2177734375, -0.20611572265625, -0.1944580078125, -0.18280029296875, -0.171142578125, -0.15948486328125, -0.1478271484375, -0.13616943359375, -0.12451171875, -0.11285400390625, -0.1011962890625, -0.08953857421875, -0.077880859375, -0.06622314453125, -0.0545654296875, -0.04290771484375, -0.03125, -0.01959228515625, -0.0079345703125, 0.00372314453125, 0.015380859375, 0.02703857421875, 0.0386962890625, 0.05035400390625, 0.06201171875, 0.07366943359375, 0.0853271484375, 0.09698486328125, 0.108642578125, 0.12030029296875, 0.1319580078125, 0.14361572265625, 0.1552734375, 0.16693115234375, 0.1785888671875, 0.19024658203125, 0.201904296875, 0.21356201171875, 0.2252197265625, 0.23687744140625, 0.24853515625, 0.26019287109375, 0.2718505859375, 0.28350830078125, 0.295166015625, 0.30682373046875, 0.3184814453125, 0.33013916015625, 0.341796875]}, "parameters/_orig_mod.model.layers.29._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 8.0, 16.0, 20.0, 30.0, 59.0, 106.0, 188.0, 264.0, 467.0, 671.0, 1187.0, 1630.0, 2407.0, 3776.0, 5074.0, 7085.0, 10054.0, 13018.0, 16881.0, 20944.0, 25652.0, 31012.0, 36743.0, 41067.0, 47683.0, 51310.0, 53667.0, 55123.0, 54915.0, 53387.0, 49114.0, 46951.0, 40545.0, 35619.0, 29694.0, 24582.0, 20087.0, 16108.0, 12150.0, 9208.0, 6677.0, 4615.0, 3485.0, 2143.0, 1489.0, 1071.0, 638.0, 424.0, 243.0, 151.0, 107.0, 55.0, 31.0, 22.0, 9.0, 8.0, 1.0, 1.0, 1.0, 3.0], "bins": [-0.07373046875, -0.07141876220703125, -0.0691070556640625, -0.06679534912109375, -0.064483642578125, -0.06217193603515625, -0.0598602294921875, -0.05754852294921875, -0.05523681640625, -0.05292510986328125, -0.0506134033203125, -0.04830169677734375, -0.045989990234375, -0.04367828369140625, -0.0413665771484375, -0.03905487060546875, -0.0367431640625, -0.03443145751953125, -0.0321197509765625, -0.02980804443359375, -0.027496337890625, -0.02518463134765625, -0.0228729248046875, -0.02056121826171875, -0.01824951171875, -0.01593780517578125, -0.0136260986328125, -0.01131439208984375, -0.009002685546875, -0.00669097900390625, -0.0043792724609375, -0.00206756591796875, 0.000244140625, 0.00255584716796875, 0.0048675537109375, 0.00717926025390625, 0.009490966796875, 0.01180267333984375, 0.0141143798828125, 0.01642608642578125, 0.01873779296875, 0.02104949951171875, 0.0233612060546875, 0.02567291259765625, 0.027984619140625, 0.03029632568359375, 0.0326080322265625, 0.03491973876953125, 0.0372314453125, 0.03954315185546875, 0.0418548583984375, 0.04416656494140625, 0.046478271484375, 0.04878997802734375, 0.0511016845703125, 0.05341339111328125, 0.05572509765625, 0.05803680419921875, 0.0603485107421875, 0.06266021728515625, 0.064971923828125, 0.06728363037109375, 0.0695953369140625, 0.07190704345703125, 0.07421875]}, "parameters/_orig_mod.model.layers.30._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 2.0, 3.0, 2.0, 6.0, 8.0, 22.0, 25.0, 35.0, 46.0, 66.0, 108.0, 168.0, 263.0, 481.0, 778.0, 1379.0, 2612.0, 5756.0, 14643.0, 37923.0, 93887.0, 237117.0, 573647.0, 1199919.0, 2251135.0, 3698384.0, 3734848.0, 2524827.0, 1303882.0, 645519.0, 266525.0, 111239.0, 42343.0, 16291.0, 6778.0, 2869.0, 1527.0, 848.0, 446.0, 314.0, 212.0, 123.0, 81.0, 33.0, 34.0, 19.0, 13.0, 7.0, 8.0, 4.0, 3.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.22265625, -0.214996337890625, -0.20733642578125, -0.199676513671875, -0.1920166015625, -0.184356689453125, -0.17669677734375, -0.169036865234375, -0.161376953125, -0.153717041015625, -0.14605712890625, -0.138397216796875, -0.1307373046875, -0.123077392578125, -0.11541748046875, -0.107757568359375, -0.10009765625, -0.092437744140625, -0.08477783203125, -0.077117919921875, -0.0694580078125, -0.061798095703125, -0.05413818359375, -0.046478271484375, -0.038818359375, -0.031158447265625, -0.02349853515625, -0.015838623046875, -0.0081787109375, -0.000518798828125, 0.00714111328125, 0.014801025390625, 0.0224609375, 0.030120849609375, 0.03778076171875, 0.045440673828125, 0.0531005859375, 0.060760498046875, 0.06842041015625, 0.076080322265625, 0.083740234375, 0.091400146484375, 0.09906005859375, 0.106719970703125, 0.1143798828125, 0.122039794921875, 0.12969970703125, 0.137359619140625, 0.14501953125, 0.152679443359375, 0.16033935546875, 0.167999267578125, 0.1756591796875, 0.183319091796875, 0.19097900390625, 0.198638916015625, 0.206298828125, 0.213958740234375, 0.22161865234375, 0.229278564453125, 0.2369384765625, 0.244598388671875, 0.25225830078125, 0.259918212890625, 0.267578125]}, "parameters/_orig_mod.model.layers.30._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [6.0, 2.0, 9.0, 4.0, 5.0, 9.0, 24.0, 13.0, 33.0, 40.0, 63.0, 79.0, 90.0, 122.0, 146.0, 244.0, 270.0, 405.0, 619.0, 1013.0, 1809.0, 3609.0, 8195.0, 19555.0, 50472.0, 119564.0, 266787.0, 522608.0, 845943.0, 906082.0, 716776.0, 398262.0, 188815.0, 83227.0, 33510.0, 13338.0, 5604.0, 2673.0, 1434.0, 836.0, 494.0, 420.0, 274.0, 199.0, 152.0, 140.0, 84.0, 68.0, 38.0, 38.0, 30.0, 23.0, 16.0, 16.0, 3.0, 1.0, 4.0, 1.0, 1.0, 1.0, 3.0, 1.0, 1.0, 1.0], "bins": [-0.3046875, -0.294281005859375, -0.28387451171875, -0.273468017578125, -0.2630615234375, -0.252655029296875, -0.24224853515625, -0.231842041015625, -0.221435546875, -0.211029052734375, -0.20062255859375, -0.190216064453125, -0.1798095703125, -0.169403076171875, -0.15899658203125, -0.148590087890625, -0.13818359375, -0.127777099609375, -0.11737060546875, -0.106964111328125, -0.0965576171875, -0.086151123046875, -0.07574462890625, -0.065338134765625, -0.054931640625, -0.044525146484375, -0.03411865234375, -0.023712158203125, -0.0133056640625, -0.002899169921875, 0.00750732421875, 0.017913818359375, 0.0283203125, 0.038726806640625, 0.04913330078125, 0.059539794921875, 0.0699462890625, 0.080352783203125, 0.09075927734375, 0.101165771484375, 0.111572265625, 0.121978759765625, 0.13238525390625, 0.142791748046875, 0.1531982421875, 0.163604736328125, 0.17401123046875, 0.184417724609375, 0.19482421875, 0.205230712890625, 0.21563720703125, 0.226043701171875, 0.2364501953125, 0.246856689453125, 0.25726318359375, 0.267669677734375, 0.278076171875, 0.288482666015625, 0.29888916015625, 0.309295654296875, 0.3197021484375, 0.330108642578125, 0.34051513671875, 0.350921630859375, 0.361328125]}, "parameters/_orig_mod.model.layers.30._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 1.0, 3.0, 3.0, 14.0, 31.0, 37.0, 83.0, 127.0, 202.0, 288.0, 511.0, 774.0, 1187.0, 1748.0, 2413.0, 3681.0, 5092.0, 6814.0, 9305.0, 11969.0, 15086.0, 19818.0, 23138.0, 28934.0, 33526.0, 39368.0, 44132.0, 50416.0, 55780.0, 57182.0, 58091.0, 56806.0, 51296.0, 45290.0, 40517.0, 36193.0, 28940.0, 25699.0, 20297.0, 16973.0, 13373.0, 9189.0, 7467.0, 5559.0, 3954.0, 2858.0, 1900.0, 1332.0, 911.0, 543.0, 320.0, 219.0, 104.0, 70.0, 48.0, 25.0, 21.0, 4.0, 6.0, 3.0, 4.0, 0.0, 2.0], "bins": [-0.08984375, -0.0869598388671875, -0.084075927734375, -0.0811920166015625, -0.07830810546875, -0.0754241943359375, -0.072540283203125, -0.0696563720703125, -0.0667724609375, -0.0638885498046875, -0.061004638671875, -0.0581207275390625, -0.05523681640625, -0.0523529052734375, -0.049468994140625, -0.0465850830078125, -0.043701171875, -0.0408172607421875, -0.037933349609375, -0.0350494384765625, -0.03216552734375, -0.0292816162109375, -0.026397705078125, -0.0235137939453125, -0.0206298828125, -0.0177459716796875, -0.014862060546875, -0.0119781494140625, -0.00909423828125, -0.0062103271484375, -0.003326416015625, -0.0004425048828125, 0.00244140625, 0.0053253173828125, 0.008209228515625, 0.0110931396484375, 0.01397705078125, 0.0168609619140625, 0.019744873046875, 0.0226287841796875, 0.0255126953125, 0.0283966064453125, 0.031280517578125, 0.0341644287109375, 0.03704833984375, 0.0399322509765625, 0.042816162109375, 0.0457000732421875, 0.048583984375, 0.0514678955078125, 0.054351806640625, 0.0572357177734375, 0.06011962890625, 0.0630035400390625, 0.065887451171875, 0.0687713623046875, 0.0716552734375, 0.0745391845703125, 0.077423095703125, 0.0803070068359375, 0.08319091796875, 0.0860748291015625, 0.088958740234375, 0.0918426513671875, 0.0947265625]}, "parameters/_orig_mod.model.layers.31._fsdp_wrapped_module.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 8.0, 4.0, 5.0, 8.0, 11.0, 13.0, 21.0, 30.0, 44.0, 50.0, 71.0, 117.0, 181.0, 305.0, 612.0, 1162.0, 3007.0, 8618.0, 34345.0, 140484.0, 546428.0, 1808674.0, 4436224.0, 5373107.0, 2974987.0, 1060631.0, 288481.0, 73182.0, 17572.0, 5035.0, 1843.0, 803.0, 469.0, 240.0, 151.0, 79.0, 66.0, 45.0, 32.0, 17.0, 11.0, 6.0, 6.0, 3.0, 4.0, 2.0, 1.0, 2.0, 2.0, 2.0, 2.0, 1.0, 2.0, 1.0], "bins": [-0.4140625, -0.401611328125, -0.38916015625, -0.376708984375, -0.3642578125, -0.351806640625, -0.33935546875, -0.326904296875, -0.314453125, -0.302001953125, -0.28955078125, -0.277099609375, -0.2646484375, -0.252197265625, -0.23974609375, -0.227294921875, -0.21484375, -0.202392578125, -0.18994140625, -0.177490234375, -0.1650390625, -0.152587890625, -0.14013671875, -0.127685546875, -0.115234375, -0.102783203125, -0.09033203125, -0.077880859375, -0.0654296875, -0.052978515625, -0.04052734375, -0.028076171875, -0.015625, -0.003173828125, 0.00927734375, 0.021728515625, 0.0341796875, 0.046630859375, 0.05908203125, 0.071533203125, 0.083984375, 0.096435546875, 0.10888671875, 0.121337890625, 0.1337890625, 0.146240234375, 0.15869140625, 0.171142578125, 0.18359375, 0.196044921875, 0.20849609375, 0.220947265625, 0.2333984375, 0.245849609375, 0.25830078125, 0.270751953125, 0.283203125, 0.295654296875, 0.30810546875, 0.320556640625, 0.3330078125, 0.345458984375, 0.35791015625, 0.370361328125, 0.3828125]}, "parameters/_orig_mod.model.layers.31._fsdp_wrapped_module.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 4.0, 3.0, 4.0, 9.0, 5.0, 10.0, 18.0, 28.0, 31.0, 38.0, 68.0, 103.0, 110.0, 191.0, 300.0, 510.0, 1107.0, 2489.0, 8208.0, 28691.0, 100774.0, 279654.0, 658738.0, 1067255.0, 1045452.0, 613708.0, 260027.0, 89749.0, 25291.0, 7078.0, 2397.0, 941.0, 471.0, 247.0, 201.0, 107.0, 78.0, 56.0, 44.0, 26.0, 22.0, 11.0, 13.0, 6.0, 4.0, 7.0, 1.0, 6.0, 2.0, 1.0, 2.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.373046875, -0.358673095703125, -0.34429931640625, -0.329925537109375, -0.3155517578125, -0.301177978515625, -0.28680419921875, -0.272430419921875, -0.258056640625, -0.243682861328125, -0.22930908203125, -0.214935302734375, -0.2005615234375, -0.186187744140625, -0.17181396484375, -0.157440185546875, -0.14306640625, -0.128692626953125, -0.11431884765625, -0.099945068359375, -0.0855712890625, -0.071197509765625, -0.05682373046875, -0.042449951171875, -0.028076171875, -0.013702392578125, 0.00067138671875, 0.015045166015625, 0.0294189453125, 0.043792724609375, 0.05816650390625, 0.072540283203125, 0.0869140625, 0.101287841796875, 0.11566162109375, 0.130035400390625, 0.1444091796875, 0.158782958984375, 0.17315673828125, 0.187530517578125, 0.201904296875, 0.216278076171875, 0.23065185546875, 0.245025634765625, 0.2593994140625, 0.273773193359375, 0.28814697265625, 0.302520751953125, 0.31689453125, 0.331268310546875, 0.34564208984375, 0.360015869140625, 0.3743896484375, 0.388763427734375, 0.40313720703125, 0.417510986328125, 0.431884765625, 0.446258544921875, 0.46063232421875, 0.475006103515625, 0.4893798828125, 0.503753662109375, 0.51812744140625, 0.532501220703125, 0.546875]}, "parameters/_orig_mod.model.layers.31._fsdp_wrapped_module.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 10.0, 13.0, 22.0, 48.0, 69.0, 96.0, 219.0, 350.0, 658.0, 1118.0, 2028.0, 2932.0, 4548.0, 6930.0, 10565.0, 13891.0, 20579.0, 26093.0, 34123.0, 41000.0, 50057.0, 58859.0, 64849.0, 67794.0, 68208.0, 67215.0, 60595.0, 53381.0, 45279.0, 35936.0, 29372.0, 22109.0, 16744.0, 11150.0, 8428.0, 5312.0, 3520.0, 2288.0, 1385.0, 798.0, 473.0, 245.0, 173.0, 99.0, 41.0, 25.0, 24.0, 8.0, 7.0, 3.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.10498046875, -0.10150146484375, -0.0980224609375, -0.09454345703125, -0.091064453125, -0.08758544921875, -0.0841064453125, -0.08062744140625, -0.0771484375, -0.07366943359375, -0.0701904296875, -0.06671142578125, -0.063232421875, -0.05975341796875, -0.0562744140625, -0.05279541015625, -0.04931640625, -0.04583740234375, -0.0423583984375, -0.03887939453125, -0.035400390625, -0.03192138671875, -0.0284423828125, -0.02496337890625, -0.021484375, -0.01800537109375, -0.0145263671875, -0.01104736328125, -0.007568359375, -0.00408935546875, -0.0006103515625, 0.00286865234375, 0.00634765625, 0.00982666015625, 0.0133056640625, 0.01678466796875, 0.020263671875, 0.02374267578125, 0.0272216796875, 0.03070068359375, 0.0341796875, 0.03765869140625, 0.0411376953125, 0.04461669921875, 0.048095703125, 0.05157470703125, 0.0550537109375, 0.05853271484375, 0.06201171875, 0.06549072265625, 0.0689697265625, 0.07244873046875, 0.075927734375, 0.07940673828125, 0.0828857421875, 0.08636474609375, 0.08984375, 0.09332275390625, 0.0968017578125, 0.10028076171875, 0.103759765625, 0.10723876953125, 0.1107177734375, 0.11419677734375, 0.11767578125]}}
\ No newline at end of file
diff --git a/wandb/run-20240512_103207-vfqtzg1a/logs/debug-internal.log b/wandb/run-20240512_103207-vfqtzg1a/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..617b8dcb188d6664c0664e7938d09e158d9b8c78
--- /dev/null
+++ b/wandb/run-20240512_103207-vfqtzg1a/logs/debug-internal.log
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40d1c2215753b9e9898035555d9a1e1f853788f57b48e53fad61fdea9300e0a8
+size 13290638
diff --git a/wandb/run-20240512_103207-vfqtzg1a/logs/debug.log b/wandb/run-20240512_103207-vfqtzg1a/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..16272d9e3c08dd11267532303d78e27802eb5944
--- /dev/null
+++ b/wandb/run-20240512_103207-vfqtzg1a/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-12 10:32:07,562 INFO MainThread:44754 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-12 10:32:07,562 INFO MainThread:44754 [wandb_setup.py:_flush():76] Configure stats pid to 44754
+2024-05-12 10:32:07,562 INFO MainThread:44754 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-12 10:32:07,562 INFO MainThread:44754 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-12 10:32:07,562 WARNING MainThread:44754 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-12 10:32:07,562 INFO MainThread:44754 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-12 10:32:07,562 INFO MainThread:44754 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-12 10:32:07,562 WARNING MainThread:44754 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-12 10:32:07,562 INFO MainThread:44754 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-12 10:32:07,562 INFO MainThread:44754 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-12 10:32:07,563 INFO MainThread:44754 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240512_103207-vfqtzg1a/logs/debug.log
+2024-05-12 10:32:07,563 INFO MainThread:44754 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240512_103207-vfqtzg1a/logs/debug-internal.log
+2024-05-12 10:32:07,563 INFO MainThread:44754 [wandb_init.py:init():560] calling init triggers
+2024-05-12 10:32:07,563 INFO MainThread:44754 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-12 10:32:07,563 INFO MainThread:44754 [wandb_init.py:init():610] starting backend
+2024-05-12 10:32:07,563 INFO MainThread:44754 [wandb_init.py:init():614] setting up manager
+2024-05-12 10:32:07,565 INFO MainThread:44754 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-12 10:32:07,567 INFO MainThread:44754 [wandb_init.py:init():622] backend started and connected
+2024-05-12 10:32:07,580 INFO MainThread:44754 [wandb_init.py:init():711] updated telemetry
+2024-05-12 10:32:07,590 INFO MainThread:44754 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-12 10:32:08,403 INFO MainThread:44754 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-12 10:32:08,481 INFO MainThread:44754 [wandb_run.py:_on_init():2405] got version response
+2024-05-12 10:32:08,482 INFO MainThread:44754 [wandb_init.py:init():795] starting run threads in backend
+2024-05-12 10:32:09,408 INFO MainThread:44754 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-12 10:32:09,408 INFO MainThread:44754 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-12 10:32:09,409 INFO MainThread:44754 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-12 10:32:09,409 INFO MainThread:44754 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-12 10:32:09,410 INFO MainThread:44754 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-12 10:32:09,413 INFO MainThread:44754 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './buzz-3b-small-v0.6', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 2, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {'num_cycles': 3}, 'warmup_ratio': 0.0, 'warmup_steps': 287, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './buzz-3b-small-v0.6/runs/May12_10-31-45_a9815c79a99f', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './buzz-3b-small-v0.6', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': ['full_shard', 'auto_wrap'], 'fsdp_min_num_params': 0, 'fsdp_config': {'fsdp_limit_all_gathers': True, 'fsdp_sync_module_states': True, 'fsdp_offload_params': True, 'fsdp_use_orig_params': False, 'fsdp_cpu_ram_efficient_loading': False, 'fsdp_auto_wrap_policy': 'TRANSFORMER_BASED_WRAP', 'fsdp_transformer_layer_cls_to_wrap': 'LlamaDecoderLayer', 'fsdp_state_dict_type': 'FULL_STATE_DICT', 'fsdp_sharding_strategy': 'FULL_SHARD', 'fsdp_forward_prefetch': 'FORWARD_PRE', 'fsdp_backward_prefetch': 'BACKWARD_PRE', 'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.3', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.92, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 1, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-12 10:32:09,415 INFO MainThread:44754 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240512_103207-vfqtzg1a/run-vfqtzg1a.wandb b/wandb/run-20240512_103207-vfqtzg1a/run-vfqtzg1a.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..af2f17f411c25f5cf9feac6ebd8920de0e4a2949
--- /dev/null
+++ b/wandb/run-20240512_103207-vfqtzg1a/run-vfqtzg1a.wandb
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1fd10125dc3a57897e9dcb92648cdee0aa1bc5e11a245c31705ed9946932e18e
+size 46037294
diff --git a/wandb/run-20240515_121108-gswsvc05/files/conda-environment.yaml b/wandb/run-20240515_121108-gswsvc05/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_121108-gswsvc05/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_121108-gswsvc05/files/config.yaml b/wandb/run-20240515_121108-gswsvc05/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..5cecbfe514d490a4bc067fb75ca4714843c0af00
--- /dev/null
+++ b/wandb/run-20240515_121108-gswsvc05/files/config.yaml
@@ -0,0 +1,739 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715775068
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 1
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.0.dev0
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./pyggel
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: false
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: 'no'
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 1
+per_device_eval_batch_size:
+ desc: null
+ value: 1
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 2
+eval_accumulation_steps:
+ desc: null
+ value: 2
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 5.0e-05
+weight_decay:
+ desc: null
+ value: 0.01
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.95
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 0.6
+num_train_epochs:
+ desc: null
+ value: 3
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value:
+ num_cycles: 3
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 2873
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./pyggel/runs/May15_12-10-38_d2827b6ba082
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.022222222222222223
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: false
+eval_steps:
+ desc: null
+ value: null
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./pyggel
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value: []
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: deepspeed_configs/zero1.json
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_torch
+optim_args:
+ desc: null
+ value: null
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: false
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.4
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: 15.0
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: true
+sample_packing_efficiency:
+ desc: null
+ value: 0.92
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 1
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240515_121108-gswsvc05/files/output.log b/wandb/run-20240515_121108-gswsvc05/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..57c25e5d93fb56af0d56928514b685aff258f93d
--- /dev/null
+++ b/wandb/run-20240515_121108-gswsvc05/files/output.log
@@ -0,0 +1,49 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/9378 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[2024-05-15 12:11:09,979] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:13217] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-15 12:11:10,973] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:13217] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+[2024-05-15 12:11:11,895] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:13217] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+Traceback (most recent call last):
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/runpy.py", line 196, in _run_module_as_main
+ return _run_code(code, main_globals, None,
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/runpy.py", line 86, in _run_code
+ exec(code, run_globals)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 70, in
+ fire.Fire(do_cli)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 143, in Fire
+ component_trace = _Fire(component, args, parsed_flag_args, context, name)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 477, in _Fire
+ component, remaining_args = _CallAndUpdateTrace(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 693, in _CallAndUpdateTrace
+ component = fn(*varargs, **kwargs)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 38, in do_cli
+ return do_train(parsed_cfg, parsed_cli_args)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 66, in do_train
+ return train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ File "/workspace/disk1/axolotl/src/axolotl/train.py", line 170, in train
+ trainer.train(resume_from_checkpoint=resume_from_checkpoint)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 1828, in train
+ return inner_training_loop(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 2181, in _inner_training_loop
+ tr_loss_step = self.training_step(model, inputs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 3125, in training_step
+ self.accelerator.backward(loss)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/accelerate/accelerator.py", line 1995, in backward
+ self.deepspeed_engine_wrapped.backward(loss, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/accelerate/utils/deepspeed.py", line 166, in backward
+ self.engine.backward(loss, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 15, in wrapped_fn
+ ret_val = func(*args, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 1964, in backward
+ self.optimizer.backward(loss, retain_graph=retain_graph)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/zero/stage_1_and_2.py", line 2040, in backward
+ self.loss_scaler.backward(loss.float(), retain_graph=retain_graph)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 63, in backward
+ scaled_loss.backward(retain_graph=retain_graph)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/_tensor.py", line 492, in backward
+ torch.autograd.backward(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/autograd/__init__.py", line 251, in backward
+ Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
+torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 3.91 GiB. GPU 0 has a total capacty of 44.34 GiB of which 3.89 GiB is free. Process 990626 has 40.44 GiB memory in use. Of the allocated memory 35.88 GiB is allocated by PyTorch, and 3.99 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
\ No newline at end of file
diff --git a/wandb/run-20240515_121108-gswsvc05/files/requirements.txt b/wandb/run-20240515_121108-gswsvc05/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_121108-gswsvc05/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_121108-gswsvc05/files/tmp/axolotl_config_1jy7wv9e.yml b/wandb/run-20240515_121108-gswsvc05/files/tmp/axolotl_config_1jy7wv9e.yml
new file mode 100644
index 0000000000000000000000000000000000000000..11b7175dbf231852e337dc9b3e074818c6ade4d2
--- /dev/null
+++ b/wandb/run-20240515_121108-gswsvc05/files/tmp/axolotl_config_1jy7wv9e.yml
@@ -0,0 +1,67 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+#push_to_hub: true
+hf_use_auth_token: true
+#resume_from_checkpoint: ./buzz-3b-small-v0.6/checkpoint-4598
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 1
+gradient_accumulation_steps: 2
+num_epochs: 3
+learning_rate: 0.00005
+optimizer: adamw_torch
+optim_args:
+adam_beta1: 0.9
+adam_beta2: 0.95
+amsgrad: true
+lr_scheduler: cosine
+lr_scheduler_kwargs:
+ num_cycles: 3
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.333
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero1.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240515_121108-gswsvc05/files/wandb-metadata.json b/wandb/run-20240515_121108-gswsvc05/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..a91347f0a257f79f831b353ed2115f86f5ac11c9
--- /dev/null
+++ b/wandb/run-20240515_121108-gswsvc05/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T12:11:09.040157",
+ "startedAt": "2024-05-15T12:11:08.385576",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1325.1610833333332,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3399.997,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.001,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.36123657226562
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_121108-gswsvc05/files/wandb-summary.json b/wandb/run-20240515_121108-gswsvc05/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..e682bae6b5eaeba8295fd0fffdc51474a259249e
--- /dev/null
+++ b/wandb/run-20240515_121108-gswsvc05/files/wandb-summary.json
@@ -0,0 +1 @@
+{"_wandb": {"runtime": 5}}
\ No newline at end of file
diff --git a/wandb/run-20240515_121108-gswsvc05/logs/debug-internal.log b/wandb/run-20240515_121108-gswsvc05/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..c9cec07ef50b67b374c9791ac45684511865c6b2
--- /dev/null
+++ b/wandb/run-20240515_121108-gswsvc05/logs/debug-internal.log
@@ -0,0 +1,171 @@
+2024-05-15 12:11:08,394 INFO StreamThr :13452 [internal.py:wandb_internal():85] W&B internal server running at pid: 13452, started at: 2024-05-15 12:11:08.393453
+2024-05-15 12:11:08,397 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: status
+2024-05-15 12:11:08,399 INFO WriterThread:13452 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/run-gswsvc05.wandb
+2024-05-15 12:11:08,400 DEBUG SenderThread:13452 [sender.py:send():378] send: header
+2024-05-15 12:11:08,407 DEBUG SenderThread:13452 [sender.py:send():378] send: run
+2024-05-15 12:11:08,811 INFO SenderThread:13452 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files
+2024-05-15 12:11:08,812 INFO SenderThread:13452 [sender.py:_start_run_threads():1123] run started: gswsvc05 with start time 1715775068.393502
+2024-05-15 12:11:08,844 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 12:11:08,845 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: check_version
+2024-05-15 12:11:08,936 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 12:11:08,968 DEBUG HandlerThread:13452 [system_info.py:__init__():26] System info init
+2024-05-15 12:11:08,968 DEBUG HandlerThread:13452 [system_info.py:__init__():41] System info init done
+2024-05-15 12:11:08,968 INFO HandlerThread:13452 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 12:11:08,969 INFO SystemMonitor:13452 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 12:11:08,969 INFO HandlerThread:13452 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 12:11:08,970 INFO SystemMonitor:13452 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 12:11:08,971 INFO SystemMonitor:13452 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 12:11:08,972 INFO SystemMonitor:13452 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 12:11:08,975 INFO SystemMonitor:13452 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 12:11:08,976 INFO SystemMonitor:13452 [interfaces.py:start():188] Started network monitoring
+2024-05-15 12:11:09,040 DEBUG HandlerThread:13452 [system_info.py:probe():150] Probing system
+2024-05-15 12:11:09,043 DEBUG HandlerThread:13452 [system_info.py:_probe_git():135] Probing git
+2024-05-15 12:11:09,052 DEBUG HandlerThread:13452 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 12:11:09,052 DEBUG HandlerThread:13452 [system_info.py:probe():198] Probing system done
+2024-05-15 12:11:09,052 DEBUG HandlerThread:13452 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T12:11:09.040157', 'startedAt': '2024-05-15T12:11:08.385576', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1325.1610833333332, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3399.997, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.001, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.36123657226562}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 12:11:09,053 INFO HandlerThread:13452 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 12:11:09,053 INFO HandlerThread:13452 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 12:11:09,053 DEBUG HandlerThread:13452 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 12:11:09,518 DEBUG HandlerThread:13452 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 12:11:09,519 DEBUG HandlerThread:13452 [system_info.py:_save_code():44] Saving code
+2024-05-15 12:11:09,519 WARNING HandlerThread:13452 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-15 12:11:09,519 DEBUG HandlerThread:13452 [system_info.py:_save_patches():82] Saving git patches
+2024-05-15 12:11:09,544 DEBUG HandlerThread:13452 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-15 12:11:09,547 INFO HandlerThread:13452 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 12:11:09,559 DEBUG SenderThread:13452 [sender.py:send():378] send: files
+2024-05-15 12:11:09,560 INFO SenderThread:13452 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 12:11:09,772 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 12:11:09,773 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:11:09,773 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 12:11:09,775 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:11:09,814 INFO Thread-12 :13452 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/conda-environment.yaml
+2024-05-15 12:11:09,814 INFO Thread-12 :13452 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/requirements.txt
+2024-05-15 12:11:09,814 INFO Thread-12 :13452 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/wandb-metadata.json
+2024-05-15 12:11:09,977 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-15 12:11:10,094 DEBUG SenderThread:13452 [sender.py:send():378] send: telemetry
+2024-05-15 12:11:10,094 DEBUG SenderThread:13452 [sender.py:send():378] send: config
+2024-05-15 12:11:10,096 DEBUG SenderThread:13452 [sender.py:send():378] send: metric
+2024-05-15 12:11:10,096 DEBUG SenderThread:13452 [sender.py:send():378] send: telemetry
+2024-05-15 12:11:10,097 DEBUG SenderThread:13452 [sender.py:send():378] send: telemetry
+2024-05-15 12:11:10,097 DEBUG SenderThread:13452 [sender.py:send():378] send: metric
+2024-05-15 12:11:10,097 WARNING SenderThread:13452 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-15 12:11:10,097 DEBUG SenderThread:13452 [sender.py:send():378] send: telemetry
+2024-05-15 12:11:10,097 DEBUG SenderThread:13452 [sender.py:send():378] send: telemetry
+2024-05-15 12:11:10,097 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: log_artifact
+2024-05-15 12:11:10,213 INFO wandb-upload_0:13452 [upload_job.py:push():130] Uploaded file /tmp/tmpbrspu7nvwandb/56cszmks-wandb-metadata.json
+2024-05-15 12:11:10,812 INFO Thread-12 :13452 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/tmp/axolotl_config_1jy7wv9e.yml
+2024-05-15 12:11:10,813 INFO Thread-12 :13452 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/tmp
+2024-05-15 12:11:11,356 INFO wandb-upload_0:13452 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmpryl06rp1
+2024-05-15 12:11:12,044 INFO SenderThread:13452 [sender.py:send_request_log_artifact():1455] logged artifact config-gswsvc05 - {'id': 'QXJ0aWZhY3Q6ODM3MzUwMTkw', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTE0NzY1NQ==', 'latestArtifact': None}}
+2024-05-15 12:11:12,046 DEBUG SenderThread:13452 [sender.py:send():378] send: telemetry
+2024-05-15 12:11:12,046 DEBUG SenderThread:13452 [sender.py:send():378] send: files
+2024-05-15 12:11:12,046 INFO SenderThread:13452 [sender.py:_save_file():1389] saving file tmp/axolotl_config_1jy7wv9e.yml with policy live
+2024-05-15 12:11:12,436 INFO wandb-upload_0:13452 [upload_job.py:push():130] Uploaded file /tmp/tmpbrspu7nvwandb/z1408wm4-tmp/axolotl_config_1jy7wv9e.yml
+2024-05-15 12:11:12,814 INFO Thread-12 :13452 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/output.log
+2024-05-15 12:11:13,897 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:11:14,591 DEBUG SenderThread:13452 [sender.py:send():378] send: exit
+2024-05-15 12:11:14,591 INFO SenderThread:13452 [sender.py:send_exit():585] handling exit code: 1
+2024-05-15 12:11:14,591 INFO SenderThread:13452 [sender.py:send_exit():587] handling runtime: 5
+2024-05-15 12:11:14,592 INFO SenderThread:13452 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:11:14,592 INFO SenderThread:13452 [sender.py:send_exit():593] send defer
+2024-05-15 12:11:14,593 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:14,593 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 0
+2024-05-15 12:11:14,593 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:14,594 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 0
+2024-05-15 12:11:14,594 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 1
+2024-05-15 12:11:14,594 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:14,594 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 1
+2024-05-15 12:11:14,594 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:14,595 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 1
+2024-05-15 12:11:14,595 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 2
+2024-05-15 12:11:14,595 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:14,595 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 2
+2024-05-15 12:11:14,595 INFO HandlerThread:13452 [system_monitor.py:finish():203] Stopping system monitor
+2024-05-15 12:11:14,595 DEBUG SystemMonitor:13452 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-15 12:11:14,596 DEBUG SystemMonitor:13452 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-15 12:11:14,596 DEBUG SystemMonitor:13452 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-15 12:11:14,597 INFO HandlerThread:13452 [interfaces.py:finish():200] Joined cpu monitor
+2024-05-15 12:11:14,598 INFO HandlerThread:13452 [interfaces.py:finish():200] Joined disk monitor
+2024-05-15 12:11:14,694 INFO HandlerThread:13452 [interfaces.py:finish():200] Joined gpu monitor
+2024-05-15 12:11:14,694 INFO HandlerThread:13452 [interfaces.py:finish():200] Joined memory monitor
+2024-05-15 12:11:14,695 INFO HandlerThread:13452 [interfaces.py:finish():200] Joined network monitor
+2024-05-15 12:11:14,696 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:14,696 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 2
+2024-05-15 12:11:14,696 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 3
+2024-05-15 12:11:14,696 DEBUG SenderThread:13452 [sender.py:send():378] send: stats
+2024-05-15 12:11:14,697 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:14,697 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 3
+2024-05-15 12:11:14,697 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:14,697 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 3
+2024-05-15 12:11:14,698 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 4
+2024-05-15 12:11:14,698 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:14,698 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 4
+2024-05-15 12:11:14,698 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:14,698 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 4
+2024-05-15 12:11:14,699 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 5
+2024-05-15 12:11:14,699 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:14,699 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 5
+2024-05-15 12:11:14,699 DEBUG SenderThread:13452 [sender.py:send():378] send: summary
+2024-05-15 12:11:14,700 INFO SenderThread:13452 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:11:14,700 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:14,701 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 5
+2024-05-15 12:11:14,701 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 6
+2024-05-15 12:11:14,701 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:14,701 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 6
+2024-05-15 12:11:14,701 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:14,702 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 6
+2024-05-15 12:11:14,707 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:11:14,815 INFO Thread-12 :13452 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/output.log
+2024-05-15 12:11:14,816 INFO Thread-12 :13452 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/wandb-summary.json
+2024-05-15 12:11:14,990 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 7
+2024-05-15 12:11:14,990 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:14,991 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 7
+2024-05-15 12:11:14,991 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:14,991 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 7
+2024-05-15 12:11:15,591 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 12:11:15,815 INFO Thread-12 :13452 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/config.yaml
+2024-05-15 12:11:16,081 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 8
+2024-05-15 12:11:16,081 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 12:11:16,082 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:16,082 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 8
+2024-05-15 12:11:16,083 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:16,083 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 8
+2024-05-15 12:11:16,083 INFO SenderThread:13452 [job_builder.py:build():432] Attempting to build job artifact
+2024-05-15 12:11:16,087 INFO SenderThread:13452 [job_builder.py:_get_source_type():565] is repo sourced job
+2024-05-15 12:11:16,087 WARNING SenderThread:13452 [job_builder.py:_log_if_verbose():267] No program path found, not creating job artifact. See https://docs.wandb.ai/guides/launch/create-job
+2024-05-15 12:11:16,087 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 9
+2024-05-15 12:11:16,087 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:16,087 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 9
+2024-05-15 12:11:16,088 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:16,088 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 9
+2024-05-15 12:11:16,088 INFO SenderThread:13452 [dir_watcher.py:finish():358] shutting down directory watcher
+2024-05-15 12:11:16,592 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 12:11:16,817 INFO SenderThread:13452 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/output.log
+2024-05-15 12:11:16,818 INFO SenderThread:13452 [dir_watcher.py:finish():388] scan: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files
+2024-05-15 12:11:16,819 INFO SenderThread:13452 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/wandb-metadata.json wandb-metadata.json
+2024-05-15 12:11:16,819 INFO SenderThread:13452 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/wandb-summary.json wandb-summary.json
+2024-05-15 12:11:16,819 INFO SenderThread:13452 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/requirements.txt requirements.txt
+2024-05-15 12:11:16,819 INFO SenderThread:13452 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/config.yaml config.yaml
+2024-05-15 12:11:16,820 INFO SenderThread:13452 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/conda-environment.yaml conda-environment.yaml
+2024-05-15 12:11:16,820 INFO SenderThread:13452 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/output.log output.log
+2024-05-15 12:11:16,821 INFO SenderThread:13452 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/tmp/axolotl_config_1jy7wv9e.yml tmp/axolotl_config_1jy7wv9e.yml
+2024-05-15 12:11:16,821 INFO SenderThread:13452 [sender.py:transition_state():613] send defer: 10
+2024-05-15 12:11:16,821 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 12:11:16,833 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 12:11:16,834 INFO HandlerThread:13452 [handler.py:handle_request_defer():184] handle defer: 10
+2024-05-15 12:11:16,836 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: defer
+2024-05-15 12:11:16,836 INFO SenderThread:13452 [sender.py:send_request_defer():609] handle sender defer: 10
+2024-05-15 12:11:16,836 INFO SenderThread:13452 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-15 12:11:17,248 INFO wandb-upload_0:13452 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/requirements.txt
+2024-05-15 12:11:17,290 INFO wandb-upload_1:13452 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/wandb-summary.json
+2024-05-15 12:11:17,364 INFO wandb-upload_3:13452 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/conda-environment.yaml
+2024-05-15 12:11:17,391 INFO wandb-upload_4:13452 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/files/output.log
+2024-05-15 12:11:17,407 WARNING StreamThr :13452 [internal.py:is_dead():413] Internal process exiting, parent pid 13217 disappeared
+2024-05-15 12:11:17,408 ERROR StreamThr :13452 [internal.py:wandb_internal():151] Internal process shutdown.
+2024-05-15 12:11:17,593 DEBUG HandlerThread:13452 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 12:11:17,593 INFO HandlerThread:13452 [handler.py:finish():882] shutting down handler
+2024-05-15 12:11:17,594 INFO WriterThread:13452 [datastore.py:close():296] close: /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/run-gswsvc05.wandb
+2024-05-15 12:11:17,594 DEBUG SenderThread:13452 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 12:11:17,595 INFO SenderThread:13452 [sender.py:finish():1545] shutting down sender
+2024-05-15 12:11:17,595 INFO SenderThread:13452 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-15 12:11:17,595 INFO SenderThread:13452 [file_pusher.py:join():175] waiting for file pusher
diff --git a/wandb/run-20240515_121108-gswsvc05/logs/debug.log b/wandb/run-20240515_121108-gswsvc05/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..ef2bc1e286ae2fea81a2981dbe5feb29f9f35f75
--- /dev/null
+++ b/wandb/run-20240515_121108-gswsvc05/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 12:11:08,387 INFO MainThread:13217 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 12:11:08,388 INFO MainThread:13217 [wandb_setup.py:_flush():76] Configure stats pid to 13217
+2024-05-15 12:11:08,388 INFO MainThread:13217 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 12:11:08,388 INFO MainThread:13217 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 12:11:08,388 WARNING MainThread:13217 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 12:11:08,388 INFO MainThread:13217 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 12:11:08,388 INFO MainThread:13217 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 12:11:08,389 WARNING MainThread:13217 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 12:11:08,389 INFO MainThread:13217 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 12:11:08,389 INFO MainThread:13217 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 12:11:08,389 INFO MainThread:13217 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/logs/debug.log
+2024-05-15 12:11:08,389 INFO MainThread:13217 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_121108-gswsvc05/logs/debug-internal.log
+2024-05-15 12:11:08,389 INFO MainThread:13217 [wandb_init.py:init():560] calling init triggers
+2024-05-15 12:11:08,389 INFO MainThread:13217 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 12:11:08,389 INFO MainThread:13217 [wandb_init.py:init():610] starting backend
+2024-05-15 12:11:08,389 INFO MainThread:13217 [wandb_init.py:init():614] setting up manager
+2024-05-15 12:11:08,391 INFO MainThread:13217 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 12:11:08,393 INFO MainThread:13217 [wandb_init.py:init():622] backend started and connected
+2024-05-15 12:11:08,396 INFO MainThread:13217 [wandb_init.py:init():711] updated telemetry
+2024-05-15 12:11:08,406 INFO MainThread:13217 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 12:11:08,843 INFO MainThread:13217 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 12:11:08,923 INFO MainThread:13217 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 12:11:08,923 INFO MainThread:13217 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 12:11:09,773 INFO MainThread:13217 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 12:11:09,773 INFO MainThread:13217 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 12:11:09,773 INFO MainThread:13217 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 12:11:09,774 INFO MainThread:13217 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 12:11:09,774 INFO MainThread:13217 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 12:11:09,777 INFO MainThread:13217 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 2, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {'num_cycles': 3}, 'warmup_ratio': 0.0, 'warmup_steps': 2873, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_12-10-38_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero1.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.92, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 1, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 12:11:09,778 INFO MainThread:13217 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_121108-gswsvc05/run-gswsvc05.wandb b/wandb/run-20240515_121108-gswsvc05/run-gswsvc05.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..8a7313f51f2790b96b8f9f097e0a113a8fa10fe1
Binary files /dev/null and b/wandb/run-20240515_121108-gswsvc05/run-gswsvc05.wandb differ
diff --git a/wandb/run-20240515_121902-us3spl8s/files/conda-environment.yaml b/wandb/run-20240515_121902-us3spl8s/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_121902-us3spl8s/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_121902-us3spl8s/files/config.yaml b/wandb/run-20240515_121902-us3spl8s/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..008a93c30cf9472aa794147108618192d5b785ef
--- /dev/null
+++ b/wandb/run-20240515_121902-us3spl8s/files/config.yaml
@@ -0,0 +1,755 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715775542
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 1
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+ - 1: train/loss
+ 5: 1
+ 6:
+ - 1
+ - 1: train/grad_norm
+ 5: 1
+ 6:
+ - 1
+ - 1: train/learning_rate
+ 5: 1
+ 6:
+ - 1
+ - 1: train/epoch
+ 5: 1
+ 6:
+ - 1
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.0.dev0
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./pyggel
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: false
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: 'no'
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 1
+per_device_eval_batch_size:
+ desc: null
+ value: 1
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 2
+eval_accumulation_steps:
+ desc: null
+ value: 2
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 5.0e-05
+weight_decay:
+ desc: null
+ value: 0.01
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.95
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 0.6
+num_train_epochs:
+ desc: null
+ value: 3
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value:
+ num_cycles: 3
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 2873
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./pyggel/runs/May15_12-17-57_d2827b6ba082
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.022222222222222223
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: false
+eval_steps:
+ desc: null
+ value: null
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./pyggel
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value: []
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: deepspeed_configs/zero2.json
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_torch
+optim_args:
+ desc: null
+ value: null
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: false
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.4
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: 15.0
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: true
+sample_packing_efficiency:
+ desc: null
+ value: 0.92
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 1
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240515_121902-us3spl8s/files/output.log b/wandb/run-20240515_121902-us3spl8s/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..c16f6210fb1a9b3b3eff4819ae0e359ece75d061
--- /dev/null
+++ b/wandb/run-20240515_121902-us3spl8s/files/output.log
@@ -0,0 +1,37 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/9378 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[2024-05-15 12:19:40,452] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:13671] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-15 12:19:41,543] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:13671] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+[2024-05-15 12:19:42,607] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:13671] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/zero/stage_1_and_2.py:1290: UserWarning: The torch.cuda.*DtypeTensor constructors are no longer recommended. It's best to use methods such as torch.tensor(data, dtype=*, device='cuda') to create tensors. (Triggered internally at ../torch/csrc/tensor/python_tensor.cpp:83.)
+ total_norm_cuda = get_accelerator().FloatTensor([float(total_norm)])
+ 0%| | 1/9378 [00:40<106:00:01, 40.70s/it]
+
+ 0%| | 2/9378 [01:11<90:57:52, 34.93s/it]
+[2024-05-15 12:20:52,036] [INFO] [axolotl.callbacks.on_step_end:125] [PID:13671] [RANK:0] GPU memory usage while training: 16.077GB (+26.679GB cache, +1.247GB misc)
+{'loss': 2.2459, 'grad_norm': 55.8736017722144, 'learning_rate': 3.480682213713888e-08, 'epoch': 0.0}
+
+
+ 0%| | 4/9378 [02:15<85:56:29, 33.01s/it]
+
+ 0%| | 5/9378 [02:46<83:57:52, 32.25s/it]
+
+ 0%| | 6/9378 [03:17<82:53:32, 31.84s/it]
+
+ 0%| | 7/9378 [03:50<83:39:27, 32.14s/it]
+
+ 0%| | 8/9378 [04:21<82:44:26, 31.79s/it]
+
+ 0%| | 9/9378 [04:51<81:28:35, 31.31s/it]
+
+ 0%| | 10/9378 [05:23<82:12:43, 31.59s/it]
+
+ 0%| | 11/9378 [05:55<82:17:21, 31.63s/it]
+
+ 0%| | 12/9378 [06:26<81:50:49, 31.46s/it]
+
+ 0%| | 13/9378 [06:57<81:10:49, 31.21s/it]
+
+ 0%| | 14/9378 [07:30<82:33:45, 31.74s/it]
diff --git a/wandb/run-20240515_121902-us3spl8s/files/requirements.txt b/wandb/run-20240515_121902-us3spl8s/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_121902-us3spl8s/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_121902-us3spl8s/files/tmp/axolotl_config_0mb2k0bm.yml b/wandb/run-20240515_121902-us3spl8s/files/tmp/axolotl_config_0mb2k0bm.yml
new file mode 100644
index 0000000000000000000000000000000000000000..787c228b860144f0e41f23c60122a50f1497d48e
--- /dev/null
+++ b/wandb/run-20240515_121902-us3spl8s/files/tmp/axolotl_config_0mb2k0bm.yml
@@ -0,0 +1,67 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+#push_to_hub: true
+hf_use_auth_token: true
+#resume_from_checkpoint: ./buzz-3b-small-v0.6/checkpoint-4598
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 1
+gradient_accumulation_steps: 2
+num_epochs: 3
+learning_rate: 0.00005
+optimizer: adamw_torch
+optim_args:
+adam_beta1: 0.9
+adam_beta2: 0.95
+amsgrad: true
+lr_scheduler: cosine
+lr_scheduler_kwargs:
+ num_cycles: 3
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.333
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero2.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240515_121902-us3spl8s/files/wandb-metadata.json b/wandb/run-20240515_121902-us3spl8s/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..f47d5d78e5160df712ea9b354da4e8ff6097dd23
--- /dev/null
+++ b/wandb/run-20240515_121902-us3spl8s/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T12:19:39.609389",
+ "startedAt": "2024-05-15T12:19:02.516779",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1285.8135,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3450.17,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3450.174,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3450.18,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3450.207,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3450.215,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 834.267,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3450.226,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.4019546508789
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json b/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..49bd58a589b7965ea35902ae8b8ca768de3bc8e1
--- /dev/null
+++ b/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
@@ -0,0 +1 @@
+{"train/loss": 2.3024, "train/grad_norm": 46.23872328741355, "train/learning_rate": 2.4364775495997213e-07, "train/epoch": 0.0, "train/global_step": 14, "_timestamp": 1715776030.8259907, "_runtime": 488.30117774009705, "_step": 13}
\ No newline at end of file
diff --git a/wandb/run-20240515_121902-us3spl8s/logs/debug-internal.log b/wandb/run-20240515_121902-us3spl8s/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..15dabc1bb624e30318abe4798994935f17937197
--- /dev/null
+++ b/wandb/run-20240515_121902-us3spl8s/logs/debug-internal.log
@@ -0,0 +1,337 @@
+2024-05-15 12:19:02,525 INFO StreamThr :14256 [internal.py:wandb_internal():85] W&B internal server running at pid: 14256, started at: 2024-05-15 12:19:02.524754
+2024-05-15 12:19:02,527 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status
+2024-05-15 12:19:02,530 INFO WriterThread:14256 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/run-us3spl8s.wandb
+2024-05-15 12:19:02,530 DEBUG SenderThread:14256 [sender.py:send():378] send: header
+2024-05-15 12:19:02,539 DEBUG SenderThread:14256 [sender.py:send():378] send: run
+2024-05-15 12:19:07,540 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:19:12,542 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:19:17,543 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:19:22,544 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:19:27,546 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:19:32,547 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:19:37,548 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:19:39,299 INFO SenderThread:14256 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files
+2024-05-15 12:19:39,299 INFO SenderThread:14256 [sender.py:_start_run_threads():1123] run started: us3spl8s with start time 1715775542.524813
+2024-05-15 12:19:39,299 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:19:39,331 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 12:19:39,332 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: check_version
+2024-05-15 12:19:39,420 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 12:19:39,467 DEBUG HandlerThread:14256 [system_info.py:__init__():26] System info init
+2024-05-15 12:19:39,468 DEBUG HandlerThread:14256 [system_info.py:__init__():41] System info init done
+2024-05-15 12:19:39,468 INFO HandlerThread:14256 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 12:19:39,468 INFO SystemMonitor:14256 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 12:19:39,468 INFO HandlerThread:14256 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 12:19:39,469 INFO SystemMonitor:14256 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 12:19:39,469 INFO SystemMonitor:14256 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 12:19:39,470 INFO SystemMonitor:14256 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 12:19:39,473 INFO SystemMonitor:14256 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 12:19:39,474 INFO SystemMonitor:14256 [interfaces.py:start():188] Started network monitoring
+2024-05-15 12:19:39,609 DEBUG HandlerThread:14256 [system_info.py:probe():150] Probing system
+2024-05-15 12:19:39,613 DEBUG HandlerThread:14256 [system_info.py:_probe_git():135] Probing git
+2024-05-15 12:19:39,626 DEBUG HandlerThread:14256 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 12:19:39,626 DEBUG HandlerThread:14256 [system_info.py:probe():198] Probing system done
+2024-05-15 12:19:39,626 DEBUG HandlerThread:14256 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T12:19:39.609389', 'startedAt': '2024-05-15T12:19:02.516779', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1285.8135, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3450.17, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3450.174, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3450.18, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3450.207, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3450.215, 'min': 800.0, 'max': 3500.0}, {'current': 834.267, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3450.226, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.4019546508789}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 12:19:39,627 INFO HandlerThread:14256 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 12:19:39,627 INFO HandlerThread:14256 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 12:19:39,627 DEBUG HandlerThread:14256 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 12:19:40,042 DEBUG HandlerThread:14256 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 12:19:40,043 DEBUG HandlerThread:14256 [system_info.py:_save_code():44] Saving code
+2024-05-15 12:19:40,043 WARNING HandlerThread:14256 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-15 12:19:40,043 DEBUG HandlerThread:14256 [system_info.py:_save_patches():82] Saving git patches
+2024-05-15 12:19:40,066 DEBUG HandlerThread:14256 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-15 12:19:40,069 INFO HandlerThread:14256 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 12:19:40,081 DEBUG SenderThread:14256 [sender.py:send():378] send: files
+2024-05-15 12:19:40,081 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 12:19:40,250 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 12:19:40,251 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:19:40,251 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 12:19:40,253 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:19:40,301 INFO Thread-12 :14256 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/requirements.txt
+2024-05-15 12:19:40,301 INFO Thread-12 :14256 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-metadata.json
+2024-05-15 12:19:40,301 INFO Thread-12 :14256 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/conda-environment.yaml
+2024-05-15 12:19:40,450 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-15 12:19:40,572 DEBUG SenderThread:14256 [sender.py:send():378] send: telemetry
+2024-05-15 12:19:40,572 DEBUG SenderThread:14256 [sender.py:send():378] send: config
+2024-05-15 12:19:40,574 DEBUG SenderThread:14256 [sender.py:send():378] send: metric
+2024-05-15 12:19:40,574 DEBUG SenderThread:14256 [sender.py:send():378] send: telemetry
+2024-05-15 12:19:40,575 DEBUG SenderThread:14256 [sender.py:send():378] send: telemetry
+2024-05-15 12:19:40,575 DEBUG SenderThread:14256 [sender.py:send():378] send: metric
+2024-05-15 12:19:40,575 WARNING SenderThread:14256 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-15 12:19:40,575 DEBUG SenderThread:14256 [sender.py:send():378] send: telemetry
+2024-05-15 12:19:40,575 DEBUG SenderThread:14256 [sender.py:send():378] send: telemetry
+2024-05-15 12:19:40,575 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: log_artifact
+2024-05-15 12:19:40,650 INFO wandb-upload_0:14256 [upload_job.py:push():130] Uploaded file /tmp/tmpy4ni4znxwandb/jyiyp9d8-wandb-metadata.json
+2024-05-15 12:19:41,300 INFO Thread-12 :14256 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/tmp/axolotl_config_0mb2k0bm.yml
+2024-05-15 12:19:41,300 INFO Thread-12 :14256 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/tmp
+2024-05-15 12:19:41,830 INFO wandb-upload_0:14256 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmpzu3_j7as
+2024-05-15 12:19:42,514 INFO SenderThread:14256 [sender.py:send_request_log_artifact():1455] logged artifact config-us3spl8s - {'id': 'QXJ0aWZhY3Q6ODM3MzU3ODYz', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTE0OTkyMw==', 'latestArtifact': None}}
+2024-05-15 12:19:42,515 DEBUG SenderThread:14256 [sender.py:send():378] send: telemetry
+2024-05-15 12:19:42,515 DEBUG SenderThread:14256 [sender.py:send():378] send: files
+2024-05-15 12:19:42,516 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file tmp/axolotl_config_0mb2k0bm.yml with policy live
+2024-05-15 12:19:42,896 INFO wandb-upload_0:14256 [upload_job.py:push():130] Uploaded file /tmp/tmpy4ni4znxwandb/ijytjant-tmp/axolotl_config_0mb2k0bm.yml
+2024-05-15 12:19:43,301 INFO Thread-12 :14256 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:19:44,679 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:19:45,302 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:19:49,680 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:19:54,681 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:19:55,251 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:19:55,252 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:20:00,448 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:05,449 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:10,251 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:20:10,252 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:20:10,475 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:11,319 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/config.yaml
+2024-05-15 12:20:16,023 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:17,323 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:20:21,024 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:21,152 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:20:21,156 DEBUG SenderThread:14256 [sender.py:send():378] send: metric
+2024-05-15 12:20:21,157 DEBUG SenderThread:14256 [sender.py:send():378] send: metric
+2024-05-15 12:20:21,158 DEBUG SenderThread:14256 [sender.py:send():378] send: metric
+2024-05-15 12:20:21,158 DEBUG SenderThread:14256 [sender.py:send():378] send: metric
+2024-05-15 12:20:21,159 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:20:21,159 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:20:21,160 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:20:21,326 INFO Thread-12 :14256 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:20:23,327 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:20:25,251 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:20:25,252 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:20:26,465 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:31,466 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:36,467 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:39,474 DEBUG SystemMonitor:14256 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-15 12:20:39,477 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:20:40,253 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:20:40,254 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:20:41,478 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:42,340 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/config.yaml
+2024-05-15 12:20:46,674 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:51,676 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:20:52,039 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:20:52,041 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:20:52,043 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:20:52,044 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:20:52,347 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:20:53,348 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:20:55,252 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:20:55,252 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:20:57,404 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:02,405 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:07,406 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:09,481 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:21:10,252 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:21:10,252 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:21:12,444 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:17,445 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:22,446 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:24,265 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:21:24,267 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:21:24,269 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:21:24,270 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:21:24,369 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:21:25,252 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:21:25,253 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:21:25,370 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:21:27,371 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:21:28,437 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:33,438 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:38,439 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:39,485 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:21:40,252 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:21:40,253 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:21:43,502 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:48,503 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:53,504 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:21:56,210 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:21:56,211 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:21:56,258 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:21:56,371 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:21:56,371 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:21:56,372 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:21:56,390 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:21:57,391 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:21:59,373 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:04,374 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:09,375 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:09,490 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:22:11,211 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:22:11,211 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:22:14,405 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:19,406 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:24,407 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:27,119 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:22:27,120 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:22:27,166 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:22:27,302 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:22:27,303 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:22:27,304 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:22:27,411 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:22:29,412 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:22:30,305 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:35,306 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:39,496 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:22:40,497 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:42,119 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:22:42,119 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:22:46,324 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:51,325 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:56,326 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:22:58,167 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:22:58,168 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:22:58,210 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:22:58,393 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:22:58,394 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:22:58,395 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:22:58,432 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:22:59,432 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:23:01,396 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:06,397 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:09,502 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:23:11,505 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:13,165 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:23:13,166 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:23:17,363 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:22,364 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:27,365 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:28,165 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:23:28,166 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:23:30,923 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:23:30,926 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:23:30,927 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:23:30,929 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:23:31,453 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:23:32,930 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:33,455 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:23:37,931 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:39,508 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:23:43,167 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:23:43,168 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:23:43,346 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:48,347 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:53,348 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:23:58,290 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:23:58,291 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:23:58,481 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:02,010 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:24:02,013 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:24:02,014 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:24:02,016 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:24:02,474 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:24:03,475 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:24:04,017 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:09,018 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:09,514 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:24:13,290 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:24:13,291 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:24:14,514 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:19,514 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:24,515 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:28,290 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:24:28,291 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:24:30,512 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:32,211 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:24:32,214 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:24:32,215 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:24:32,217 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:24:32,494 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:24:33,495 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:24:36,218 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:39,520 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:24:41,521 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:43,290 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:24:43,291 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:24:47,504 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:52,505 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:57,506 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:24:58,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:24:58,292 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:25:03,498 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:04,445 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:25:04,447 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:25:04,448 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:25:04,451 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:25:04,515 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:25:05,516 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:25:09,452 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:09,525 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:25:13,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:25:13,291 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:25:14,465 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:19,466 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:24,467 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:28,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:25:28,292 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:25:29,488 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:34,490 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:36,146 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:25:36,149 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:25:36,150 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:25:36,152 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:25:36,536 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:25:37,537 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:25:39,531 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:25:39,533 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:43,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:25:43,291 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:25:44,542 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:49,543 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:54,544 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:25:58,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:25:58,292 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:26:00,524 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:05,525 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:07,225 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:26:07,228 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:26:07,230 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:26:07,231 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:26:07,557 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:26:09,537 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:26:09,558 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:26:10,540 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:13,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:26:13,291 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:26:16,535 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:21,535 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:26,536 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:28,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:26:28,292 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:26:32,468 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:37,469 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:37,889 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:26:37,892 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:26:37,892 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:26:37,894 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:26:38,577 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:26:39,543 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:26:39,578 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:26:42,546 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:43,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:26:43,292 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:26:47,548 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:52,549 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:57,550 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:26:58,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:26:58,292 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:27:03,473 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:27:08,474 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:27:09,550 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:27:10,826 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 12:27:10,828 DEBUG SenderThread:14256 [sender.py:send():378] send: history
+2024-05-15 12:27:10,830 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 12:27:10,830 INFO SenderThread:14256 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 12:27:11,633 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/wandb-summary.json
+2024-05-15 12:27:11,633 INFO Thread-12 :14256 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/files/output.log
+2024-05-15 12:27:13,291 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:27:13,292 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:27:13,487 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:27:18,488 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:27:23,489 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:27:28,292 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:27:28,292 DEBUG SenderThread:14256 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:27:29,487 DEBUG HandlerThread:14256 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:27:33,685 INFO memory :14256 [interfaces.py:monitor():138] Process proc.memory.rssMB has exited.
+2024-05-15 12:27:33,686 DEBUG SystemMonitor:14256 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-15 12:27:33,686 DEBUG SystemMonitor:14256 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-15 12:27:33,696 DEBUG SenderThread:14256 [sender.py:send():378] send: stats
+2024-05-15 12:27:34,526 INFO MainThread:14256 [internal.py:handle_exit():75] Internal process exited
diff --git a/wandb/run-20240515_121902-us3spl8s/logs/debug.log b/wandb/run-20240515_121902-us3spl8s/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..8b09db2e622034c999ab1e8850d97e6b2a58d7ec
--- /dev/null
+++ b/wandb/run-20240515_121902-us3spl8s/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 12:19:02,519 INFO MainThread:13671 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 12:19:02,519 INFO MainThread:13671 [wandb_setup.py:_flush():76] Configure stats pid to 13671
+2024-05-15 12:19:02,519 INFO MainThread:13671 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 12:19:02,520 INFO MainThread:13671 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 12:19:02,520 WARNING MainThread:13671 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 12:19:02,520 INFO MainThread:13671 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 12:19:02,520 INFO MainThread:13671 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 12:19:02,520 WARNING MainThread:13671 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 12:19:02,520 INFO MainThread:13671 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 12:19:02,520 INFO MainThread:13671 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 12:19:02,520 INFO MainThread:13671 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/logs/debug.log
+2024-05-15 12:19:02,520 INFO MainThread:13671 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_121902-us3spl8s/logs/debug-internal.log
+2024-05-15 12:19:02,520 INFO MainThread:13671 [wandb_init.py:init():560] calling init triggers
+2024-05-15 12:19:02,520 INFO MainThread:13671 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 12:19:02,521 INFO MainThread:13671 [wandb_init.py:init():610] starting backend
+2024-05-15 12:19:02,521 INFO MainThread:13671 [wandb_init.py:init():614] setting up manager
+2024-05-15 12:19:02,523 INFO MainThread:13671 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 12:19:02,524 INFO MainThread:13671 [wandb_init.py:init():622] backend started and connected
+2024-05-15 12:19:02,528 INFO MainThread:13671 [wandb_init.py:init():711] updated telemetry
+2024-05-15 12:19:02,538 INFO MainThread:13671 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 12:19:39,330 INFO MainThread:13671 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 12:19:39,408 INFO MainThread:13671 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 12:19:39,409 INFO MainThread:13671 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 12:19:40,251 INFO MainThread:13671 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 12:19:40,251 INFO MainThread:13671 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 12:19:40,251 INFO MainThread:13671 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 12:19:40,251 INFO MainThread:13671 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 12:19:40,252 INFO MainThread:13671 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 12:19:40,255 INFO MainThread:13671 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 2, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {'num_cycles': 3}, 'warmup_ratio': 0.0, 'warmup_steps': 2873, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_12-17-57_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero2.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.92, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 1, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 12:19:40,257 INFO MainThread:13671 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_121902-us3spl8s/run-us3spl8s.wandb b/wandb/run-20240515_121902-us3spl8s/run-us3spl8s.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..994b9baa21714f9a265abed492355546f60ace8b
Binary files /dev/null and b/wandb/run-20240515_121902-us3spl8s/run-us3spl8s.wandb differ
diff --git a/wandb/run-20240515_123016-qssouxf5/files/conda-environment.yaml b/wandb/run-20240515_123016-qssouxf5/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_123016-qssouxf5/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_123016-qssouxf5/files/config.yaml b/wandb/run-20240515_123016-qssouxf5/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..fcbb1816ba42e35a14aa96bd61bf4fa83c408758
--- /dev/null
+++ b/wandb/run-20240515_123016-qssouxf5/files/config.yaml
@@ -0,0 +1,35 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715776216
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 23
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 13: linux-x86_64
diff --git a/wandb/run-20240515_123016-qssouxf5/files/output.log b/wandb/run-20240515_123016-qssouxf5/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..a01d16b5c3f7536945402c480c634046db8bb6c7
--- /dev/null
+++ b/wandb/run-20240515_123016-qssouxf5/files/output.log
@@ -0,0 +1,6 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/366 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[2024-05-15 12:30:23,492] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:14470] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-15 12:30:24,161] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:14470] [RANK:0] packing_efficiency_estimate: 0.98 total_num_tokens per device: 47609834
diff --git a/wandb/run-20240515_123016-qssouxf5/files/requirements.txt b/wandb/run-20240515_123016-qssouxf5/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_123016-qssouxf5/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_123016-qssouxf5/files/tmp/axolotl_config_rnrykqff.yml b/wandb/run-20240515_123016-qssouxf5/files/tmp/axolotl_config_rnrykqff.yml
new file mode 100644
index 0000000000000000000000000000000000000000..0c7b34ad2d712622ee91a25ab95e55caef5da6d5
--- /dev/null
+++ b/wandb/run-20240515_123016-qssouxf5/files/tmp/axolotl_config_rnrykqff.yml
@@ -0,0 +1,67 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+#push_to_hub: true
+hf_use_auth_token: true
+#resume_from_checkpoint: ./buzz-3b-small-v0.6/checkpoint-4598
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 3
+gradient_accumulation_steps: 16
+num_epochs: 3
+learning_rate: 0.00005
+optimizer: adamw_torch
+optim_args:
+adam_beta1: 0.9
+adam_beta2: 0.95
+amsgrad: true
+lr_scheduler: cosine
+lr_scheduler_kwargs:
+ num_cycles: 3
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.333
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero2.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240515_123016-qssouxf5/files/wandb-metadata.json b/wandb/run-20240515_123016-qssouxf5/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..3f648c0360149af7204ef8d30b5bbb8e34a8a754
--- /dev/null
+++ b/wandb/run-20240515_123016-qssouxf5/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T12:30:22.564228",
+ "startedAt": "2024-05-15T12:30:16.856378",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1184.8509583333334,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.002,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3324.493,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1700.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.002,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.001,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.40234375
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_123016-qssouxf5/logs/debug-internal.log b/wandb/run-20240515_123016-qssouxf5/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..ff79aec59fbfc3187a29d8bd0cea092f04fc3342
--- /dev/null
+++ b/wandb/run-20240515_123016-qssouxf5/logs/debug-internal.log
@@ -0,0 +1,69 @@
+2024-05-15 12:30:16,863 INFO StreamThr :15051 [internal.py:wandb_internal():85] W&B internal server running at pid: 15051, started at: 2024-05-15 12:30:16.863403
+2024-05-15 12:30:16,866 DEBUG HandlerThread:15051 [handler.py:handle_request():158] handle_request: status
+2024-05-15 12:30:16,868 INFO WriterThread:15051 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/run-qssouxf5.wandb
+2024-05-15 12:30:16,869 DEBUG SenderThread:15051 [sender.py:send():378] send: header
+2024-05-15 12:30:16,878 DEBUG SenderThread:15051 [sender.py:send():378] send: run
+2024-05-15 12:30:17,233 INFO SenderThread:15051 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/files
+2024-05-15 12:30:17,233 INFO SenderThread:15051 [sender.py:_start_run_threads():1123] run started: qssouxf5 with start time 1715776216.863766
+2024-05-15 12:30:17,265 DEBUG HandlerThread:15051 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 12:30:17,265 DEBUG SenderThread:15051 [sender.py:send_request():405] send_request: check_version
+2024-05-15 12:30:22,266 DEBUG HandlerThread:15051 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:30:22,268 DEBUG HandlerThread:15051 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:30:22,322 DEBUG HandlerThread:15051 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 12:30:22,368 DEBUG HandlerThread:15051 [system_info.py:__init__():26] System info init
+2024-05-15 12:30:22,368 DEBUG HandlerThread:15051 [system_info.py:__init__():41] System info init done
+2024-05-15 12:30:22,368 INFO HandlerThread:15051 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 12:30:22,368 INFO SystemMonitor:15051 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 12:30:22,369 INFO HandlerThread:15051 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 12:30:22,370 INFO SystemMonitor:15051 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 12:30:22,371 INFO SystemMonitor:15051 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 12:30:22,371 INFO SystemMonitor:15051 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 12:30:22,372 INFO SystemMonitor:15051 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 12:30:22,373 INFO SystemMonitor:15051 [interfaces.py:start():188] Started network monitoring
+2024-05-15 12:30:22,564 DEBUG HandlerThread:15051 [system_info.py:probe():150] Probing system
+2024-05-15 12:30:22,566 DEBUG HandlerThread:15051 [system_info.py:_probe_git():135] Probing git
+2024-05-15 12:30:22,575 DEBUG HandlerThread:15051 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 12:30:22,575 DEBUG HandlerThread:15051 [system_info.py:probe():198] Probing system done
+2024-05-15 12:30:22,575 DEBUG HandlerThread:15051 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T12:30:22.564228', 'startedAt': '2024-05-15T12:30:16.856378', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1184.8509583333334, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.002, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3324.493, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1700.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.002, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.001, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.40234375}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 12:30:22,575 INFO HandlerThread:15051 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 12:30:22,575 INFO HandlerThread:15051 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 12:30:22,575 DEBUG HandlerThread:15051 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 12:30:23,041 DEBUG HandlerThread:15051 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 12:30:23,041 DEBUG HandlerThread:15051 [system_info.py:_save_code():44] Saving code
+2024-05-15 12:30:23,041 WARNING HandlerThread:15051 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-15 12:30:23,041 DEBUG HandlerThread:15051 [system_info.py:_save_patches():82] Saving git patches
+2024-05-15 12:30:23,063 DEBUG HandlerThread:15051 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-15 12:30:23,066 INFO HandlerThread:15051 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 12:30:23,077 DEBUG SenderThread:15051 [sender.py:send():378] send: files
+2024-05-15 12:30:23,078 INFO SenderThread:15051 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 12:30:23,239 INFO Thread-12 :15051 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/files/conda-environment.yaml
+2024-05-15 12:30:23,239 INFO Thread-12 :15051 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/files/wandb-metadata.json
+2024-05-15 12:30:23,288 DEBUG HandlerThread:15051 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 12:30:23,289 DEBUG HandlerThread:15051 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:30:23,290 DEBUG SenderThread:15051 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 12:30:23,291 DEBUG SenderThread:15051 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:30:23,490 DEBUG HandlerThread:15051 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-15 12:30:23,681 INFO wandb-upload_0:15051 [upload_job.py:push():130] Uploaded file /tmp/tmp8el9xc76wandb/lqmxlpmx-wandb-metadata.json
+2024-05-15 12:30:23,688 DEBUG SenderThread:15051 [sender.py:send():378] send: telemetry
+2024-05-15 12:30:23,688 DEBUG SenderThread:15051 [sender.py:send():378] send: config
+2024-05-15 12:30:23,690 DEBUG SenderThread:15051 [sender.py:send():378] send: metric
+2024-05-15 12:30:23,691 DEBUG SenderThread:15051 [sender.py:send():378] send: telemetry
+2024-05-15 12:30:23,691 DEBUG SenderThread:15051 [sender.py:send():378] send: telemetry
+2024-05-15 12:30:23,691 DEBUG SenderThread:15051 [sender.py:send():378] send: metric
+2024-05-15 12:30:23,691 WARNING SenderThread:15051 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-15 12:30:23,691 DEBUG SenderThread:15051 [sender.py:send():378] send: telemetry
+2024-05-15 12:30:23,691 DEBUG SenderThread:15051 [sender.py:send():378] send: telemetry
+2024-05-15 12:30:23,691 DEBUG SenderThread:15051 [sender.py:send_request():405] send_request: log_artifact
+2024-05-15 12:30:24,235 INFO Thread-12 :15051 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/files/requirements.txt
+2024-05-15 12:30:24,235 INFO Thread-12 :15051 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/files/tmp/axolotl_config_rnrykqff.yml
+2024-05-15 12:30:24,236 INFO Thread-12 :15051 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/files/tmp
+2024-05-15 12:30:24,986 INFO wandb-upload_0:15051 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmpzkrau3hs
+2024-05-15 12:30:25,709 INFO SenderThread:15051 [sender.py:send_request_log_artifact():1455] logged artifact config-qssouxf5 - {'id': 'QXJ0aWZhY3Q6ODM3MzY5Mzk3', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTE1MzcyMg==', 'latestArtifact': None}}
+2024-05-15 12:30:25,710 DEBUG SenderThread:15051 [sender.py:send():378] send: telemetry
+2024-05-15 12:30:25,710 DEBUG SenderThread:15051 [sender.py:send():378] send: files
+2024-05-15 12:30:25,710 INFO SenderThread:15051 [sender.py:_save_file():1389] saving file tmp/axolotl_config_rnrykqff.yml with policy live
+2024-05-15 12:30:26,101 INFO wandb-upload_0:15051 [upload_job.py:push():130] Uploaded file /tmp/tmp8el9xc76wandb/k1a2sswh-tmp/axolotl_config_rnrykqff.yml
+2024-05-15 12:30:26,236 INFO Thread-12 :15051 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/files/output.log
+2024-05-15 12:30:27,858 DEBUG HandlerThread:15051 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:30:28,237 INFO Thread-12 :15051 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/files/output.log
+2024-05-15 12:30:30,284 INFO MainThread:15051 [internal.py:handle_exit():75] Internal process exited
diff --git a/wandb/run-20240515_123016-qssouxf5/logs/debug.log b/wandb/run-20240515_123016-qssouxf5/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..d726eccc7890efc5dafcbcf4e2c121f650795fc5
--- /dev/null
+++ b/wandb/run-20240515_123016-qssouxf5/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 12:30:16,858 INFO MainThread:14470 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 12:30:16,858 INFO MainThread:14470 [wandb_setup.py:_flush():76] Configure stats pid to 14470
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 12:30:16,859 WARNING MainThread:14470 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 12:30:16,859 WARNING MainThread:14470 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/logs/debug.log
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_123016-qssouxf5/logs/debug-internal.log
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_init.py:init():560] calling init triggers
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_init.py:init():610] starting backend
+2024-05-15 12:30:16,859 INFO MainThread:14470 [wandb_init.py:init():614] setting up manager
+2024-05-15 12:30:16,861 INFO MainThread:14470 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 12:30:16,863 INFO MainThread:14470 [wandb_init.py:init():622] backend started and connected
+2024-05-15 12:30:16,867 INFO MainThread:14470 [wandb_init.py:init():711] updated telemetry
+2024-05-15 12:30:16,877 INFO MainThread:14470 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 12:30:17,264 INFO MainThread:14470 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 12:30:22,268 INFO MainThread:14470 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 12:30:22,269 INFO MainThread:14470 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 12:30:23,289 INFO MainThread:14470 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 12:30:23,289 INFO MainThread:14470 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 12:30:23,289 INFO MainThread:14470 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 12:30:23,289 INFO MainThread:14470 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 12:30:23,290 INFO MainThread:14470 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 12:30:23,292 INFO MainThread:14470 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 3, 'per_device_eval_batch_size': 3, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 16, 'eval_accumulation_steps': 16, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {'num_cycles': 3}, 'warmup_ratio': 0.0, 'warmup_steps': 118, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_12-29-24_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero2.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.98, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 3, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 12:30:23,294 INFO MainThread:14470 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_123016-qssouxf5/run-qssouxf5.wandb b/wandb/run-20240515_123016-qssouxf5/run-qssouxf5.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..c522321d27ff322ace202f72237511f3644d2189
Binary files /dev/null and b/wandb/run-20240515_123016-qssouxf5/run-qssouxf5.wandb differ
diff --git a/wandb/run-20240515_125149-c28h2y8m/files/conda-environment.yaml b/wandb/run-20240515_125149-c28h2y8m/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_125149-c28h2y8m/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_125149-c28h2y8m/files/config.yaml b/wandb/run-20240515_125149-c28h2y8m/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..d9818d53f55e433638b7c1cab9412c1d85509d60
--- /dev/null
+++ b/wandb/run-20240515_125149-c28h2y8m/files/config.yaml
@@ -0,0 +1,35 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715777509
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 23
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 13: linux-x86_64
diff --git a/wandb/run-20240515_125149-c28h2y8m/files/requirements.txt b/wandb/run-20240515_125149-c28h2y8m/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_125149-c28h2y8m/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_125149-c28h2y8m/files/tmp/axolotl_config_n2j_g17w.yml b/wandb/run-20240515_125149-c28h2y8m/files/tmp/axolotl_config_n2j_g17w.yml
new file mode 100644
index 0000000000000000000000000000000000000000..735f30302ecab0962ade9b2240b23d1815d021f9
--- /dev/null
+++ b/wandb/run-20240515_125149-c28h2y8m/files/tmp/axolotl_config_n2j_g17w.yml
@@ -0,0 +1,66 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+push_to_hub: true
+hf_use_auth_token: true
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 2
+gradient_accumulation_steps: 16
+num_epochs: 3
+learning_rate: 0.00005
+optimizer: adamw_bnb_8bit
+optim_args:
+adam_beta1: 0.9
+adam_beta2: 0.95
+amsgrad: true
+lr_scheduler: cosine
+lr_scheduler_kwargs:
+ num_cycles: 3
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.333
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero1.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240515_125149-c28h2y8m/files/wandb-metadata.json b/wandb/run-20240515_125149-c28h2y8m/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..bc3307cc6323bf063050002a383f2c2f4ded681e
--- /dev/null
+++ b/wandb/run-20240515_125149-c28h2y8m/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T12:51:50.052475",
+ "startedAt": "2024-05-15T12:51:49.385250",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1260.7434791666667,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3399.997,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3399.997,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.40304565429688
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_125149-c28h2y8m/logs/debug-internal.log b/wandb/run-20240515_125149-c28h2y8m/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..92df08bf0f280e47404f344acecb52cd5bc8e0fe
--- /dev/null
+++ b/wandb/run-20240515_125149-c28h2y8m/logs/debug-internal.log
@@ -0,0 +1,63 @@
+2024-05-15 12:51:49,393 INFO StreamThr :15735 [internal.py:wandb_internal():85] W&B internal server running at pid: 15735, started at: 2024-05-15 12:51:49.392388
+2024-05-15 12:51:49,396 DEBUG HandlerThread:15735 [handler.py:handle_request():158] handle_request: status
+2024-05-15 12:51:49,399 INFO WriterThread:15735 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/run-c28h2y8m.wandb
+2024-05-15 12:51:49,400 DEBUG SenderThread:15735 [sender.py:send():378] send: header
+2024-05-15 12:51:49,406 DEBUG SenderThread:15735 [sender.py:send():378] send: run
+2024-05-15 12:51:49,820 INFO SenderThread:15735 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/files
+2024-05-15 12:51:49,820 INFO SenderThread:15735 [sender.py:_start_run_threads():1123] run started: c28h2y8m with start time 1715777509.392538
+2024-05-15 12:51:49,852 DEBUG HandlerThread:15735 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 12:51:49,852 DEBUG SenderThread:15735 [sender.py:send_request():405] send_request: check_version
+2024-05-15 12:51:49,937 DEBUG HandlerThread:15735 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 12:51:49,972 DEBUG HandlerThread:15735 [system_info.py:__init__():26] System info init
+2024-05-15 12:51:49,972 DEBUG HandlerThread:15735 [system_info.py:__init__():41] System info init done
+2024-05-15 12:51:49,972 INFO HandlerThread:15735 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 12:51:49,972 INFO SystemMonitor:15735 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 12:51:49,972 INFO HandlerThread:15735 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 12:51:49,973 INFO SystemMonitor:15735 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 12:51:49,974 INFO SystemMonitor:15735 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 12:51:49,975 INFO SystemMonitor:15735 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 12:51:49,978 INFO SystemMonitor:15735 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 12:51:49,979 INFO SystemMonitor:15735 [interfaces.py:start():188] Started network monitoring
+2024-05-15 12:51:50,052 DEBUG HandlerThread:15735 [system_info.py:probe():150] Probing system
+2024-05-15 12:51:50,055 DEBUG HandlerThread:15735 [system_info.py:_probe_git():135] Probing git
+2024-05-15 12:51:50,064 DEBUG HandlerThread:15735 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 12:51:50,064 DEBUG HandlerThread:15735 [system_info.py:probe():198] Probing system done
+2024-05-15 12:51:50,064 DEBUG HandlerThread:15735 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T12:51:50.052475', 'startedAt': '2024-05-15T12:51:49.385250', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1260.7434791666667, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 1800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3399.997, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3399.997, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.40304565429688}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 12:51:50,065 INFO HandlerThread:15735 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 12:51:50,065 INFO HandlerThread:15735 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 12:51:50,065 DEBUG HandlerThread:15735 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 12:51:50,530 DEBUG HandlerThread:15735 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 12:51:50,530 DEBUG HandlerThread:15735 [system_info.py:_save_code():44] Saving code
+2024-05-15 12:51:50,531 WARNING HandlerThread:15735 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-15 12:51:50,531 DEBUG HandlerThread:15735 [system_info.py:_save_patches():82] Saving git patches
+2024-05-15 12:51:50,553 DEBUG HandlerThread:15735 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-15 12:51:50,556 INFO HandlerThread:15735 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 12:51:50,568 DEBUG SenderThread:15735 [sender.py:send():378] send: files
+2024-05-15 12:51:50,568 INFO SenderThread:15735 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 12:51:50,781 DEBUG HandlerThread:15735 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 12:51:50,781 DEBUG HandlerThread:15735 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:51:50,782 DEBUG SenderThread:15735 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 12:51:50,784 DEBUG SenderThread:15735 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:51:50,823 INFO Thread-12 :15735 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/files/wandb-metadata.json
+2024-05-15 12:51:50,823 INFO Thread-12 :15735 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/files/conda-environment.yaml
+2024-05-15 12:51:50,823 INFO Thread-12 :15735 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/files/requirements.txt
+2024-05-15 12:51:50,974 DEBUG HandlerThread:15735 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-15 12:51:51,154 INFO wandb-upload_0:15735 [upload_job.py:push():130] Uploaded file /tmp/tmpkqao2lj8wandb/t97csxyz-wandb-metadata.json
+2024-05-15 12:51:51,822 INFO Thread-12 :15735 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/files/tmp/axolotl_config_n2j_g17w.yml
+2024-05-15 12:51:51,822 INFO Thread-12 :15735 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/files/tmp
+2024-05-15 12:51:56,014 DEBUG HandlerThread:15735 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:51:57,990 INFO cpu :15735 [interfaces.py:monitor():138] Process cpu has exited.
+2024-05-15 12:51:57,991 DEBUG SystemMonitor:15735 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-15 12:51:57,991 DEBUG SystemMonitor:15735 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-15 12:51:57,992 DEBUG SystemMonitor:15735 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-15 12:51:58,406 WARNING StreamThr :15735 [internal.py:is_dead():413] Internal process exiting, parent pid 15500 disappeared
+2024-05-15 12:51:58,407 ERROR StreamThr :15735 [internal.py:wandb_internal():151] Internal process shutdown.
+2024-05-15 12:51:58,993 INFO HandlerThread:15735 [handler.py:finish():882] shutting down handler
+2024-05-15 12:51:58,994 INFO HandlerThread:15735 [system_monitor.py:finish():203] Stopping system monitor
+2024-05-15 12:51:58,994 INFO HandlerThread:15735 [interfaces.py:finish():200] Joined cpu monitor
+2024-05-15 12:51:58,994 INFO WriterThread:15735 [datastore.py:close():296] close: /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/run-c28h2y8m.wandb
+2024-05-15 12:51:58,995 INFO HandlerThread:15735 [interfaces.py:finish():200] Joined disk monitor
+2024-05-15 12:51:58,995 INFO HandlerThread:15735 [interfaces.py:finish():200] Joined gpu monitor
+2024-05-15 12:51:58,996 INFO HandlerThread:15735 [interfaces.py:finish():200] Joined memory monitor
+2024-05-15 12:51:58,996 INFO HandlerThread:15735 [interfaces.py:finish():200] Joined network monitor
+2024-05-15 12:51:59,939 INFO MainThread:15735 [internal.py:handle_exit():75] Internal process exited
diff --git a/wandb/run-20240515_125149-c28h2y8m/logs/debug.log b/wandb/run-20240515_125149-c28h2y8m/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..0429bee9864692ab4c1fc4eb6e090fdfd3ff64e6
--- /dev/null
+++ b/wandb/run-20240515_125149-c28h2y8m/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 12:51:49,387 INFO MainThread:15500 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 12:51:49,387 INFO MainThread:15500 [wandb_setup.py:_flush():76] Configure stats pid to 15500
+2024-05-15 12:51:49,387 INFO MainThread:15500 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 12:51:49,387 INFO MainThread:15500 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 12:51:49,387 WARNING MainThread:15500 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 12:51:49,388 WARNING MainThread:15500 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/logs/debug.log
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_125149-c28h2y8m/logs/debug-internal.log
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_init.py:init():560] calling init triggers
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_init.py:init():610] starting backend
+2024-05-15 12:51:49,388 INFO MainThread:15500 [wandb_init.py:init():614] setting up manager
+2024-05-15 12:51:49,390 INFO MainThread:15500 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 12:51:49,392 INFO MainThread:15500 [wandb_init.py:init():622] backend started and connected
+2024-05-15 12:51:49,395 INFO MainThread:15500 [wandb_init.py:init():711] updated telemetry
+2024-05-15 12:51:49,405 INFO MainThread:15500 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 12:51:49,851 INFO MainThread:15500 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 12:51:49,924 INFO MainThread:15500 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 12:51:49,924 INFO MainThread:15500 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 12:51:50,782 INFO MainThread:15500 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 12:51:50,782 INFO MainThread:15500 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 12:51:50,782 INFO MainThread:15500 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 12:51:50,782 INFO MainThread:15500 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 12:51:50,783 INFO MainThread:15500 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 12:51:50,785 INFO MainThread:15500 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 16, 'eval_accumulation_steps': 16, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {'num_cycles': 3}, 'warmup_ratio': 0.0, 'warmup_steps': 178, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_12-51-21_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero1.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_bnb_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.96, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 2, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 12:51:50,787 INFO MainThread:15500 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_125149-c28h2y8m/run-c28h2y8m.wandb b/wandb/run-20240515_125149-c28h2y8m/run-c28h2y8m.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..62fd33b0e23e878db532c5f7e9e93d409a98f5ba
Binary files /dev/null and b/wandb/run-20240515_125149-c28h2y8m/run-c28h2y8m.wandb differ
diff --git a/wandb/run-20240515_125656-5z8o4gdu/files/conda-environment.yaml b/wandb/run-20240515_125656-5z8o4gdu/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_125656-5z8o4gdu/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_125656-5z8o4gdu/files/config.yaml b/wandb/run-20240515_125656-5z8o4gdu/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..ac21b09de2a3d03964b408c85127bc223a9f2558
--- /dev/null
+++ b/wandb/run-20240515_125656-5z8o4gdu/files/config.yaml
@@ -0,0 +1,35 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715777816
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 23
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 13: linux-x86_64
diff --git a/wandb/run-20240515_125656-5z8o4gdu/files/requirements.txt b/wandb/run-20240515_125656-5z8o4gdu/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_125656-5z8o4gdu/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_125656-5z8o4gdu/files/wandb-metadata.json b/wandb/run-20240515_125656-5z8o4gdu/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..f3db3faa56b023550624f0bd03cd0064f3c7c4a8
--- /dev/null
+++ b/wandb/run-20240515_125656-5z8o4gdu/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T12:56:57.090838",
+ "startedAt": "2024-05-15T12:56:56.431611",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1447.9168541666666,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3399.997,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3399.996,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.003,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1000.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.40332794189453
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_125656-5z8o4gdu/logs/debug-internal.log b/wandb/run-20240515_125656-5z8o4gdu/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..e8b27f6a1c5d803c0857e8391b2dd66a4094528c
--- /dev/null
+++ b/wandb/run-20240515_125656-5z8o4gdu/logs/debug-internal.log
@@ -0,0 +1,61 @@
+2024-05-15 12:56:56,436 INFO StreamThr :16168 [internal.py:wandb_internal():85] W&B internal server running at pid: 16168, started at: 2024-05-15 12:56:56.435709
+2024-05-15 12:56:56,440 DEBUG HandlerThread:16168 [handler.py:handle_request():158] handle_request: status
+2024-05-15 12:56:56,442 INFO WriterThread:16168 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_125656-5z8o4gdu/run-5z8o4gdu.wandb
+2024-05-15 12:56:56,443 DEBUG SenderThread:16168 [sender.py:send():378] send: header
+2024-05-15 12:56:56,450 DEBUG SenderThread:16168 [sender.py:send():378] send: run
+2024-05-15 12:56:56,870 INFO SenderThread:16168 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_125656-5z8o4gdu/files
+2024-05-15 12:56:56,870 INFO SenderThread:16168 [sender.py:_start_run_threads():1123] run started: 5z8o4gdu with start time 1715777816.436042
+2024-05-15 12:56:56,898 DEBUG HandlerThread:16168 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 12:56:56,899 DEBUG SenderThread:16168 [sender.py:send_request():405] send_request: check_version
+2024-05-15 12:56:56,983 DEBUG HandlerThread:16168 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 12:56:57,016 DEBUG HandlerThread:16168 [system_info.py:__init__():26] System info init
+2024-05-15 12:56:57,016 DEBUG HandlerThread:16168 [system_info.py:__init__():41] System info init done
+2024-05-15 12:56:57,016 INFO HandlerThread:16168 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 12:56:57,016 INFO SystemMonitor:16168 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 12:56:57,017 INFO HandlerThread:16168 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 12:56:57,018 INFO SystemMonitor:16168 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 12:56:57,019 INFO SystemMonitor:16168 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 12:56:57,021 INFO SystemMonitor:16168 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 12:56:57,023 INFO SystemMonitor:16168 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 12:56:57,024 INFO SystemMonitor:16168 [interfaces.py:start():188] Started network monitoring
+2024-05-15 12:56:57,090 DEBUG HandlerThread:16168 [system_info.py:probe():150] Probing system
+2024-05-15 12:56:57,093 DEBUG HandlerThread:16168 [system_info.py:_probe_git():135] Probing git
+2024-05-15 12:56:57,103 DEBUG HandlerThread:16168 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 12:56:57,103 DEBUG HandlerThread:16168 [system_info.py:probe():198] Probing system done
+2024-05-15 12:56:57,103 DEBUG HandlerThread:16168 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T12:56:57.090838', 'startedAt': '2024-05-15T12:56:56.431611', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1447.9168541666666, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 900.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3399.997, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3399.996, 'min': 800.0, 'max': 3500.0}, {'current': 3400.003, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1900.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 1000.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.40332794189453}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 12:56:57,103 INFO HandlerThread:16168 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 12:56:57,103 INFO HandlerThread:16168 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 12:56:57,103 DEBUG HandlerThread:16168 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 12:56:57,569 DEBUG HandlerThread:16168 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 12:56:57,569 DEBUG HandlerThread:16168 [system_info.py:_save_code():44] Saving code
+2024-05-15 12:56:57,569 WARNING HandlerThread:16168 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-15 12:56:57,569 DEBUG HandlerThread:16168 [system_info.py:_save_patches():82] Saving git patches
+2024-05-15 12:56:57,588 DEBUG HandlerThread:16168 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-15 12:56:57,589 INFO HandlerThread:16168 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 12:56:57,595 DEBUG SenderThread:16168 [sender.py:send():378] send: files
+2024-05-15 12:56:57,595 INFO SenderThread:16168 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 12:56:57,809 DEBUG HandlerThread:16168 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 12:56:57,810 DEBUG SenderThread:16168 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 12:56:57,810 DEBUG HandlerThread:16168 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 12:56:57,811 DEBUG SenderThread:16168 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 12:56:57,872 INFO Thread-12 :16168 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_125656-5z8o4gdu/files/conda-environment.yaml
+2024-05-15 12:56:57,873 INFO Thread-12 :16168 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_125656-5z8o4gdu/files/wandb-metadata.json
+2024-05-15 12:56:57,873 INFO Thread-12 :16168 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_125656-5z8o4gdu/files/requirements.txt
+2024-05-15 12:56:58,001 DEBUG SenderThread:16168 [sender.py:send():378] send: telemetry
+2024-05-15 12:56:58,001 DEBUG SenderThread:16168 [sender.py:send():378] send: config
+2024-05-15 12:56:58,003 DEBUG SenderThread:16168 [sender.py:send():378] send: metric
+2024-05-15 12:56:58,003 DEBUG SenderThread:16168 [sender.py:send():378] send: telemetry
+2024-05-15 12:56:58,003 DEBUG SenderThread:16168 [sender.py:send():378] send: telemetry
+2024-05-15 12:56:58,003 DEBUG SenderThread:16168 [sender.py:send():378] send: metric
+2024-05-15 12:56:58,003 WARNING SenderThread:16168 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-15 12:56:58,003 DEBUG SenderThread:16168 [sender.py:send():378] send: telemetry
+2024-05-15 12:56:58,003 DEBUG SenderThread:16168 [sender.py:send():378] send: telemetry
+2024-05-15 12:56:58,202 INFO wandb-upload_0:16168 [upload_job.py:push():130] Uploaded file /tmp/tmp5pu5dewxwandb/gnbk305h-wandb-metadata.json
+2024-05-15 12:57:02,004 DEBUG HandlerThread:16168 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:57:07,005 DEBUG HandlerThread:16168 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 12:57:07,041 INFO cpu :16168 [interfaces.py:monitor():138] Process cpu has exited.
+2024-05-15 12:57:07,043 DEBUG SystemMonitor:16168 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-15 12:57:07,043 DEBUG SystemMonitor:16168 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-15 12:57:07,043 DEBUG SystemMonitor:16168 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-15 12:57:07,048 DEBUG SenderThread:16168 [sender.py:send():378] send: stats
+2024-05-15 12:57:07,985 INFO MainThread:16168 [internal.py:handle_exit():75] Internal process exited
diff --git a/wandb/run-20240515_125656-5z8o4gdu/logs/debug.log b/wandb/run-20240515_125656-5z8o4gdu/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..5368438cd838b95cf49e2136ef4ed9f7fe6c518c
--- /dev/null
+++ b/wandb/run-20240515_125656-5z8o4gdu/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 12:56:56,432 INFO MainThread:15933 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 12:56:56,432 INFO MainThread:15933 [wandb_setup.py:_flush():76] Configure stats pid to 15933
+2024-05-15 12:56:56,432 INFO MainThread:15933 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 12:56:56,432 INFO MainThread:15933 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 12:56:56,432 WARNING MainThread:15933 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 12:56:56,432 INFO MainThread:15933 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 12:56:56,432 INFO MainThread:15933 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 12:56:56,432 WARNING MainThread:15933 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 12:56:56,432 INFO MainThread:15933 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 12:56:56,433 INFO MainThread:15933 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 12:56:56,433 INFO MainThread:15933 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_125656-5z8o4gdu/logs/debug.log
+2024-05-15 12:56:56,433 INFO MainThread:15933 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_125656-5z8o4gdu/logs/debug-internal.log
+2024-05-15 12:56:56,433 INFO MainThread:15933 [wandb_init.py:init():560] calling init triggers
+2024-05-15 12:56:56,433 INFO MainThread:15933 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 12:56:56,433 INFO MainThread:15933 [wandb_init.py:init():610] starting backend
+2024-05-15 12:56:56,433 INFO MainThread:15933 [wandb_init.py:init():614] setting up manager
+2024-05-15 12:56:56,434 INFO MainThread:15933 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 12:56:56,435 INFO MainThread:15933 [wandb_init.py:init():622] backend started and connected
+2024-05-15 12:56:56,439 INFO MainThread:15933 [wandb_init.py:init():711] updated telemetry
+2024-05-15 12:56:56,449 INFO MainThread:15933 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 12:56:56,897 INFO MainThread:15933 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 12:56:56,969 INFO MainThread:15933 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 12:56:56,970 INFO MainThread:15933 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 12:56:57,810 INFO MainThread:15933 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 12:56:57,810 INFO MainThread:15933 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 12:56:57,810 INFO MainThread:15933 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 12:56:57,810 INFO MainThread:15933 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 12:56:57,811 INFO MainThread:15933 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 12:56:57,814 INFO MainThread:15933 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': 4, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.95, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {'num_cycles': 3}, 'warmup_ratio': 0.0, 'warmup_steps': 717, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_12-56-29_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero1.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_bnb_8bit', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.96, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 2, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 12:56:57,815 INFO MainThread:15933 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_125656-5z8o4gdu/run-5z8o4gdu.wandb b/wandb/run-20240515_125656-5z8o4gdu/run-5z8o4gdu.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..d30ba6895e341fa4620c9c042bf5b510ea66fad8
Binary files /dev/null and b/wandb/run-20240515_125656-5z8o4gdu/run-5z8o4gdu.wandb differ
diff --git a/wandb/run-20240515_125924-zyixog64/files/config.yaml b/wandb/run-20240515_125924-zyixog64/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..a2ce83f980da7a115c303eb737f9c60dcb2196b6
--- /dev/null
+++ b/wandb/run-20240515_125924-zyixog64/files/config.yaml
@@ -0,0 +1,35 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715777964
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 23
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 13: linux-x86_64
diff --git a/wandb/run-20240515_125924-zyixog64/logs/debug-internal.log b/wandb/run-20240515_125924-zyixog64/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..01f42037686cd6a342a31aac07ce97df341cca3b
--- /dev/null
+++ b/wandb/run-20240515_125924-zyixog64/logs/debug-internal.log
@@ -0,0 +1,7 @@
+2024-05-15 12:59:24,340 INFO StreamThr :16602 [internal.py:wandb_internal():85] W&B internal server running at pid: 16602, started at: 2024-05-15 12:59:24.339694
+2024-05-15 12:59:24,343 DEBUG HandlerThread:16602 [handler.py:handle_request():158] handle_request: status
+2024-05-15 12:59:24,344 INFO WriterThread:16602 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_125924-zyixog64/run-zyixog64.wandb
+2024-05-15 12:59:24,344 DEBUG SenderThread:16602 [sender.py:send():378] send: header
+2024-05-15 12:59:24,351 DEBUG SenderThread:16602 [sender.py:send():378] send: run
+2024-05-15 12:59:29,352 DEBUG HandlerThread:16602 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 12:59:33,346 INFO MainThread:16602 [internal.py:handle_exit():75] Internal process exited
diff --git a/wandb/run-20240515_125924-zyixog64/logs/debug.log b/wandb/run-20240515_125924-zyixog64/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..7e0fbeb122ac21689c0b5781f17516a80ee492d2
--- /dev/null
+++ b/wandb/run-20240515_125924-zyixog64/logs/debug.log
@@ -0,0 +1,21 @@
+2024-05-15 12:59:24,335 INFO MainThread:16367 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 12:59:24,335 INFO MainThread:16367 [wandb_setup.py:_flush():76] Configure stats pid to 16367
+2024-05-15 12:59:24,335 INFO MainThread:16367 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 12:59:24,336 WARNING MainThread:16367 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 12:59:24,336 WARNING MainThread:16367 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_125924-zyixog64/logs/debug.log
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_125924-zyixog64/logs/debug-internal.log
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_init.py:init():560] calling init triggers
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_init.py:init():610] starting backend
+2024-05-15 12:59:24,336 INFO MainThread:16367 [wandb_init.py:init():614] setting up manager
+2024-05-15 12:59:24,338 INFO MainThread:16367 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 12:59:24,339 INFO MainThread:16367 [wandb_init.py:init():622] backend started and connected
+2024-05-15 12:59:24,342 INFO MainThread:16367 [wandb_init.py:init():711] updated telemetry
+2024-05-15 12:59:24,350 INFO MainThread:16367 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
diff --git a/wandb/run-20240515_125924-zyixog64/run-zyixog64.wandb b/wandb/run-20240515_125924-zyixog64/run-zyixog64.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391
diff --git a/wandb/run-20240515_130237-wp6u8fwb/files/conda-environment.yaml b/wandb/run-20240515_130237-wp6u8fwb/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_130237-wp6u8fwb/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_130237-wp6u8fwb/files/config.yaml b/wandb/run-20240515_130237-wp6u8fwb/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..c7f2a2c92126f22abbcaa651a6ae14ce7da13a96
--- /dev/null
+++ b/wandb/run-20240515_130237-wp6u8fwb/files/config.yaml
@@ -0,0 +1,739 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715778157
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 1
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.0.dev0
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./pyggel
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: false
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: 'no'
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 4
+per_device_eval_batch_size:
+ desc: null
+ value: 4
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 1
+eval_accumulation_steps:
+ desc: null
+ value: 1
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 5.0e-05
+weight_decay:
+ desc: null
+ value: 0.01
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.999
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 0.6
+num_train_epochs:
+ desc: null
+ value: 3
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value:
+ num_cycles: 3
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 1435
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./pyggel/runs/May15_13-01-56_d2827b6ba082
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.022222222222222223
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: false
+eval_steps:
+ desc: null
+ value: null
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./pyggel
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value: []
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: deepspeed_configs/zero2.json
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_bnb_8bit
+optim_args:
+ desc: null
+ value: adam_beta1=0.9,adam_beta2=0.95,amsgrad=True
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: false
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.4
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: 15.0
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: true
+sample_packing_efficiency:
+ desc: null
+ value: 0.98
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 4
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240515_130237-wp6u8fwb/files/output.log b/wandb/run-20240515_130237-wp6u8fwb/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..6e9c15415952717c0b53181462e80865040b6949
--- /dev/null
+++ b/wandb/run-20240515_130237-wp6u8fwb/files/output.log
@@ -0,0 +1,51 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/4398 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[2024-05-15 13:02:38,919] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:16778] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-15 13:02:39,605] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:16778] [RANK:0] packing_efficiency_estimate: 0.98 total_num_tokens per device: 47609834
+[2024-05-15 13:02:40,288] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:16778] [RANK:0] packing_efficiency_estimate: 0.98 total_num_tokens per device: 47609834
+Traceback (most recent call last):
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/runpy.py", line 196, in _run_module_as_main
+ return _run_code(code, main_globals, None,
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/runpy.py", line 86, in _run_code
+ exec(code, run_globals)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 70, in
+ fire.Fire(do_cli)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 143, in Fire
+ component_trace = _Fire(component, args, parsed_flag_args, context, name)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 477, in _Fire
+ component, remaining_args = _CallAndUpdateTrace(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 693, in _CallAndUpdateTrace
+ component = fn(*varargs, **kwargs)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 38, in do_cli
+ return do_train(parsed_cfg, parsed_cli_args)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 66, in do_train
+ return train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ File "/workspace/disk1/axolotl/src/axolotl/train.py", line 170, in train
+ trainer.train(resume_from_checkpoint=resume_from_checkpoint)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 1828, in train
+ return inner_training_loop(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 2181, in _inner_training_loop
+ tr_loss_step = self.training_step(model, inputs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 3116, in training_step
+ loss = self.compute_loss(model, inputs)
+ File "/workspace/disk1/axolotl/src/axolotl/core/trainer_builder.py", line 493, in compute_loss
+ return super().compute_loss(model, inputs, return_outputs=return_outputs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 3139, in compute_loss
+ outputs = model(**inputs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
+ return forward_call(*args, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 15, in wrapped_fn
+ ret_val = func(*args, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 1842, in forward
+ loss = self.module(*inputs, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
+ return self._call_impl(*args, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1568, in _call_impl
+ result = forward_call(*args, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 1209, in forward
+ logits = logits.float()
+torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 15.66 GiB. GPU 0 has a total capacty of 44.34 GiB of which 9.61 GiB is free. Process 1003248 has 34.72 GiB memory in use. Of the allocated memory 32.40 GiB is allocated by PyTorch, and 1.74 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
\ No newline at end of file
diff --git a/wandb/run-20240515_130237-wp6u8fwb/files/requirements.txt b/wandb/run-20240515_130237-wp6u8fwb/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_130237-wp6u8fwb/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_130237-wp6u8fwb/files/tmp/axolotl_config_aehzlufk.yml b/wandb/run-20240515_130237-wp6u8fwb/files/tmp/axolotl_config_aehzlufk.yml
new file mode 100644
index 0000000000000000000000000000000000000000..9b0d3204580fdde8d1bbe0537c6e9ed6a7bc6251
--- /dev/null
+++ b/wandb/run-20240515_130237-wp6u8fwb/files/tmp/axolotl_config_aehzlufk.yml
@@ -0,0 +1,66 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+push_to_hub: true
+hf_use_auth_token: true
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 4
+gradient_accumulation_steps: 1
+num_epochs: 3
+learning_rate: 0.00005
+optimizer: adamw_bnb_8bit
+optim_args:
+ adam_beta1: 0.9
+ adam_beta2: 0.95
+ amsgrad: true
+lr_scheduler: cosine
+lr_scheduler_kwargs:
+ num_cycles: 3
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.333
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero2.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240515_130237-wp6u8fwb/files/wandb-metadata.json b/wandb/run-20240515_130237-wp6u8fwb/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..dfc9aef223f51b287c06f3781158b6b534e1d5a5
--- /dev/null
+++ b/wandb/run-20240515_130237-wp6u8fwb/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T13:02:37.985292",
+ "startedAt": "2024-05-15T13:02:37.330237",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1564.5833333333333,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3393.947,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3399.998,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1700.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1599.839,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.40398788452148
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_130237-wp6u8fwb/files/wandb-summary.json b/wandb/run-20240515_130237-wp6u8fwb/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..a594e25ada167c5ad54a828e8a5b1a6223620b44
--- /dev/null
+++ b/wandb/run-20240515_130237-wp6u8fwb/files/wandb-summary.json
@@ -0,0 +1 @@
+{"_wandb": {"runtime": 8}}
\ No newline at end of file
diff --git a/wandb/run-20240515_130237-wp6u8fwb/logs/debug-internal.log b/wandb/run-20240515_130237-wp6u8fwb/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..a55a776a35a449dfab7b5cc8bdae473757be99d8
--- /dev/null
+++ b/wandb/run-20240515_130237-wp6u8fwb/logs/debug-internal.log
@@ -0,0 +1,174 @@
+2024-05-15 13:02:37,337 INFO StreamThr :17347 [internal.py:wandb_internal():85] W&B internal server running at pid: 17347, started at: 2024-05-15 13:02:37.336478
+2024-05-15 13:02:37,339 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: status
+2024-05-15 13:02:37,341 INFO WriterThread:17347 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/run-wp6u8fwb.wandb
+2024-05-15 13:02:37,342 DEBUG SenderThread:17347 [sender.py:send():378] send: header
+2024-05-15 13:02:37,350 DEBUG SenderThread:17347 [sender.py:send():378] send: run
+2024-05-15 13:02:37,746 INFO SenderThread:17347 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files
+2024-05-15 13:02:37,746 INFO SenderThread:17347 [sender.py:_start_run_threads():1123] run started: wp6u8fwb with start time 1715778157.336952
+2024-05-15 13:02:37,780 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 13:02:37,781 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: check_version
+2024-05-15 13:02:37,866 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 13:02:37,900 DEBUG HandlerThread:17347 [system_info.py:__init__():26] System info init
+2024-05-15 13:02:37,900 DEBUG HandlerThread:17347 [system_info.py:__init__():41] System info init done
+2024-05-15 13:02:37,900 INFO HandlerThread:17347 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 13:02:37,900 INFO SystemMonitor:17347 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 13:02:37,900 INFO HandlerThread:17347 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 13:02:37,901 INFO SystemMonitor:17347 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 13:02:37,902 INFO SystemMonitor:17347 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 13:02:37,903 INFO SystemMonitor:17347 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 13:02:37,906 INFO SystemMonitor:17347 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 13:02:37,907 INFO SystemMonitor:17347 [interfaces.py:start():188] Started network monitoring
+2024-05-15 13:02:37,985 DEBUG HandlerThread:17347 [system_info.py:probe():150] Probing system
+2024-05-15 13:02:37,987 DEBUG HandlerThread:17347 [system_info.py:_probe_git():135] Probing git
+2024-05-15 13:02:37,996 DEBUG HandlerThread:17347 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 13:02:37,996 DEBUG HandlerThread:17347 [system_info.py:probe():198] Probing system done
+2024-05-15 13:02:37,996 DEBUG HandlerThread:17347 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T13:02:37.985292', 'startedAt': '2024-05-15T13:02:37.330237', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1564.5833333333333, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3393.947, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3399.998, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 1200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1700.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2400.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 1599.839, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.40398788452148}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 13:02:37,996 INFO HandlerThread:17347 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 13:02:37,996 INFO HandlerThread:17347 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 13:02:37,996 DEBUG HandlerThread:17347 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 13:02:38,462 DEBUG HandlerThread:17347 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 13:02:38,462 DEBUG HandlerThread:17347 [system_info.py:_save_code():44] Saving code
+2024-05-15 13:02:38,462 WARNING HandlerThread:17347 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-15 13:02:38,462 DEBUG HandlerThread:17347 [system_info.py:_save_patches():82] Saving git patches
+2024-05-15 13:02:38,482 DEBUG HandlerThread:17347 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-15 13:02:38,485 INFO HandlerThread:17347 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 13:02:38,497 DEBUG SenderThread:17347 [sender.py:send():378] send: files
+2024-05-15 13:02:38,497 INFO SenderThread:17347 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 13:02:38,722 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 13:02:38,723 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 13:02:38,723 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:02:38,724 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:02:38,748 INFO Thread-12 :17347 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/requirements.txt
+2024-05-15 13:02:38,748 INFO Thread-12 :17347 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/wandb-metadata.json
+2024-05-15 13:02:38,749 INFO Thread-12 :17347 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/conda-environment.yaml
+2024-05-15 13:02:38,917 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-15 13:02:38,975 DEBUG SenderThread:17347 [sender.py:send():378] send: telemetry
+2024-05-15 13:02:38,975 DEBUG SenderThread:17347 [sender.py:send():378] send: config
+2024-05-15 13:02:38,976 DEBUG SenderThread:17347 [sender.py:send():378] send: metric
+2024-05-15 13:02:38,977 DEBUG SenderThread:17347 [sender.py:send():378] send: telemetry
+2024-05-15 13:02:38,977 DEBUG SenderThread:17347 [sender.py:send():378] send: telemetry
+2024-05-15 13:02:38,977 DEBUG SenderThread:17347 [sender.py:send():378] send: metric
+2024-05-15 13:02:38,977 WARNING SenderThread:17347 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-15 13:02:38,977 DEBUG SenderThread:17347 [sender.py:send():378] send: telemetry
+2024-05-15 13:02:38,977 DEBUG SenderThread:17347 [sender.py:send():378] send: telemetry
+2024-05-15 13:02:38,977 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: log_artifact
+2024-05-15 13:02:39,114 INFO wandb-upload_0:17347 [upload_job.py:push():130] Uploaded file /tmp/tmprewwkx9ewandb/anonligk-wandb-metadata.json
+2024-05-15 13:02:39,747 INFO Thread-12 :17347 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/tmp/axolotl_config_aehzlufk.yml
+2024-05-15 13:02:39,748 INFO Thread-12 :17347 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/tmp
+2024-05-15 13:02:40,160 INFO wandb-upload_0:17347 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmph6f07cw2
+2024-05-15 13:02:40,896 INFO SenderThread:17347 [sender.py:send_request_log_artifact():1455] logged artifact config-wp6u8fwb - {'id': 'QXJ0aWZhY3Q6ODM3NDAyMzUy', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTE2NDUzMw==', 'latestArtifact': None}}
+2024-05-15 13:02:40,898 DEBUG SenderThread:17347 [sender.py:send():378] send: telemetry
+2024-05-15 13:02:40,898 DEBUG SenderThread:17347 [sender.py:send():378] send: files
+2024-05-15 13:02:40,898 INFO SenderThread:17347 [sender.py:_save_file():1389] saving file tmp/axolotl_config_aehzlufk.yml with policy live
+2024-05-15 13:02:41,281 INFO wandb-upload_0:17347 [upload_job.py:push():130] Uploaded file /tmp/tmprewwkx9ewandb/ytu9cih6-tmp/axolotl_config_aehzlufk.yml
+2024-05-15 13:02:41,748 INFO Thread-12 :17347 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/output.log
+2024-05-15 13:02:43,113 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:02:43,750 INFO Thread-12 :17347 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/output.log
+2024-05-15 13:02:46,550 DEBUG SenderThread:17347 [sender.py:send():378] send: exit
+2024-05-15 13:02:46,550 INFO SenderThread:17347 [sender.py:send_exit():585] handling exit code: 1
+2024-05-15 13:02:46,550 INFO SenderThread:17347 [sender.py:send_exit():587] handling runtime: 8
+2024-05-15 13:02:46,553 INFO SenderThread:17347 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:02:46,554 INFO SenderThread:17347 [sender.py:send_exit():593] send defer
+2024-05-15 13:02:46,554 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:46,555 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 0
+2024-05-15 13:02:46,555 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:46,555 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 0
+2024-05-15 13:02:46,555 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 1
+2024-05-15 13:02:46,556 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:46,556 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 1
+2024-05-15 13:02:46,556 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:46,556 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 1
+2024-05-15 13:02:46,556 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 2
+2024-05-15 13:02:46,556 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:46,557 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 2
+2024-05-15 13:02:46,557 INFO HandlerThread:17347 [system_monitor.py:finish():203] Stopping system monitor
+2024-05-15 13:02:46,558 INFO HandlerThread:17347 [interfaces.py:finish():200] Joined cpu monitor
+2024-05-15 13:02:46,558 DEBUG SystemMonitor:17347 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-15 13:02:46,559 INFO HandlerThread:17347 [interfaces.py:finish():200] Joined disk monitor
+2024-05-15 13:02:46,559 DEBUG SystemMonitor:17347 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-15 13:02:46,560 DEBUG SystemMonitor:17347 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-15 13:02:46,675 INFO HandlerThread:17347 [interfaces.py:finish():200] Joined gpu monitor
+2024-05-15 13:02:46,676 INFO HandlerThread:17347 [interfaces.py:finish():200] Joined memory monitor
+2024-05-15 13:02:46,676 INFO HandlerThread:17347 [interfaces.py:finish():200] Joined network monitor
+2024-05-15 13:02:46,677 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:46,678 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 2
+2024-05-15 13:02:46,678 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 3
+2024-05-15 13:02:46,678 DEBUG SenderThread:17347 [sender.py:send():378] send: stats
+2024-05-15 13:02:46,680 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:46,680 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 3
+2024-05-15 13:02:46,680 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:46,680 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 3
+2024-05-15 13:02:46,680 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 4
+2024-05-15 13:02:46,681 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:46,681 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 4
+2024-05-15 13:02:46,681 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:46,681 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 4
+2024-05-15 13:02:46,681 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 5
+2024-05-15 13:02:46,682 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:46,682 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 5
+2024-05-15 13:02:46,682 DEBUG SenderThread:17347 [sender.py:send():378] send: summary
+2024-05-15 13:02:46,683 INFO SenderThread:17347 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:02:46,683 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:46,684 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 5
+2024-05-15 13:02:46,684 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 6
+2024-05-15 13:02:46,684 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:46,684 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 6
+2024-05-15 13:02:46,684 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:46,684 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 6
+2024-05-15 13:02:46,690 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:02:46,752 INFO Thread-12 :17347 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/wandb-summary.json
+2024-05-15 13:02:47,029 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 7
+2024-05-15 13:02:47,030 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:47,030 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 7
+2024-05-15 13:02:47,030 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:47,031 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 7
+2024-05-15 13:02:47,550 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:02:47,753 INFO Thread-12 :17347 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/config.yaml
+2024-05-15 13:02:49,755 INFO Thread-12 :17347 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/output.log
+2024-05-15 13:02:50,957 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 8
+2024-05-15 13:02:50,958 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:02:50,962 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:50,962 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 8
+2024-05-15 13:02:50,963 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:50,963 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 8
+2024-05-15 13:02:50,963 INFO SenderThread:17347 [job_builder.py:build():432] Attempting to build job artifact
+2024-05-15 13:02:50,968 INFO SenderThread:17347 [job_builder.py:_get_source_type():565] is repo sourced job
+2024-05-15 13:02:50,968 WARNING SenderThread:17347 [job_builder.py:_log_if_verbose():267] No program path found, not creating job artifact. See https://docs.wandb.ai/guides/launch/create-job
+2024-05-15 13:02:50,968 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 9
+2024-05-15 13:02:50,969 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:50,969 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 9
+2024-05-15 13:02:50,970 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:50,970 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 9
+2024-05-15 13:02:50,970 INFO SenderThread:17347 [dir_watcher.py:finish():358] shutting down directory watcher
+2024-05-15 13:02:51,552 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:02:51,757 INFO SenderThread:17347 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/output.log
+2024-05-15 13:02:51,757 INFO SenderThread:17347 [dir_watcher.py:finish():388] scan: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files
+2024-05-15 13:02:51,758 INFO SenderThread:17347 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/wandb-metadata.json wandb-metadata.json
+2024-05-15 13:02:51,758 INFO SenderThread:17347 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/wandb-summary.json wandb-summary.json
+2024-05-15 13:02:51,758 INFO SenderThread:17347 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/requirements.txt requirements.txt
+2024-05-15 13:02:51,759 INFO SenderThread:17347 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/config.yaml config.yaml
+2024-05-15 13:02:51,765 INFO SenderThread:17347 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/conda-environment.yaml conda-environment.yaml
+2024-05-15 13:02:51,766 INFO SenderThread:17347 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/output.log output.log
+2024-05-15 13:02:51,779 INFO SenderThread:17347 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/tmp/axolotl_config_aehzlufk.yml tmp/axolotl_config_aehzlufk.yml
+2024-05-15 13:02:51,780 INFO SenderThread:17347 [sender.py:transition_state():613] send defer: 10
+2024-05-15 13:02:51,780 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:02:51,781 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:02:51,781 INFO HandlerThread:17347 [handler.py:handle_request_defer():184] handle defer: 10
+2024-05-15 13:02:51,786 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:02:51,787 INFO SenderThread:17347 [sender.py:send_request_defer():609] handle sender defer: 10
+2024-05-15 13:02:51,788 INFO SenderThread:17347 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-15 13:02:52,248 INFO wandb-upload_1:17347 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/requirements.txt
+2024-05-15 13:02:52,270 INFO wandb-upload_3:17347 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/conda-environment.yaml
+2024-05-15 13:02:52,553 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:02:52,553 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:02:53,553 DEBUG HandlerThread:17347 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:02:53,554 DEBUG SenderThread:17347 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:02:54,358 WARNING StreamThr :17347 [internal.py:is_dead():413] Internal process exiting, parent pid 16778 disappeared
+2024-05-15 13:02:54,358 ERROR StreamThr :17347 [internal.py:wandb_internal():151] Internal process shutdown.
+2024-05-15 13:02:54,554 INFO HandlerThread:17347 [handler.py:finish():882] shutting down handler
+2024-05-15 13:02:54,554 INFO WriterThread:17347 [datastore.py:close():296] close: /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/run-wp6u8fwb.wandb
+2024-05-15 13:02:54,555 INFO SenderThread:17347 [sender.py:finish():1545] shutting down sender
+2024-05-15 13:02:54,556 INFO SenderThread:17347 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-15 13:02:54,556 INFO SenderThread:17347 [file_pusher.py:join():175] waiting for file pusher
+2024-05-15 13:03:02,254 INFO wandb-upload_0:17347 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/wandb-summary.json
+2024-05-15 13:03:02,316 INFO wandb-upload_4:17347 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/files/output.log
diff --git a/wandb/run-20240515_130237-wp6u8fwb/logs/debug.log b/wandb/run-20240515_130237-wp6u8fwb/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..8649bdda88913f2580cb5fc52c71ab5e21eb4c1f
--- /dev/null
+++ b/wandb/run-20240515_130237-wp6u8fwb/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 13:02:37,332 INFO MainThread:16778 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 13:02:37,332 INFO MainThread:16778 [wandb_setup.py:_flush():76] Configure stats pid to 16778
+2024-05-15 13:02:37,332 INFO MainThread:16778 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 13:02:37,332 INFO MainThread:16778 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 13:02:37,332 WARNING MainThread:16778 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 13:02:37,332 INFO MainThread:16778 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 13:02:37,332 INFO MainThread:16778 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 13:02:37,332 WARNING MainThread:16778 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 13:02:37,332 INFO MainThread:16778 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 13:02:37,332 INFO MainThread:16778 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 13:02:37,333 INFO MainThread:16778 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/logs/debug.log
+2024-05-15 13:02:37,333 INFO MainThread:16778 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_130237-wp6u8fwb/logs/debug-internal.log
+2024-05-15 13:02:37,333 INFO MainThread:16778 [wandb_init.py:init():560] calling init triggers
+2024-05-15 13:02:37,333 INFO MainThread:16778 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 13:02:37,333 INFO MainThread:16778 [wandb_init.py:init():610] starting backend
+2024-05-15 13:02:37,333 INFO MainThread:16778 [wandb_init.py:init():614] setting up manager
+2024-05-15 13:02:37,335 INFO MainThread:16778 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 13:02:37,336 INFO MainThread:16778 [wandb_init.py:init():622] backend started and connected
+2024-05-15 13:02:37,339 INFO MainThread:16778 [wandb_init.py:init():711] updated telemetry
+2024-05-15 13:02:37,349 INFO MainThread:16778 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 13:02:37,779 INFO MainThread:16778 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 13:02:37,851 INFO MainThread:16778 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 13:02:37,852 INFO MainThread:16778 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 13:02:38,723 INFO MainThread:16778 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 13:02:38,723 INFO MainThread:16778 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 13:02:38,723 INFO MainThread:16778 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 13:02:38,723 INFO MainThread:16778 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 13:02:38,724 INFO MainThread:16778 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 13:02:38,726 INFO MainThread:16778 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 1, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {'num_cycles': 3}, 'warmup_ratio': 0.0, 'warmup_steps': 1435, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_13-01-56_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero2.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_bnb_8bit', 'optim_args': 'adam_beta1=0.9,adam_beta2=0.95,amsgrad=True', 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.98, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 4, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 13:02:38,728 INFO MainThread:16778 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_130237-wp6u8fwb/run-wp6u8fwb.wandb b/wandb/run-20240515_130237-wp6u8fwb/run-wp6u8fwb.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..fe9ff3db8e76583f611ccbbbe048b933c3a3cd91
Binary files /dev/null and b/wandb/run-20240515_130237-wp6u8fwb/run-wp6u8fwb.wandb differ
diff --git a/wandb/run-20240515_130544-jg2l489k/files/conda-environment.yaml b/wandb/run-20240515_130544-jg2l489k/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_130544-jg2l489k/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_130544-jg2l489k/files/config.yaml b/wandb/run-20240515_130544-jg2l489k/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..e826c4ffd42facadffcb3640660f13108a02f80c
--- /dev/null
+++ b/wandb/run-20240515_130544-jg2l489k/files/config.yaml
@@ -0,0 +1,739 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715778344
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 1
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.0.dev0
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./pyggel
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: false
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: 'no'
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 2
+per_device_eval_batch_size:
+ desc: null
+ value: 2
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 2
+eval_accumulation_steps:
+ desc: null
+ value: 2
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 5.0e-05
+weight_decay:
+ desc: null
+ value: 0.01
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.999
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 0.6
+num_train_epochs:
+ desc: null
+ value: 3
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value:
+ num_cycles: 3
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 1435
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./pyggel/runs/May15_13-05-05_d2827b6ba082
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.022222222222222223
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: false
+eval_steps:
+ desc: null
+ value: null
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./pyggel
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value: []
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: deepspeed_configs/zero2.json
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_bnb_8bit
+optim_args:
+ desc: null
+ value: adam_beta1=0.9,adam_beta2=0.95,amsgrad=True
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: false
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.4
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: 15.0
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: true
+sample_packing_efficiency:
+ desc: null
+ value: 0.96
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 2
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240515_130544-jg2l489k/files/output.log b/wandb/run-20240515_130544-jg2l489k/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..84a65ddd5cd58067ad4421a19c641bf93b20f567
--- /dev/null
+++ b/wandb/run-20240515_130544-jg2l489k/files/output.log
@@ -0,0 +1,49 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/4491 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[2024-05-15 13:05:46,433] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:17467] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-15 13:05:47,076] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:17467] [RANK:0] packing_efficiency_estimate: 0.96 total_num_tokens per device: 47609834
+[2024-05-15 13:05:47,897] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:17467] [RANK:0] packing_efficiency_estimate: 0.96 total_num_tokens per device: 47609834
+Traceback (most recent call last):
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/runpy.py", line 196, in _run_module_as_main
+ return _run_code(code, main_globals, None,
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/runpy.py", line 86, in _run_code
+ exec(code, run_globals)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 70, in
+ fire.Fire(do_cli)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 143, in Fire
+ component_trace = _Fire(component, args, parsed_flag_args, context, name)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 477, in _Fire
+ component, remaining_args = _CallAndUpdateTrace(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/fire/core.py", line 693, in _CallAndUpdateTrace
+ component = fn(*varargs, **kwargs)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 38, in do_cli
+ return do_train(parsed_cfg, parsed_cli_args)
+ File "/workspace/disk1/axolotl/src/axolotl/cli/train.py", line 66, in do_train
+ return train(cfg=cfg, cli_args=cli_args, dataset_meta=dataset_meta)
+ File "/workspace/disk1/axolotl/src/axolotl/train.py", line 170, in train
+ trainer.train(resume_from_checkpoint=resume_from_checkpoint)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 1828, in train
+ return inner_training_loop(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 2181, in _inner_training_loop
+ tr_loss_step = self.training_step(model, inputs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/transformers/trainer.py", line 3125, in training_step
+ self.accelerator.backward(loss)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/accelerate/accelerator.py", line 1995, in backward
+ self.deepspeed_engine_wrapped.backward(loss, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/accelerate/utils/deepspeed.py", line 166, in backward
+ self.engine.backward(loss, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 15, in wrapped_fn
+ ret_val = func(*args, **kwargs)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 1964, in backward
+ self.optimizer.backward(loss, retain_graph=retain_graph)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/zero/stage_1_and_2.py", line 2040, in backward
+ self.loss_scaler.backward(loss.float(), retain_graph=retain_graph)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 63, in backward
+ scaled_loss.backward(retain_graph=retain_graph)
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/_tensor.py", line 492, in backward
+ torch.autograd.backward(
+ File "/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/autograd/__init__.py", line 251, in backward
+ Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
+torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 7.83 GiB. GPU 0 has a total capacty of 44.34 GiB of which 2.74 GiB is free. Process 1004690 has 41.59 GiB memory in use. Of the allocated memory 37.84 GiB is allocated by PyTorch, and 3.17 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
\ No newline at end of file
diff --git a/wandb/run-20240515_130544-jg2l489k/files/requirements.txt b/wandb/run-20240515_130544-jg2l489k/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_130544-jg2l489k/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_130544-jg2l489k/files/tmp/axolotl_config_wlhr25m3.yml b/wandb/run-20240515_130544-jg2l489k/files/tmp/axolotl_config_wlhr25m3.yml
new file mode 100644
index 0000000000000000000000000000000000000000..84f536aa57044eb653058027df0bafdd64bb1924
--- /dev/null
+++ b/wandb/run-20240515_130544-jg2l489k/files/tmp/axolotl_config_wlhr25m3.yml
@@ -0,0 +1,66 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+push_to_hub: true
+hf_use_auth_token: true
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 2
+gradient_accumulation_steps: 2
+num_epochs: 3
+learning_rate: 0.00005
+optimizer: adamw_bnb_8bit
+optim_args:
+ adam_beta1: 0.9
+ adam_beta2: 0.95
+ amsgrad: true
+lr_scheduler: cosine
+lr_scheduler_kwargs:
+ num_cycles: 3
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.333
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero2.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240515_130544-jg2l489k/files/wandb-metadata.json b/wandb/run-20240515_130544-jg2l489k/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..a7b4315b6c40ab1302221811b72863cfb5039670
--- /dev/null
+++ b/wandb/run-20240515_130544-jg2l489k/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T13:05:45.513881",
+ "startedAt": "2024-05-15T13:05:44.641356",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1369.7915833333336,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3301.25,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3301.255,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3301.256,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1599.999,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.003,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.40430450439453
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_130544-jg2l489k/files/wandb-summary.json b/wandb/run-20240515_130544-jg2l489k/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..74ae82ca002f26112178f0cd636ac5b92bf8e035
--- /dev/null
+++ b/wandb/run-20240515_130544-jg2l489k/files/wandb-summary.json
@@ -0,0 +1 @@
+{"_wandb": {"runtime": 6}}
\ No newline at end of file
diff --git a/wandb/run-20240515_130544-jg2l489k/logs/debug-internal.log b/wandb/run-20240515_130544-jg2l489k/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..eb06b11231e931c1f924043fa925a1ad4704cc78
--- /dev/null
+++ b/wandb/run-20240515_130544-jg2l489k/logs/debug-internal.log
@@ -0,0 +1,186 @@
+2024-05-15 13:05:44,649 INFO StreamThr :18066 [internal.py:wandb_internal():85] W&B internal server running at pid: 18066, started at: 2024-05-15 13:05:44.648771
+2024-05-15 13:05:44,651 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: status
+2024-05-15 13:05:44,653 INFO WriterThread:18066 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/run-jg2l489k.wandb
+2024-05-15 13:05:44,653 DEBUG SenderThread:18066 [sender.py:send():378] send: header
+2024-05-15 13:05:44,662 DEBUG SenderThread:18066 [sender.py:send():378] send: run
+2024-05-15 13:05:45,280 INFO SenderThread:18066 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files
+2024-05-15 13:05:45,281 INFO SenderThread:18066 [sender.py:_start_run_threads():1123] run started: jg2l489k with start time 1715778344.6488
+2024-05-15 13:05:45,312 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 13:05:45,312 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: check_version
+2024-05-15 13:05:45,400 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 13:05:45,431 DEBUG HandlerThread:18066 [system_info.py:__init__():26] System info init
+2024-05-15 13:05:45,432 DEBUG HandlerThread:18066 [system_info.py:__init__():41] System info init done
+2024-05-15 13:05:45,432 INFO HandlerThread:18066 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 13:05:45,432 INFO SystemMonitor:18066 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 13:05:45,432 INFO HandlerThread:18066 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 13:05:45,433 INFO SystemMonitor:18066 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 13:05:45,434 INFO SystemMonitor:18066 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 13:05:45,435 INFO SystemMonitor:18066 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 13:05:45,438 INFO SystemMonitor:18066 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 13:05:45,439 INFO SystemMonitor:18066 [interfaces.py:start():188] Started network monitoring
+2024-05-15 13:05:45,513 DEBUG HandlerThread:18066 [system_info.py:probe():150] Probing system
+2024-05-15 13:05:45,515 DEBUG HandlerThread:18066 [system_info.py:_probe_git():135] Probing git
+2024-05-15 13:05:45,524 DEBUG HandlerThread:18066 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 13:05:45,524 DEBUG HandlerThread:18066 [system_info.py:probe():198] Probing system done
+2024-05-15 13:05:45,524 DEBUG HandlerThread:18066 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T13:05:45.513881', 'startedAt': '2024-05-15T13:05:44.641356', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1369.7915833333336, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 3301.25, 'min': 800.0, 'max': 3500.0}, {'current': 3301.255, 'min': 800.0, 'max': 3500.0}, {'current': 3301.256, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 1599.999, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3100.0, 'min': 800.0, 'max': 3500.0}, {'current': 1300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 900.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.003, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.40430450439453}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 13:05:45,525 INFO HandlerThread:18066 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 13:05:45,525 INFO HandlerThread:18066 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 13:05:45,525 DEBUG HandlerThread:18066 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 13:05:45,990 DEBUG HandlerThread:18066 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 13:05:45,990 DEBUG HandlerThread:18066 [system_info.py:_save_code():44] Saving code
+2024-05-15 13:05:45,991 WARNING HandlerThread:18066 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-15 13:05:45,991 DEBUG HandlerThread:18066 [system_info.py:_save_patches():82] Saving git patches
+2024-05-15 13:05:46,012 DEBUG HandlerThread:18066 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-15 13:05:46,015 INFO HandlerThread:18066 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 13:05:46,026 DEBUG SenderThread:18066 [sender.py:send():378] send: files
+2024-05-15 13:05:46,027 INFO SenderThread:18066 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 13:05:46,240 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 13:05:46,240 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:05:46,241 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 13:05:46,243 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:05:46,286 INFO Thread-12 :18066 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/conda-environment.yaml
+2024-05-15 13:05:46,286 INFO Thread-12 :18066 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/requirements.txt
+2024-05-15 13:05:46,287 INFO Thread-12 :18066 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/wandb-metadata.json
+2024-05-15 13:05:46,431 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-15 13:05:46,479 DEBUG SenderThread:18066 [sender.py:send():378] send: telemetry
+2024-05-15 13:05:46,479 DEBUG SenderThread:18066 [sender.py:send():378] send: config
+2024-05-15 13:05:46,480 DEBUG SenderThread:18066 [sender.py:send():378] send: metric
+2024-05-15 13:05:46,481 DEBUG SenderThread:18066 [sender.py:send():378] send: telemetry
+2024-05-15 13:05:46,481 DEBUG SenderThread:18066 [sender.py:send():378] send: telemetry
+2024-05-15 13:05:46,481 DEBUG SenderThread:18066 [sender.py:send():378] send: metric
+2024-05-15 13:05:46,481 WARNING SenderThread:18066 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-15 13:05:46,481 DEBUG SenderThread:18066 [sender.py:send():378] send: telemetry
+2024-05-15 13:05:46,481 DEBUG SenderThread:18066 [sender.py:send():378] send: telemetry
+2024-05-15 13:05:46,481 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: log_artifact
+2024-05-15 13:05:46,646 INFO wandb-upload_0:18066 [upload_job.py:push():130] Uploaded file /tmp/tmpl9oouhnjwandb/ibm3f2yx-wandb-metadata.json
+2024-05-15 13:05:47,282 INFO Thread-12 :18066 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/tmp/axolotl_config_wlhr25m3.yml
+2024-05-15 13:05:47,283 INFO Thread-12 :18066 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/tmp
+2024-05-15 13:05:47,772 INFO wandb-upload_0:18066 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmp308p5yww
+2024-05-15 13:05:48,549 INFO SenderThread:18066 [sender.py:send_request_log_artifact():1455] logged artifact config-jg2l489k - {'id': 'QXJ0aWZhY3Q6ODM3NDA2MTQ3', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTE2NjQ0OQ==', 'latestArtifact': None}}
+2024-05-15 13:05:48,551 DEBUG SenderThread:18066 [sender.py:send():378] send: telemetry
+2024-05-15 13:05:48,552 DEBUG SenderThread:18066 [sender.py:send():378] send: files
+2024-05-15 13:05:48,552 INFO SenderThread:18066 [sender.py:_save_file():1389] saving file tmp/axolotl_config_wlhr25m3.yml with policy live
+2024-05-15 13:05:48,947 INFO wandb-upload_0:18066 [upload_job.py:push():130] Uploaded file /tmp/tmpl9oouhnjwandb/q3diu5f0-tmp/axolotl_config_wlhr25m3.yml
+2024-05-15 13:05:49,283 INFO Thread-12 :18066 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/output.log
+2024-05-15 13:05:50,652 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:05:51,285 INFO Thread-12 :18066 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/output.log
+2024-05-15 13:05:51,483 DEBUG SenderThread:18066 [sender.py:send():378] send: exit
+2024-05-15 13:05:51,483 INFO SenderThread:18066 [sender.py:send_exit():585] handling exit code: 1
+2024-05-15 13:05:51,483 INFO SenderThread:18066 [sender.py:send_exit():587] handling runtime: 6
+2024-05-15 13:05:51,483 INFO SenderThread:18066 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:05:51,484 INFO SenderThread:18066 [sender.py:send_exit():593] send defer
+2024-05-15 13:05:51,484 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:51,484 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 0
+2024-05-15 13:05:51,485 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:51,485 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 0
+2024-05-15 13:05:51,485 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 1
+2024-05-15 13:05:51,485 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:51,485 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 1
+2024-05-15 13:05:51,485 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:51,485 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 1
+2024-05-15 13:05:51,485 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 2
+2024-05-15 13:05:51,486 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:51,486 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 2
+2024-05-15 13:05:51,486 INFO HandlerThread:18066 [system_monitor.py:finish():203] Stopping system monitor
+2024-05-15 13:05:51,486 DEBUG SystemMonitor:18066 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-15 13:05:51,487 INFO HandlerThread:18066 [interfaces.py:finish():200] Joined cpu monitor
+2024-05-15 13:05:51,488 DEBUG SystemMonitor:18066 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-15 13:05:51,488 INFO HandlerThread:18066 [interfaces.py:finish():200] Joined disk monitor
+2024-05-15 13:05:51,488 DEBUG SystemMonitor:18066 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-15 13:05:51,778 INFO HandlerThread:18066 [interfaces.py:finish():200] Joined gpu monitor
+2024-05-15 13:05:51,778 INFO HandlerThread:18066 [interfaces.py:finish():200] Joined memory monitor
+2024-05-15 13:05:51,779 INFO HandlerThread:18066 [interfaces.py:finish():200] Joined network monitor
+2024-05-15 13:05:51,780 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:51,780 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 2
+2024-05-15 13:05:51,780 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 3
+2024-05-15 13:05:51,780 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:51,781 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 3
+2024-05-15 13:05:51,781 DEBUG SenderThread:18066 [sender.py:send():378] send: stats
+2024-05-15 13:05:51,783 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:51,783 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 3
+2024-05-15 13:05:51,783 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 4
+2024-05-15 13:05:51,784 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:51,784 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 4
+2024-05-15 13:05:51,784 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:51,784 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 4
+2024-05-15 13:05:51,785 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 5
+2024-05-15 13:05:51,785 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:51,785 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 5
+2024-05-15 13:05:51,786 DEBUG SenderThread:18066 [sender.py:send():378] send: summary
+2024-05-15 13:05:51,786 INFO SenderThread:18066 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:05:51,787 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:51,787 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 5
+2024-05-15 13:05:51,787 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 6
+2024-05-15 13:05:51,788 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:51,788 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 6
+2024-05-15 13:05:51,788 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:51,789 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 6
+2024-05-15 13:05:51,794 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:05:52,235 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 7
+2024-05-15 13:05:52,235 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:52,236 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 7
+2024-05-15 13:05:52,236 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:52,236 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 7
+2024-05-15 13:05:52,286 INFO Thread-12 :18066 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/config.yaml
+2024-05-15 13:05:52,286 INFO Thread-12 :18066 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/wandb-summary.json
+2024-05-15 13:05:52,481 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:05:53,286 INFO Thread-12 :18066 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/output.log
+2024-05-15 13:05:54,568 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 8
+2024-05-15 13:05:54,569 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:05:54,569 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:54,569 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 8
+2024-05-15 13:05:54,570 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:54,570 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 8
+2024-05-15 13:05:54,571 INFO SenderThread:18066 [job_builder.py:build():432] Attempting to build job artifact
+2024-05-15 13:05:54,574 INFO SenderThread:18066 [job_builder.py:_get_source_type():565] is repo sourced job
+2024-05-15 13:05:54,574 WARNING SenderThread:18066 [job_builder.py:_log_if_verbose():267] No program path found, not creating job artifact. See https://docs.wandb.ai/guides/launch/create-job
+2024-05-15 13:05:54,574 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 9
+2024-05-15 13:05:54,574 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:54,574 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 9
+2024-05-15 13:05:54,575 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:54,575 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 9
+2024-05-15 13:05:54,575 INFO SenderThread:18066 [dir_watcher.py:finish():358] shutting down directory watcher
+2024-05-15 13:05:55,289 INFO SenderThread:18066 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/output.log
+2024-05-15 13:05:55,290 INFO SenderThread:18066 [dir_watcher.py:finish():388] scan: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files
+2024-05-15 13:05:55,291 INFO SenderThread:18066 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/wandb-metadata.json wandb-metadata.json
+2024-05-15 13:05:55,291 INFO SenderThread:18066 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/wandb-summary.json wandb-summary.json
+2024-05-15 13:05:55,291 INFO SenderThread:18066 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/requirements.txt requirements.txt
+2024-05-15 13:05:55,292 INFO SenderThread:18066 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/config.yaml config.yaml
+2024-05-15 13:05:55,292 INFO SenderThread:18066 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/conda-environment.yaml conda-environment.yaml
+2024-05-15 13:05:55,292 INFO SenderThread:18066 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/output.log output.log
+2024-05-15 13:05:55,294 INFO SenderThread:18066 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/tmp/axolotl_config_wlhr25m3.yml tmp/axolotl_config_wlhr25m3.yml
+2024-05-15 13:05:55,294 INFO SenderThread:18066 [sender.py:transition_state():613] send defer: 10
+2024-05-15 13:05:55,295 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: defer
+2024-05-15 13:05:55,295 INFO HandlerThread:18066 [handler.py:handle_request_defer():184] handle defer: 10
+2024-05-15 13:05:55,304 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: defer
+2024-05-15 13:05:55,311 INFO SenderThread:18066 [sender.py:send_request_defer():609] handle sender defer: 10
+2024-05-15 13:05:55,315 INFO SenderThread:18066 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-15 13:05:55,483 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:05:55,484 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:05:55,684 INFO wandb-upload_1:18066 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/requirements.txt
+2024-05-15 13:05:55,745 INFO wandb-upload_2:18066 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/config.yaml
+2024-05-15 13:05:55,812 INFO wandb-upload_4:18066 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/output.log
+2024-05-15 13:05:56,485 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:05:56,486 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:05:57,485 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:05:57,486 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:05:58,487 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:05:58,487 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:05:59,487 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:05:59,488 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:06:00,488 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:06:00,489 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:06:01,490 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:06:01,490 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:06:01,492 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:06:01,669 WARNING StreamThr :18066 [internal.py:is_dead():413] Internal process exiting, parent pid 17467 disappeared
+2024-05-15 13:06:01,670 ERROR StreamThr :18066 [internal.py:wandb_internal():151] Internal process shutdown.
+2024-05-15 13:06:02,490 DEBUG HandlerThread:18066 [handler.py:handle_request():158] handle_request: poll_exit
+2024-05-15 13:06:02,490 INFO HandlerThread:18066 [handler.py:finish():882] shutting down handler
+2024-05-15 13:06:02,491 INFO WriterThread:18066 [datastore.py:close():296] close: /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/run-jg2l489k.wandb
+2024-05-15 13:06:02,491 DEBUG SenderThread:18066 [sender.py:send_request():405] send_request: poll_exit
+2024-05-15 13:06:02,492 INFO SenderThread:18066 [sender.py:finish():1545] shutting down sender
+2024-05-15 13:06:02,492 INFO SenderThread:18066 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-15 13:06:02,493 INFO SenderThread:18066 [file_pusher.py:join():175] waiting for file pusher
+2024-05-15 13:06:05,792 INFO wandb-upload_0:18066 [upload_job.py:push():130] Uploaded file /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/files/wandb-summary.json
+2024-05-15 13:06:39,245 WARNING FileStreamThread:18066 [file_stream.py:request_with_retry():674] requests_with_retry encountered retryable exception: HTTPSConnectionPool(host='api.wandb.ai', port=443): Max retries exceeded with url: /files/autometa/pyg-1-m-2.3/jg2l489k/file_stream (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)'))). func: functools.partial(>, timeout=180.0), args: ('https://api.wandb.ai/files/autometa/pyg-1-m-2.3/jg2l489k/file_stream',), kwargs: {'json': {'files': {'output.log': {'offset': 0, 'content': ['ERROR 2024-05-15T13:05:50.557105 \x1b[34m\x1b[1mwandb\x1b[39m\x1b[22m: \x1b[33mWARNING\x1b[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")\n', "ERROR 2024-05-15T13:05:50.557105 0%| | 0/4491 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.\n", 'ERROR 2024-05-15T13:05:50.557105 /root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.\n', 'ERROR 2024-05-15T13:05:50.557105 warnings.warn(\n', '2024-05-15T13:05:50.559751 [2024-05-15 13:05:46,433] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:17467] [RANK:0] The Axolotl config has been saved to the WandB run under files.\n', '2024-05-15T13:05:50.559751 [2024-05-15 13:05:47,076] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:17467] [RANK:0] packing_efficiency_estimate: 0.96 total_num_tokens per device: 47609834\n', '2024-05-15T13:05:50.559751 [2024-05-15 13:05:47,897] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:17467] [RANK:0] packing_efficiency_estimate: 0.96 total_num_tokens per device: 47609834\n']}}, 'dropped': 0}}
diff --git a/wandb/run-20240515_130544-jg2l489k/logs/debug.log b/wandb/run-20240515_130544-jg2l489k/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..752534f2f9fafcce236a603babff4ca9ee41facb
--- /dev/null
+++ b/wandb/run-20240515_130544-jg2l489k/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 13:05:44,643 INFO MainThread:17467 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 13:05:44,643 INFO MainThread:17467 [wandb_setup.py:_flush():76] Configure stats pid to 17467
+2024-05-15 13:05:44,643 INFO MainThread:17467 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 13:05:44,644 WARNING MainThread:17467 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 13:05:44,644 WARNING MainThread:17467 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/logs/debug.log
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_130544-jg2l489k/logs/debug-internal.log
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_init.py:init():560] calling init triggers
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_init.py:init():610] starting backend
+2024-05-15 13:05:44,644 INFO MainThread:17467 [wandb_init.py:init():614] setting up manager
+2024-05-15 13:05:44,646 INFO MainThread:17467 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 13:05:44,648 INFO MainThread:17467 [wandb_init.py:init():622] backend started and connected
+2024-05-15 13:05:44,652 INFO MainThread:17467 [wandb_init.py:init():711] updated telemetry
+2024-05-15 13:05:44,661 INFO MainThread:17467 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 13:05:45,311 INFO MainThread:17467 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 13:05:45,386 INFO MainThread:17467 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 13:05:45,387 INFO MainThread:17467 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 13:05:46,240 INFO MainThread:17467 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 13:05:46,240 INFO MainThread:17467 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 13:05:46,241 INFO MainThread:17467 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 13:05:46,241 INFO MainThread:17467 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 13:05:46,242 INFO MainThread:17467 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 13:05:46,245 INFO MainThread:17467 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 2, 'eval_accumulation_steps': 2, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {'num_cycles': 3}, 'warmup_ratio': 0.0, 'warmup_steps': 1435, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_13-05-05_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero2.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_bnb_8bit', 'optim_args': 'adam_beta1=0.9,adam_beta2=0.95,amsgrad=True', 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.96, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 2, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 13:05:46,246 INFO MainThread:17467 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_130544-jg2l489k/run-jg2l489k.wandb b/wandb/run-20240515_130544-jg2l489k/run-jg2l489k.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..d6ec56738448643b2d6de23068c30127fbd55a3d
Binary files /dev/null and b/wandb/run-20240515_130544-jg2l489k/run-jg2l489k.wandb differ
diff --git a/wandb/run-20240515_131101-qldtc48s/files/conda-environment.yaml b/wandb/run-20240515_131101-qldtc48s/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_131101-qldtc48s/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_131101-qldtc48s/files/config.yaml b/wandb/run-20240515_131101-qldtc48s/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..634d7e29cac5c914a9a82a75c9f3ee46c293d84f
--- /dev/null
+++ b/wandb/run-20240515_131101-qldtc48s/files/config.yaml
@@ -0,0 +1,35 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715778661
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 23
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 13: linux-x86_64
diff --git a/wandb/run-20240515_131101-qldtc48s/files/output.log b/wandb/run-20240515_131101-qldtc48s/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..52cf57ee251de0aecae9466b19b68dd37331f95f
--- /dev/null
+++ b/wandb/run-20240515_131101-qldtc48s/files/output.log
@@ -0,0 +1,6 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/4689 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[2024-05-15 13:11:02,587] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:18517] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-15 13:11:03,677] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:18517] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
diff --git a/wandb/run-20240515_131101-qldtc48s/files/requirements.txt b/wandb/run-20240515_131101-qldtc48s/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_131101-qldtc48s/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_131101-qldtc48s/files/tmp/axolotl_config_zegz9yk1.yml b/wandb/run-20240515_131101-qldtc48s/files/tmp/axolotl_config_zegz9yk1.yml
new file mode 100644
index 0000000000000000000000000000000000000000..6f2b639506a5430edf26a91b76cbba5f16dc43c7
--- /dev/null
+++ b/wandb/run-20240515_131101-qldtc48s/files/tmp/axolotl_config_zegz9yk1.yml
@@ -0,0 +1,66 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+push_to_hub: true
+hf_use_auth_token: true
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 1
+gradient_accumulation_steps: 4
+num_epochs: 3
+learning_rate: 0.00005
+optimizer: adamw_bnb_8bit
+optim_args:
+ adam_beta1: 0.9
+ adam_beta2: 0.95
+ amsgrad: true
+lr_scheduler: cosine
+lr_scheduler_kwargs:
+ num_cycles: 3
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.333
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero1.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240515_131101-qldtc48s/files/wandb-metadata.json b/wandb/run-20240515_131101-qldtc48s/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..ccc975dddcd1ad2a9b2345aafe0bf47a2871ba32
--- /dev/null
+++ b/wandb/run-20240515_131101-qldtc48s/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T13:11:01.694774",
+ "startedAt": "2024-05-15T13:11:01.016914",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1342.7082500000001,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2799.998,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3499.997,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3419.018,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3419.17,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3419.344,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.001,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.40483474731445
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_131101-qldtc48s/logs/debug-internal.log b/wandb/run-20240515_131101-qldtc48s/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..02f4ac4705cdd8e46f67e9e5b7d64c98252f212b
--- /dev/null
+++ b/wandb/run-20240515_131101-qldtc48s/logs/debug-internal.log
@@ -0,0 +1,80 @@
+2024-05-15 13:11:01,025 INFO StreamThr :18752 [internal.py:wandb_internal():85] W&B internal server running at pid: 18752, started at: 2024-05-15 13:11:01.024209
+2024-05-15 13:11:01,027 DEBUG HandlerThread:18752 [handler.py:handle_request():158] handle_request: status
+2024-05-15 13:11:01,030 INFO WriterThread:18752 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/run-qldtc48s.wandb
+2024-05-15 13:11:01,030 DEBUG SenderThread:18752 [sender.py:send():378] send: header
+2024-05-15 13:11:01,038 DEBUG SenderThread:18752 [sender.py:send():378] send: run
+2024-05-15 13:11:01,458 INFO SenderThread:18752 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/files
+2024-05-15 13:11:01,458 INFO SenderThread:18752 [sender.py:_start_run_threads():1123] run started: qldtc48s with start time 1715778661.025542
+2024-05-15 13:11:01,490 DEBUG HandlerThread:18752 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 13:11:01,490 DEBUG SenderThread:18752 [sender.py:send_request():405] send_request: check_version
+2024-05-15 13:11:01,578 DEBUG HandlerThread:18752 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 13:11:01,611 DEBUG HandlerThread:18752 [system_info.py:__init__():26] System info init
+2024-05-15 13:11:01,611 DEBUG HandlerThread:18752 [system_info.py:__init__():41] System info init done
+2024-05-15 13:11:01,612 INFO HandlerThread:18752 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 13:11:01,612 INFO SystemMonitor:18752 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 13:11:01,612 INFO HandlerThread:18752 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 13:11:01,613 INFO SystemMonitor:18752 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 13:11:01,614 INFO SystemMonitor:18752 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 13:11:01,616 INFO SystemMonitor:18752 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 13:11:01,616 INFO SystemMonitor:18752 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 13:11:01,618 INFO SystemMonitor:18752 [interfaces.py:start():188] Started network monitoring
+2024-05-15 13:11:01,694 DEBUG HandlerThread:18752 [system_info.py:probe():150] Probing system
+2024-05-15 13:11:01,697 DEBUG HandlerThread:18752 [system_info.py:_probe_git():135] Probing git
+2024-05-15 13:11:01,709 DEBUG HandlerThread:18752 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 13:11:01,710 DEBUG HandlerThread:18752 [system_info.py:probe():198] Probing system done
+2024-05-15 13:11:01,710 DEBUG HandlerThread:18752 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T13:11:01.694774', 'startedAt': '2024-05-15T13:11:01.016914', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1342.7082500000001, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2799.998, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3499.997, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3419.018, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 3419.17, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3419.344, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.001, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.40483474731445}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 13:11:01,710 INFO HandlerThread:18752 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 13:11:01,710 INFO HandlerThread:18752 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 13:11:01,710 DEBUG HandlerThread:18752 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 13:11:02,125 DEBUG HandlerThread:18752 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 13:11:02,128 INFO HandlerThread:18752 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 13:11:02,140 DEBUG SenderThread:18752 [sender.py:send():378] send: files
+2024-05-15 13:11:02,140 INFO SenderThread:18752 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 13:11:02,370 DEBUG HandlerThread:18752 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 13:11:02,370 DEBUG SenderThread:18752 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 13:11:02,371 DEBUG HandlerThread:18752 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:11:02,372 DEBUG SenderThread:18752 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:11:02,460 INFO Thread-12 :18752 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/files/wandb-metadata.json
+2024-05-15 13:11:02,460 INFO Thread-12 :18752 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/files/conda-environment.yaml
+2024-05-15 13:11:02,460 INFO Thread-12 :18752 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/files/requirements.txt
+2024-05-15 13:11:02,585 DEBUG HandlerThread:18752 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-15 13:11:02,627 DEBUG SenderThread:18752 [sender.py:send():378] send: telemetry
+2024-05-15 13:11:02,627 DEBUG SenderThread:18752 [sender.py:send():378] send: config
+2024-05-15 13:11:02,628 DEBUG SenderThread:18752 [sender.py:send():378] send: metric
+2024-05-15 13:11:02,629 DEBUG SenderThread:18752 [sender.py:send():378] send: telemetry
+2024-05-15 13:11:02,629 DEBUG SenderThread:18752 [sender.py:send():378] send: telemetry
+2024-05-15 13:11:02,629 DEBUG SenderThread:18752 [sender.py:send():378] send: metric
+2024-05-15 13:11:02,629 WARNING SenderThread:18752 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-15 13:11:02,629 DEBUG SenderThread:18752 [sender.py:send():378] send: telemetry
+2024-05-15 13:11:02,629 DEBUG SenderThread:18752 [sender.py:send():378] send: telemetry
+2024-05-15 13:11:02,629 DEBUG SenderThread:18752 [sender.py:send_request():405] send_request: log_artifact
+2024-05-15 13:11:02,779 INFO wandb-upload_0:18752 [upload_job.py:push():130] Uploaded file /tmp/tmp17ifgj1lwandb/s2uix5i1-wandb-metadata.json
+2024-05-15 13:11:03,458 INFO Thread-12 :18752 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/files/tmp/axolotl_config_zegz9yk1.yml
+2024-05-15 13:11:03,459 INFO Thread-12 :18752 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/files/tmp
+2024-05-15 13:11:03,860 INFO wandb-upload_0:18752 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmpjjgjxk4f
+2024-05-15 13:11:04,560 INFO SenderThread:18752 [sender.py:send_request_log_artifact():1455] logged artifact config-qldtc48s - {'id': 'QXJ0aWZhY3Q6ODM3NDEyODUy', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTE2OTg4Mw==', 'latestArtifact': None}}
+2024-05-15 13:11:04,562 DEBUG SenderThread:18752 [sender.py:send():378] send: telemetry
+2024-05-15 13:11:04,562 DEBUG SenderThread:18752 [sender.py:send():378] send: files
+2024-05-15 13:11:04,562 INFO SenderThread:18752 [sender.py:_save_file():1389] saving file tmp/axolotl_config_zegz9yk1.yml with policy live
+2024-05-15 13:11:04,948 INFO wandb-upload_0:18752 [upload_job.py:push():130] Uploaded file /tmp/tmp17ifgj1lwandb/6mgyr534-tmp/axolotl_config_zegz9yk1.yml
+2024-05-15 13:11:05,460 INFO Thread-12 :18752 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/files/output.log
+2024-05-15 13:11:06,959 DEBUG HandlerThread:18752 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:11:07,461 INFO Thread-12 :18752 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/files/output.log
+2024-05-15 13:11:09,631 INFO cpu :18752 [interfaces.py:monitor():138] Process cpu has exited.
+2024-05-15 13:11:09,633 DEBUG SystemMonitor:18752 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-15 13:11:09,633 DEBUG SystemMonitor:18752 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-15 13:11:09,633 DEBUG SystemMonitor:18752 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-15 13:11:09,643 INFO memory :18752 [interfaces.py:monitor():138] Process proc.memory.rssMB has exited.
+2024-05-15 13:11:09,644 DEBUG SenderThread:18752 [sender.py:send():378] send: stats
+2024-05-15 13:11:10,037 WARNING StreamThr :18752 [internal.py:is_dead():413] Internal process exiting, parent pid 18517 disappeared
+2024-05-15 13:11:10,037 ERROR StreamThr :18752 [internal.py:wandb_internal():151] Internal process shutdown.
+2024-05-15 13:11:10,638 INFO HandlerThread:18752 [handler.py:finish():882] shutting down handler
+2024-05-15 13:11:10,638 INFO HandlerThread:18752 [system_monitor.py:finish():203] Stopping system monitor
+2024-05-15 13:11:10,638 INFO HandlerThread:18752 [interfaces.py:finish():200] Joined cpu monitor
+2024-05-15 13:11:10,638 INFO HandlerThread:18752 [interfaces.py:finish():200] Joined disk monitor
+2024-05-15 13:11:10,639 INFO HandlerThread:18752 [interfaces.py:finish():200] Joined gpu monitor
+2024-05-15 13:11:10,639 INFO HandlerThread:18752 [interfaces.py:finish():200] Joined memory monitor
+2024-05-15 13:11:10,639 INFO HandlerThread:18752 [interfaces.py:finish():200] Joined network monitor
+2024-05-15 13:11:10,644 INFO WriterThread:18752 [datastore.py:close():296] close: /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/run-qldtc48s.wandb
+2024-05-15 13:11:10,646 INFO SenderThread:18752 [sender.py:finish():1545] shutting down sender
+2024-05-15 13:11:11,581 INFO MainThread:18752 [internal.py:handle_exit():75] Internal process exited
diff --git a/wandb/run-20240515_131101-qldtc48s/logs/debug.log b/wandb/run-20240515_131101-qldtc48s/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..29cde6611974d09d0dd04cf73c763b100288470f
--- /dev/null
+++ b/wandb/run-20240515_131101-qldtc48s/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 13:11:01,019 INFO MainThread:18517 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 13:11:01,019 INFO MainThread:18517 [wandb_setup.py:_flush():76] Configure stats pid to 18517
+2024-05-15 13:11:01,019 INFO MainThread:18517 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 13:11:01,019 INFO MainThread:18517 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 13:11:01,019 WARNING MainThread:18517 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 13:11:01,019 INFO MainThread:18517 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 13:11:01,019 INFO MainThread:18517 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 13:11:01,019 WARNING MainThread:18517 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 13:11:01,019 INFO MainThread:18517 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 13:11:01,019 INFO MainThread:18517 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 13:11:01,019 INFO MainThread:18517 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/logs/debug.log
+2024-05-15 13:11:01,020 INFO MainThread:18517 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_131101-qldtc48s/logs/debug-internal.log
+2024-05-15 13:11:01,020 INFO MainThread:18517 [wandb_init.py:init():560] calling init triggers
+2024-05-15 13:11:01,020 INFO MainThread:18517 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 13:11:01,020 INFO MainThread:18517 [wandb_init.py:init():610] starting backend
+2024-05-15 13:11:01,020 INFO MainThread:18517 [wandb_init.py:init():614] setting up manager
+2024-05-15 13:11:01,022 INFO MainThread:18517 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 13:11:01,025 INFO MainThread:18517 [wandb_init.py:init():622] backend started and connected
+2024-05-15 13:11:01,028 INFO MainThread:18517 [wandb_init.py:init():711] updated telemetry
+2024-05-15 13:11:01,038 INFO MainThread:18517 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 13:11:01,489 INFO MainThread:18517 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 13:11:01,566 INFO MainThread:18517 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 13:11:01,566 INFO MainThread:18517 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 13:11:02,370 INFO MainThread:18517 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 13:11:02,370 INFO MainThread:18517 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 13:11:02,370 INFO MainThread:18517 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 13:11:02,371 INFO MainThread:18517 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 13:11:02,372 INFO MainThread:18517 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 13:11:02,374 INFO MainThread:18517 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': 4, 'eval_delay': 0, 'learning_rate': 5e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {'num_cycles': 3}, 'warmup_ratio': 0.0, 'warmup_steps': 1436, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_13-10-27_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero1.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_bnb_8bit', 'optim_args': 'adam_beta1=0.9,adam_beta2=0.95,amsgrad=True', 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.92, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 1, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 13:11:02,376 INFO MainThread:18517 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_131101-qldtc48s/run-qldtc48s.wandb b/wandb/run-20240515_131101-qldtc48s/run-qldtc48s.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..8ffba10fc1691d78af829aba84dd6a53036ab9b7
Binary files /dev/null and b/wandb/run-20240515_131101-qldtc48s/run-qldtc48s.wandb differ
diff --git a/wandb/run-20240515_131923-4pdik1oq/files/conda-environment.yaml b/wandb/run-20240515_131923-4pdik1oq/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_131923-4pdik1oq/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_131923-4pdik1oq/files/config.yaml b/wandb/run-20240515_131923-4pdik1oq/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..23c31cd5f9baf6de680e521d3ec0c9422aa0f1e7
--- /dev/null
+++ b/wandb/run-20240515_131923-4pdik1oq/files/config.yaml
@@ -0,0 +1,754 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715779163
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 1
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+ - 1: train/loss
+ 5: 1
+ 6:
+ - 1
+ - 1: train/grad_norm
+ 5: 1
+ 6:
+ - 1
+ - 1: train/learning_rate
+ 5: 1
+ 6:
+ - 1
+ - 1: train/epoch
+ 5: 1
+ 6:
+ - 1
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.0.dev0
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./pyggel
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: false
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: 'no'
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 1
+per_device_eval_batch_size:
+ desc: null
+ value: 1
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 1
+eval_accumulation_steps:
+ desc: null
+ value: 1
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 0.0001
+weight_decay:
+ desc: null
+ value: 0.01
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.999
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 0.6
+num_train_epochs:
+ desc: null
+ value: 3
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value: {}
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 5746
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./pyggel/runs/May15_13-18-41_d2827b6ba082
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.022222222222222223
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: false
+eval_steps:
+ desc: null
+ value: null
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./pyggel
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value: []
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: deepspeed_configs/zero2.json
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_bnb_8bit
+optim_args:
+ desc: null
+ value: amsgrad=True
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: false
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.4
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: 15.0
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: true
+sample_packing_efficiency:
+ desc: null
+ value: 0.92
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 1
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240515_131923-4pdik1oq/files/output.log b/wandb/run-20240515_131923-4pdik1oq/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..e810ed2ce986379efb6e7fc5dd1a16187ff3e4a0
--- /dev/null
+++ b/wandb/run-20240515_131923-4pdik1oq/files/output.log
@@ -0,0 +1,16 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/18756 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[2024-05-15 13:19:25,279] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:19691] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-15 13:19:26,274] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:19691] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+[2024-05-15 13:19:27,246] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:19691] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/zero/stage_1_and_2.py:1290: UserWarning: The torch.cuda.*DtypeTensor constructors are no longer recommended. It's best to use methods such as torch.tensor(data, dtype=*, device='cuda') to create tensors. (Triggered internally at ../torch/csrc/tensor/python_tensor.cpp:83.)
+ total_norm_cuda = get_accelerator().FloatTensor([float(total_norm)])
+{'loss': 1.9946, 'grad_norm': 78.64436429367491, 'learning_rate': 1.740341106856944e-08, 'epoch': 0.0}
+
+ 0%| | 2/18756 [00:36<92:53:39, 17.83s/it]
+[2024-05-15 13:20:01,476] [INFO] [axolotl.callbacks.on_step_end:125] [PID:19691] [RANK:0] GPU memory usage while training: 16.077GB (+25.159GB cache, +1.247GB misc)
+
+ 0%| | 3/18756 [00:55<96:31:29, 18.53s/it]
+{'loss': 2.1278, 'grad_norm': 39.31913664916789, 'learning_rate': 5.2210233205708324e-08, 'epoch': 0.0}
\ No newline at end of file
diff --git a/wandb/run-20240515_131923-4pdik1oq/files/requirements.txt b/wandb/run-20240515_131923-4pdik1oq/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_131923-4pdik1oq/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_131923-4pdik1oq/files/tmp/axolotl_config_2xva31_6.yml b/wandb/run-20240515_131923-4pdik1oq/files/tmp/axolotl_config_2xva31_6.yml
new file mode 100644
index 0000000000000000000000000000000000000000..2956e466118feb05e9a2f9ddb4056b3f6789d12d
--- /dev/null
+++ b/wandb/run-20240515_131923-4pdik1oq/files/tmp/axolotl_config_2xva31_6.yml
@@ -0,0 +1,62 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+push_to_hub: true
+hf_use_auth_token: true
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 1
+gradient_accumulation_steps: 1
+num_epochs: 3
+learning_rate: 0.0001
+optimizer: adamw_bnb_8bit
+optim_args:
+ amsgrad: true
+lr_scheduler: cosine
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.333
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero2.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240515_131923-4pdik1oq/files/wandb-metadata.json b/wandb/run-20240515_131923-4pdik1oq/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..4b1c2e7d19ba2357de8089104040f51fad95ff9f
--- /dev/null
+++ b/wandb/run-20240515_131923-4pdik1oq/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T13:19:24.343311",
+ "startedAt": "2024-05-15T13:19:23.699372",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1436.462427083333,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.002,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.001,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.40570068359375
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_131923-4pdik1oq/files/wandb-summary.json b/wandb/run-20240515_131923-4pdik1oq/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..8ae3ff999312dbaefcebbff2f88b10573175d670
--- /dev/null
+++ b/wandb/run-20240515_131923-4pdik1oq/files/wandb-summary.json
@@ -0,0 +1 @@
+{"train/loss": 2.1278, "train/grad_norm": 39.31913664916789, "train/learning_rate": 5.2210233205708324e-08, "train/epoch": 0.0, "train/global_step": 3, "_timestamp": 1715779220.8392699, "_runtime": 57.13422393798828, "_step": 2}
\ No newline at end of file
diff --git a/wandb/run-20240515_131923-4pdik1oq/logs/debug-internal.log b/wandb/run-20240515_131923-4pdik1oq/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..8698aad16745f6c532c5db013217bbdf7ddbd335
--- /dev/null
+++ b/wandb/run-20240515_131923-4pdik1oq/logs/debug-internal.log
@@ -0,0 +1,141 @@
+2024-05-15 13:19:23,706 INFO StreamThr :20266 [internal.py:wandb_internal():85] W&B internal server running at pid: 20266, started at: 2024-05-15 13:19:23.705034
+2024-05-15 13:19:23,708 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status
+2024-05-15 13:19:23,711 INFO WriterThread:20266 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/run-4pdik1oq.wandb
+2024-05-15 13:19:23,711 DEBUG SenderThread:20266 [sender.py:send():378] send: header
+2024-05-15 13:19:23,717 DEBUG SenderThread:20266 [sender.py:send():378] send: run
+2024-05-15 13:19:24,109 INFO SenderThread:20266 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files
+2024-05-15 13:19:24,109 INFO SenderThread:20266 [sender.py:_start_run_threads():1123] run started: 4pdik1oq with start time 1715779163.705046
+2024-05-15 13:19:24,146 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 13:19:24,147 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: check_version
+2024-05-15 13:19:24,234 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 13:19:24,268 DEBUG HandlerThread:20266 [system_info.py:__init__():26] System info init
+2024-05-15 13:19:24,268 DEBUG HandlerThread:20266 [system_info.py:__init__():41] System info init done
+2024-05-15 13:19:24,268 INFO HandlerThread:20266 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 13:19:24,268 INFO SystemMonitor:20266 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 13:19:24,269 INFO HandlerThread:20266 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 13:19:24,269 INFO SystemMonitor:20266 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 13:19:24,270 INFO SystemMonitor:20266 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 13:19:24,271 INFO SystemMonitor:20266 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 13:19:24,274 INFO SystemMonitor:20266 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 13:19:24,275 INFO SystemMonitor:20266 [interfaces.py:start():188] Started network monitoring
+2024-05-15 13:19:24,343 DEBUG HandlerThread:20266 [system_info.py:probe():150] Probing system
+2024-05-15 13:19:24,346 DEBUG HandlerThread:20266 [system_info.py:_probe_git():135] Probing git
+2024-05-15 13:19:24,356 DEBUG HandlerThread:20266 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 13:19:24,356 DEBUG HandlerThread:20266 [system_info.py:probe():198] Probing system done
+2024-05-15 13:19:24,356 DEBUG HandlerThread:20266 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T13:19:24.343311', 'startedAt': '2024-05-15T13:19:23.699372', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1436.462427083333, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.002, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.001, 'min': 800.0, 'max': 3500.0}, {'current': 2800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.40570068359375}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 13:19:24,357 INFO HandlerThread:20266 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 13:19:24,357 INFO HandlerThread:20266 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 13:19:24,357 DEBUG HandlerThread:20266 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 13:19:24,822 DEBUG HandlerThread:20266 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 13:19:24,822 DEBUG HandlerThread:20266 [system_info.py:_save_code():44] Saving code
+2024-05-15 13:19:24,822 WARNING HandlerThread:20266 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-15 13:19:24,822 DEBUG HandlerThread:20266 [system_info.py:_save_patches():82] Saving git patches
+2024-05-15 13:19:24,843 DEBUG HandlerThread:20266 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-15 13:19:24,844 INFO HandlerThread:20266 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 13:19:24,851 DEBUG SenderThread:20266 [sender.py:send():378] send: files
+2024-05-15 13:19:24,851 INFO SenderThread:20266 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 13:19:25,077 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 13:19:25,077 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 13:19:25,078 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:19:25,079 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:19:25,111 INFO Thread-12 :20266 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/wandb-metadata.json
+2024-05-15 13:19:25,112 INFO Thread-12 :20266 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/conda-environment.yaml
+2024-05-15 13:19:25,112 INFO Thread-12 :20266 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/requirements.txt
+2024-05-15 13:19:25,277 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-15 13:19:25,329 DEBUG SenderThread:20266 [sender.py:send():378] send: telemetry
+2024-05-15 13:19:25,329 DEBUG SenderThread:20266 [sender.py:send():378] send: config
+2024-05-15 13:19:25,331 DEBUG SenderThread:20266 [sender.py:send():378] send: metric
+2024-05-15 13:19:25,331 DEBUG SenderThread:20266 [sender.py:send():378] send: telemetry
+2024-05-15 13:19:25,331 DEBUG SenderThread:20266 [sender.py:send():378] send: telemetry
+2024-05-15 13:19:25,331 DEBUG SenderThread:20266 [sender.py:send():378] send: metric
+2024-05-15 13:19:25,331 WARNING SenderThread:20266 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-15 13:19:25,331 DEBUG SenderThread:20266 [sender.py:send():378] send: telemetry
+2024-05-15 13:19:25,332 DEBUG SenderThread:20266 [sender.py:send():378] send: telemetry
+2024-05-15 13:19:25,332 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: log_artifact
+2024-05-15 13:19:26,110 INFO Thread-12 :20266 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/tmp/axolotl_config_2xva31_6.yml
+2024-05-15 13:19:26,111 INFO Thread-12 :20266 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/tmp
+2024-05-15 13:19:26,590 INFO wandb-upload_1:20266 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmp2v2gj0fx
+2024-05-15 13:19:27,324 INFO SenderThread:20266 [sender.py:send_request_log_artifact():1455] logged artifact config-4pdik1oq - {'id': 'QXJ0aWZhY3Q6ODM3NDIyMDgz', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTE3MzA5OQ==', 'latestArtifact': None}}
+2024-05-15 13:19:27,325 DEBUG SenderThread:20266 [sender.py:send():378] send: telemetry
+2024-05-15 13:19:27,326 DEBUG SenderThread:20266 [sender.py:send():378] send: files
+2024-05-15 13:19:27,326 INFO SenderThread:20266 [sender.py:_save_file():1389] saving file tmp/axolotl_config_2xva31_6.yml with policy live
+2024-05-15 13:19:27,686 INFO wandb-upload_1:20266 [upload_job.py:push():130] Uploaded file /tmp/tmpjhq_o5olwandb/y5q8eyq1-tmp/axolotl_config_2xva31_6.yml
+2024-05-15 13:19:28,111 INFO Thread-12 :20266 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/output.log
+2024-05-15 13:19:29,237 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:19:30,112 INFO Thread-12 :20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/output.log
+2024-05-15 13:19:34,238 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:19:39,239 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:19:41,526 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:19:41,527 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:19:42,120 INFO Thread-12 :20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/output.log
+2024-05-15 13:19:44,695 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:19:44,938 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:19:44,942 DEBUG SenderThread:20266 [sender.py:send():378] send: metric
+2024-05-15 13:19:44,944 DEBUG SenderThread:20266 [sender.py:send():378] send: metric
+2024-05-15 13:19:44,944 DEBUG SenderThread:20266 [sender.py:send():378] send: metric
+2024-05-15 13:19:44,944 DEBUG SenderThread:20266 [sender.py:send():378] send: metric
+2024-05-15 13:19:44,944 DEBUG SenderThread:20266 [sender.py:send():378] send: history
+2024-05-15 13:19:44,945 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:19:44,948 INFO SenderThread:20266 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:19:45,122 INFO Thread-12 :20266 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/wandb-summary.json
+2024-05-15 13:19:46,123 INFO Thread-12 :20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/output.log
+2024-05-15 13:19:48,124 INFO Thread-12 :20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/output.log
+2024-05-15 13:19:49,950 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:19:54,955 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:19:56,129 INFO Thread-12 :20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/config.yaml
+2024-05-15 13:19:56,526 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:19:56,527 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:20:00,684 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:20:01,479 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:20:01,481 DEBUG SenderThread:20266 [sender.py:send():378] send: history
+2024-05-15 13:20:01,481 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:20:01,486 INFO SenderThread:20266 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:20:02,134 INFO Thread-12 :20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/wandb-summary.json
+2024-05-15 13:20:04,135 INFO Thread-12 :20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/output.log
+2024-05-15 13:20:06,488 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:20:11,489 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:20:11,526 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:20:11,526 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:20:16,768 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:20:20,839 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:20:20,842 DEBUG SenderThread:20266 [sender.py:send():378] send: history
+2024-05-15 13:20:20,843 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:20:20,847 INFO SenderThread:20266 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:20:21,146 INFO Thread-12 :20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/wandb-summary.json
+2024-05-15 13:20:21,848 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:20:22,147 INFO Thread-12 :20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/output.log
+2024-05-15 13:20:24,275 DEBUG SystemMonitor:20266 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-15 13:20:24,278 DEBUG SenderThread:20266 [sender.py:send():378] send: stats
+2024-05-15 13:20:26,526 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:20:26,527 DEBUG SenderThread:20266 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:20:27,687 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:20:32,310 INFO memory :20266 [interfaces.py:monitor():138] Process proc.memory.rssMB has exited.
+2024-05-15 13:20:32,310 DEBUG SystemMonitor:20266 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-15 13:20:32,311 DEBUG SystemMonitor:20266 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-15 13:20:32,323 DEBUG SenderThread:20266 [sender.py:send():378] send: stats
+2024-05-15 13:20:33,326 DEBUG HandlerThread:20266 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:20:36,786 WARNING StreamThr :20266 [internal.py:is_dead():413] Internal process exiting, parent pid 19691 disappeared
+2024-05-15 13:20:36,786 ERROR StreamThr :20266 [internal.py:wandb_internal():151] Internal process shutdown.
+2024-05-15 13:20:37,326 INFO SenderThread:20266 [sender.py:finish():1545] shutting down sender
+2024-05-15 13:20:37,327 INFO HandlerThread:20266 [handler.py:finish():882] shutting down handler
+2024-05-15 13:20:37,327 INFO HandlerThread:20266 [system_monitor.py:finish():203] Stopping system monitor
+2024-05-15 13:20:37,327 INFO HandlerThread:20266 [interfaces.py:finish():200] Joined cpu monitor
+2024-05-15 13:20:37,327 INFO HandlerThread:20266 [interfaces.py:finish():200] Joined disk monitor
+2024-05-15 13:20:37,328 INFO HandlerThread:20266 [interfaces.py:finish():200] Joined gpu monitor
+2024-05-15 13:20:37,328 INFO HandlerThread:20266 [interfaces.py:finish():200] Joined memory monitor
+2024-05-15 13:20:37,329 INFO HandlerThread:20266 [interfaces.py:finish():200] Joined network monitor
+2024-05-15 13:20:37,329 INFO WriterThread:20266 [datastore.py:close():296] close: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/run-4pdik1oq.wandb
+2024-05-15 13:20:39,443 INFO SenderThread:20266 [dir_watcher.py:finish():358] shutting down directory watcher
+2024-05-15 13:20:40,151 INFO SenderThread:20266 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/output.log
+2024-05-15 13:20:40,152 INFO SenderThread:20266 [dir_watcher.py:finish():388] scan: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files
+2024-05-15 13:20:40,152 INFO SenderThread:20266 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/wandb-metadata.json wandb-metadata.json
+2024-05-15 13:20:40,153 INFO SenderThread:20266 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/wandb-summary.json wandb-summary.json
+2024-05-15 13:20:40,153 INFO SenderThread:20266 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/requirements.txt requirements.txt
+2024-05-15 13:20:40,154 INFO SenderThread:20266 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/config.yaml config.yaml
+2024-05-15 13:20:40,154 INFO SenderThread:20266 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/conda-environment.yaml conda-environment.yaml
+2024-05-15 13:20:40,155 INFO SenderThread:20266 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/output.log output.log
+2024-05-15 13:20:40,155 INFO SenderThread:20266 [dir_watcher.py:finish():402] scan save: /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/files/tmp/axolotl_config_2xva31_6.yml tmp/axolotl_config_2xva31_6.yml
+2024-05-15 13:20:40,156 INFO SenderThread:20266 [file_pusher.py:finish():169] shutting down file pusher
+2024-05-15 13:20:40,156 INFO SenderThread:20266 [file_pusher.py:join():175] waiting for file pusher
+2024-05-15 13:20:40,657 INFO SenderThread:20266 [file_stream.py:finish():601] file stream finish called
+2024-05-15 13:20:40,977 INFO SenderThread:20266 [file_stream.py:finish():605] file stream finish is done
diff --git a/wandb/run-20240515_131923-4pdik1oq/logs/debug.log b/wandb/run-20240515_131923-4pdik1oq/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..353516877c19d90a002703e7914d0754a3f0e6d0
--- /dev/null
+++ b/wandb/run-20240515_131923-4pdik1oq/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_setup.py:_flush():76] Configure stats pid to 19691
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 13:19:23,701 WARNING MainThread:19691 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 13:19:23,701 WARNING MainThread:19691 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/logs/debug.log
+2024-05-15 13:19:23,701 INFO MainThread:19691 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_131923-4pdik1oq/logs/debug-internal.log
+2024-05-15 13:19:23,702 INFO MainThread:19691 [wandb_init.py:init():560] calling init triggers
+2024-05-15 13:19:23,702 INFO MainThread:19691 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 13:19:23,702 INFO MainThread:19691 [wandb_init.py:init():610] starting backend
+2024-05-15 13:19:23,702 INFO MainThread:19691 [wandb_init.py:init():614] setting up manager
+2024-05-15 13:19:23,703 INFO MainThread:19691 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 13:19:23,704 INFO MainThread:19691 [wandb_init.py:init():622] backend started and connected
+2024-05-15 13:19:23,707 INFO MainThread:19691 [wandb_init.py:init():711] updated telemetry
+2024-05-15 13:19:23,716 INFO MainThread:19691 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 13:19:24,145 INFO MainThread:19691 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 13:19:24,218 INFO MainThread:19691 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 13:19:24,219 INFO MainThread:19691 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 13:19:25,078 INFO MainThread:19691 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 13:19:25,078 INFO MainThread:19691 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 13:19:25,078 INFO MainThread:19691 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 13:19:25,078 INFO MainThread:19691 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 13:19:25,079 INFO MainThread:19691 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 13:19:25,082 INFO MainThread:19691 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 1, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 5746, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_13-18-41_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero2.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_bnb_8bit', 'optim_args': 'amsgrad=True', 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.92, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 1, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 13:19:25,084 INFO MainThread:19691 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_131923-4pdik1oq/run-4pdik1oq.wandb b/wandb/run-20240515_131923-4pdik1oq/run-4pdik1oq.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..05417876d2f12f35d2daf20f7664c7ef7d67fe02
Binary files /dev/null and b/wandb/run-20240515_131923-4pdik1oq/run-4pdik1oq.wandb differ
diff --git a/wandb/run-20240515_132302-ky6h7dv9/files/conda-environment.yaml b/wandb/run-20240515_132302-ky6h7dv9/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240515_132302-ky6h7dv9/files/config.yaml b/wandb/run-20240515_132302-ky6h7dv9/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..aa7fb06fcf3a8bb7402e916e71d1f0b948d0b218
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/files/config.yaml
@@ -0,0 +1,6970 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.0.dev0
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715779382
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 1
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.0.dev0
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+ - 1: train/loss
+ 5: 1
+ 6:
+ - 1
+ - 1: train/grad_norm
+ 5: 1
+ 6:
+ - 1
+ - 1: train/learning_rate
+ 5: 1
+ 6:
+ - 1
+ - 1: train/epoch
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.embed_tokens\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.embed_tokens\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.embed_tokens\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.0\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.1\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.2\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.3\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.4\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.5\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.6\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.7\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.8\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.9\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.10\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.11\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.12\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.13\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.14\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.15\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.16\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.17\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.18\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.19\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.20\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.21\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.22\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.23\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.24\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.25\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.26\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.27\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.28\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.29\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.30\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.layers\.31\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.norm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.norm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/model\.norm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/lm_head\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/lm_head\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: parameters/lm_head\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/lm_head\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/lm_head\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/lm_head\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.norm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.norm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.norm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.31\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.30\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.29\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.28\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.27\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.26\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.25\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.24\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.23\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.22\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.21\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.20\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.19\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.18\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.17\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.16\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.15\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.14\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.13\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.12\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.11\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.10\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.9\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.8\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.7\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.6\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.5\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.4\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.3\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.2\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.1\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.mlp\.swiglu\.w3\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.mlp\.swiglu\.w3\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.mlp\.swiglu\.w3\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.mlp\.swiglu\.w12\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.mlp\.swiglu\.w12\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.mlp\.swiglu\.w12\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.post_attention_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.post_attention_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.post_attention_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.o_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.o_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.o_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.v_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.v_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.v_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.k_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.k_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.k_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.q_proj\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.q_proj\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.self_attn\.q_proj\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.input_layernorm\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.input_layernorm\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.layers\.0\.input_layernorm\.weight.bins
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.embed_tokens\.weight._type
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.embed_tokens\.weight.values
+ 5: 1
+ 6:
+ - 1
+ - 1: gradients/model\.embed_tokens\.weight.bins
+ 5: 1
+ 6:
+ - 1
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.0.dev0
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./pyggel
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: false
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: 'no'
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 1
+per_device_eval_batch_size:
+ desc: null
+ value: 1
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 1
+eval_accumulation_steps:
+ desc: null
+ value: 1
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 0.0001
+weight_decay:
+ desc: null
+ value: 0.01
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.999
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 0.6
+num_train_epochs:
+ desc: null
+ value: 3
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value: {}
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 8628
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./pyggel/runs/May15_13-22-23_d2827b6ba082
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.022222222222222223
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: false
+eval_steps:
+ desc: null
+ value: null
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./pyggel
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value: []
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: deepspeed_configs/zero2.json
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_bnb_8bit
+optim_args:
+ desc: null
+ value: amsgrad=True
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: false
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.4
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: 15.0
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: true
+sample_packing_efficiency:
+ desc: null
+ value: 0.92
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 1
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240515_132302-ky6h7dv9/files/output.log b/wandb/run-20240515_132302-ky6h7dv9/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..b334fc0b7a45cca6d982f864c9128a39b2aa370e
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/files/output.log
@@ -0,0 +1,2522 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/18756 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[2024-05-15 13:23:04,109] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:20482] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-15 13:23:05,212] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:20482] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+[2024-05-15 13:23:06,317] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:20482] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/deepspeed/runtime/zero/stage_1_and_2.py:1290: UserWarning: The torch.cuda.*DtypeTensor constructors are no longer recommended. It's best to use methods such as torch.tensor(data, dtype=*, device='cuda') to create tensors. (Triggered internally at ../torch/csrc/tensor/python_tensor.cpp:83.)
+ total_norm_cuda = get_accelerator().FloatTensor([float(total_norm)])
+{'loss': 1.9946, 'grad_norm': 66.66815997676477, 'learning_rate': 1.1590171534538711e-08, 'epoch': 0.0}
+ 0%| | 1/18756 [00:22<119:42:17, 22.98s/it]
+[2024-05-15 13:23:44,016] [INFO] [axolotl.callbacks.on_step_end:125] [PID:20482] [RANK:0] GPU memory usage while training: 16.077GB (+25.159GB cache, +1.247GB misc)
+
+ 0%| | 2/18756 [00:39<101:10:04, 19.42s/it]
+
+ 0%| | 3/18756 [00:59<102:10:36, 19.61s/it]
+
+ 0%| | 4/18756 [01:17<98:15:34, 18.86s/it]
+
+ 0%| | 5/18756 [01:35<96:37:06, 18.55s/it]
+
+ 0%| | 6/18756 [01:54<96:51:35, 18.60s/it]
+
+ 0%| | 7/18756 [02:11<93:59:21, 18.05s/it]
+
+ 0%| | 8/18756 [02:30<96:15:29, 18.48s/it]
+
+ 0%| | 9/18756 [02:47<93:30:40, 17.96s/it]
+
+ 0%| | 10/18756 [03:05<93:17:09, 17.91s/it]
+
+ 0%| | 11/18756 [03:22<92:32:37, 17.77s/it]
+
+ 0%| | 12/18756 [03:39<91:27:33, 17.57s/it]
+
+ 0%| | 13/18756 [03:58<92:42:49, 17.81s/it]
+
+ 0%| | 14/18756 [04:15<91:30:49, 17.58s/it]
+
+
+ 0%| | 16/18756 [04:49<90:50:36, 17.45s/it]
+{'loss': 2.303, 'grad_norm': 57.84785236668255, 'learning_rate': 1.8544274455261938e-07, 'epoch': 0.0}
+
+ 0%| | 17/18756 [05:07<90:59:31, 17.48s/it]
+
+ 0%| | 18/18756 [05:25<91:40:43, 17.61s/it]
+
+ 0%| | 19/18756 [05:42<90:13:32, 17.34s/it]
+
+ 0%| | 20/18756 [06:00<92:02:06, 17.68s/it]
+
+ 0%| | 21/18756 [06:17<90:28:47, 17.39s/it]
+
+ 0%| | 22/18756 [06:34<90:29:19, 17.39s/it]
+
+ 0%| | 23/18756 [06:52<90:29:28, 17.39s/it]
+
+ 0%| | 24/18756 [07:09<89:47:11, 17.26s/it]
+
+ 0%| | 25/18756 [07:27<92:10:28, 17.72s/it]
+
+ 0%| | 26/18756 [07:44<90:42:29, 17.43s/it]
+
+ 0%| | 27/18756 [08:02<90:41:08, 17.43s/it]
+
+ 0%| | 28/18756 [08:19<90:02:53, 17.31s/it]
+
+ 0%| | 29/18756 [08:36<90:07:18, 17.32s/it]
+
+ 0%| | 30/18756 [08:55<92:34:11, 17.80s/it]
+
+ 0%| | 31/18756 [09:11<90:46:59, 17.45s/it]
+
+ 0%| | 32/18756 [09:30<91:53:07, 17.67s/it]
+
+ 0%| | 33/18756 [09:46<90:27:31, 17.39s/it]
+
+ 0%| | 34/18756 [10:04<90:37:39, 17.43s/it]
+
+ 0%| | 35/18756 [10:22<91:36:28, 17.62s/it]
+
+ 0%| | 36/18756 [10:39<90:22:05, 17.38s/it]
+
+ 0%|▏ | 37/18756 [10:58<92:42:01, 17.83s/it]
+
+
+ 0%|▏ | 39/18756 [11:32<91:06:12, 17.52s/it]
+
+ 0%|▏ | 40/18756 [11:49<90:17:58, 17.37s/it]
+{'loss': 1.9954, 'grad_norm': 48.64512206162736, 'learning_rate': 4.636068613815485e-07, 'epoch': 0.01}
+
+
+ 0%|▏ | 42/18756 [12:25<92:20:40, 17.76s/it]
+
+ 0%|▏ | 43/18756 [12:42<90:38:38, 17.44s/it]
+
+ 0%|▏ | 44/18756 [13:00<91:20:46, 17.57s/it]
+{'loss': 1.7941, 'grad_norm': 24.18707353871459, 'learning_rate': 5.099675475197033e-07, 'epoch': 0.01}
+
+
+ 0%|▏ | 46/18756 [13:34<90:31:03, 17.42s/it]
+
+ 0%|▏ | 47/18756 [13:52<91:14:09, 17.56s/it]
+
+ 0%|▏ | 48/18756 [14:09<89:48:42, 17.28s/it]
+
+ 0%|▏ | 49/18756 [14:27<91:55:08, 17.69s/it]
+{'loss': 1.9483, 'grad_norm': 28.518231901386347, 'learning_rate': 5.679184051923968e-07, 'epoch': 0.01}
+
+ 0%|▏ | 50/18756 [14:44<90:34:44, 17.43s/it]
+
+ 0%|▏ | 51/18756 [15:01<90:14:16, 17.37s/it]
+
+
+ 0%|▏ | 53/18756 [15:35<89:22:15, 17.20s/it]
+
+ 0%|▏ | 54/18756 [15:54<91:58:04, 17.70s/it]
+
+ 0%|▏ | 55/18756 [16:11<90:45:25, 17.47s/it]
+{'loss': 1.8958, 'grad_norm': 26.755127602524038, 'learning_rate': 6.374594343996291e-07, 'epoch': 0.01}
+
+
+ 0%|▏ | 57/18756 [16:46<90:15:41, 17.38s/it]
+
+ 0%|▏ | 58/18756 [17:03<90:16:04, 17.38s/it]
+
+ 0%|▏ | 59/18756 [17:21<91:34:51, 17.63s/it]
+
+ 0%|▏ | 60/18756 [17:38<90:15:28, 17.38s/it]
+
+ 0%|▏ | 61/18756 [17:56<91:27:52, 17.61s/it]
+
+ 0%|▏ | 62/18756 [18:13<90:14:42, 17.38s/it]
+
+ 0%|▏ | 63/18756 [18:31<90:25:34, 17.41s/it]
+
+ 0%|▏ | 64/18756 [18:48<90:58:34, 17.52s/it]
+{'loss': 1.6818, 'grad_norm': 17.505881057005706, 'learning_rate': 7.417709782104775e-07, 'epoch': 0.01}
+
+ 0%|▏ | 65/18756 [19:06<90:16:35, 17.39s/it]
+
+ 0%|▏ | 66/18756 [19:24<91:40:38, 17.66s/it]
+
+ 0%|▏ | 67/18756 [19:41<90:26:46, 17.42s/it]
+
+ 0%|▏ | 68/18756 [19:58<90:34:28, 17.45s/it]
+
+ 0%|▏ | 69/18756 [20:15<90:17:02, 17.39s/it]
+
+ 0%|▏ | 70/18756 [20:33<90:12:25, 17.38s/it]
+
+ 0%|▏ | 71/18756 [20:51<91:18:24, 17.59s/it]
+
+ 0%|▏ | 72/18756 [21:08<90:15:03, 17.39s/it]
+
+ 0%|▏ | 73/18756 [21:26<91:19:30, 17.60s/it]
+
+ 0%|▎ | 74/18756 [21:43<90:08:15, 17.37s/it]
+
+ 0%|▎ | 75/18756 [22:00<90:34:09, 17.45s/it]
+
+ 0%|▎ | 76/18756 [22:18<91:08:33, 17.56s/it]
+
+ 0%|▎ | 77/18756 [22:35<90:20:20, 17.41s/it]
+
+ 0%|▎ | 78/18756 [22:53<91:04:05, 17.55s/it]
+
+ 0%|▎ | 79/18756 [23:10<90:13:13, 17.39s/it]
+
+ 0%|▎ | 80/18756 [23:28<90:45:08, 17.49s/it]
+
+ 0%|▎ | 81/18756 [23:45<90:30:33, 17.45s/it]
+
+ 0%|▎ | 82/18756 [24:03<90:30:32, 17.45s/it]
+
+
+ 0%|▎ | 84/18756 [24:37<90:10:30, 17.39s/it]
+
+ 0%|▎ | 85/18756 [24:56<91:18:59, 17.61s/it]
+
+ 0%|▎ | 86/18756 [25:12<90:11:30, 17.39s/it]
+
+ 0%|▎ | 87/18756 [25:30<90:43:45, 17.50s/it]
+
+ 0%|▎ | 88/18756 [25:47<90:19:59, 17.42s/it]
+
+ 0%|▎ | 89/18756 [26:05<90:12:34, 17.40s/it]
+{'loss': 1.7514, 'grad_norm': 11.653287209332959, 'learning_rate': 1.0315252665739455e-06, 'epoch': 0.01}
+
+ 0%|▎ | 90/18756 [26:22<90:41:49, 17.49s/it]
+
+ 0%|▎ | 91/18756 [26:39<89:28:51, 17.26s/it]
+
+ 0%|▎ | 92/18756 [26:57<90:18:23, 17.42s/it]
+
+ 0%|▎ | 93/18756 [27:14<90:01:11, 17.36s/it]
+
+ 1%|▎ | 94/18756 [27:32<90:23:01, 17.44s/it]
+
+ 1%|▎ | 95/18756 [27:49<90:08:45, 17.39s/it]
+
+ 1%|▎ | 96/18756 [28:06<89:25:02, 17.25s/it]
+
+ 1%|▎ | 97/18756 [28:24<90:19:56, 17.43s/it]
+
+ 1%|▎ | 98/18756 [28:41<89:35:24, 17.29s/it]
+
+
+ 1%|▎ | 100/18756 [36:55<803:26:09, 155.04s/it]
+{'loss': 1.4187, 'grad_norm': 9.269101601926474, 'learning_rate': 1.1590171534538713e-06, 'epoch': 0.02}
+
+ 1%|▎ | 101/18756 [37:12<588:30:07, 113.57s/it]
+
+ 1%|▎ | 102/18756 [37:29<438:44:59, 84.67s/it]
+
+ 1%|▎ | 103/18756 [37:47<335:02:42, 64.66s/it]
+
+ 1%|▎ | 104/18756 [38:04<261:13:43, 50.42s/it]
+
+ 1%|▎ | 105/18756 [38:22<211:12:29, 40.77s/it]
+
+ 1%|▎ | 106/18756 [38:39<173:30:55, 33.49s/it]
+
+
+ 1%|▎ | 108/18756 [39:13<130:22:17, 25.17s/it]
+{'loss': 1.5966, 'grad_norm': 9.10386991271818, 'learning_rate': 1.2517385257301808e-06, 'epoch': 0.02}
+
+
+ 1%|▎ | 110/18756 [39:49<111:55:05, 21.61s/it]
+{'loss': 1.3813, 'grad_norm': 5.883278391350094, 'learning_rate': 1.2749188687992583e-06, 'epoch': 0.02}
+
+ 1%|▎ | 111/18756 [40:06<103:54:07, 20.06s/it]
+
+ 1%|▎ | 112/18756 [40:23<100:11:40, 19.35s/it]
+
+ 1%|▍ | 113/18756 [40:40<96:13:25, 18.58s/it]
+
+ 1%|▍ | 114/18756 [40:58<94:20:51, 18.22s/it]
+
+ 1%|▍ | 115/18756 [41:16<94:27:06, 18.24s/it]
+
+ 1%|▍ | 116/18756 [41:33<92:03:27, 17.78s/it]
+
+ 1%|▍ | 117/18756 [41:51<92:59:36, 17.96s/it]
+
+ 1%|▍ | 118/18756 [42:08<91:00:28, 17.58s/it]
+
+ 1%|▍ | 119/18756 [42:25<90:18:27, 17.44s/it]
+
+ 1%|▍ | 120/18756 [42:42<90:02:07, 17.39s/it]
+
+ 1%|▍ | 121/18756 [42:59<88:59:03, 17.19s/it]
+
+ 1%|▍ | 122/18756 [43:18<91:52:30, 17.75s/it]
+
+ 1%|▍ | 123/18756 [43:35<90:35:04, 17.50s/it]
+
+ 1%|▍ | 124/18756 [43:52<90:32:29, 17.49s/it]
+
+ 1%|▍ | 125/18756 [44:09<89:22:15, 17.27s/it]
+
+ 1%|▍ | 126/18756 [44:26<88:21:59, 17.08s/it]
+
+ 1%|▍ | 127/18756 [44:45<91:45:57, 17.73s/it]
+
+ 1%|▍ | 128/18756 [45:02<90:20:27, 17.46s/it]
+
+ 1%|▍ | 129/18756 [45:20<91:46:08, 17.74s/it]
+
+ 1%|▍ | 130/18756 [45:37<90:12:57, 17.44s/it]
+
+ 1%|▍ | 131/18756 [45:53<88:52:33, 17.18s/it]
+
+ 1%|▍ | 132/18756 [46:11<90:13:07, 17.44s/it]
+
+ 1%|▍ | 133/18756 [46:28<89:00:22, 17.21s/it]
+
+ 1%|▍ | 134/18756 [46:47<91:44:21, 17.74s/it]
+
+ 1%|▍ | 135/18756 [47:04<90:37:01, 17.52s/it]
+
+ 1%|▍ | 136/18756 [47:21<89:53:35, 17.38s/it]
+
+
+ 1%|▍ | 138/18756 [47:55<89:10:43, 17.24s/it]
+{'loss': 1.5655, 'grad_norm': 8.378564787622677, 'learning_rate': 1.5994436717663423e-06, 'epoch': 0.02}
+
+ 1%|▍ | 139/18756 [48:14<91:14:08, 17.64s/it]
+
+ 1%|▍ | 140/18756 [48:31<90:10:48, 17.44s/it]
+
+ 1%|▍ | 141/18756 [48:49<91:08:21, 17.63s/it]
+
+ 1%|▍ | 142/18756 [49:06<90:46:15, 17.56s/it]
+
+ 1%|▍ | 143/18756 [49:23<89:17:10, 17.27s/it]
+
+ 1%|▍ | 144/18756 [49:41<89:57:58, 17.40s/it]
+
+ 1%|▍ | 145/18756 [49:58<89:04:35, 17.23s/it]
+
+ 1%|▍ | 146/18756 [50:16<90:54:43, 17.59s/it]
+
+ 1%|▍ | 147/18756 [50:34<90:48:37, 17.57s/it]
+
+ 1%|▍ | 148/18756 [50:51<90:09:52, 17.44s/it]
+
+ 1%|▌ | 149/18756 [51:08<90:13:41, 17.46s/it]
+
+ 1%|▌ | 150/18756 [51:25<89:04:16, 17.23s/it]
+
+ 1%|▌ | 151/18756 [51:43<90:13:49, 17.46s/it]
+
+ 1%|▌ | 152/18756 [52:00<90:26:48, 17.50s/it]
+
+ 1%|▌ | 153/18756 [52:18<90:59:06, 17.61s/it]
+
+ 1%|▌ | 154/18756 [52:36<90:51:21, 17.58s/it]
+
+ 1%|▌ | 155/18756 [52:52<89:22:19, 17.30s/it]
+
+ 1%|▌ | 156/18756 [53:09<88:55:55, 17.21s/it]
+
+ 1%|▌ | 157/18756 [53:27<89:34:36, 17.34s/it]
+
+ 1%|▌ | 158/18756 [53:46<91:14:37, 17.66s/it]
+
+ 1%|▌ | 159/18756 [54:03<90:36:07, 17.54s/it]
+
+ 1%|▌ | 160/18756 [54:20<90:37:05, 17.54s/it]
+
+ 1%|▌ | 161/18756 [54:37<89:58:50, 17.42s/it]
+
+ 1%|▌ | 162/18756 [54:55<90:06:35, 17.45s/it]
+
+ 1%|▌ | 163/18756 [55:13<90:16:44, 17.48s/it]
+
+ 1%|▌ | 164/18756 [55:30<90:08:00, 17.45s/it]
+
+ 1%|▌ | 165/18756 [55:48<91:29:13, 17.72s/it]
+
+ 1%|▌ | 166/18756 [56:05<90:27:17, 17.52s/it]
+
+ 1%|▌ | 167/18756 [56:23<90:03:27, 17.44s/it]
+
+ 1%|▌ | 168/18756 [56:39<89:06:57, 17.26s/it]
+
+ 1%|▌ | 169/18756 [56:57<89:06:16, 17.26s/it]
+
+ 1%|▌ | 170/18756 [57:15<90:58:22, 17.62s/it]
+
+ 1%|▌ | 171/18756 [57:32<90:13:16, 17.48s/it]
+
+ 1%|▌ | 172/18756 [57:50<91:13:06, 17.67s/it]
+
+ 1%|▌ | 173/18756 [58:08<90:26:38, 17.52s/it]
+
+ 1%|▌ | 174/18756 [58:24<89:27:40, 17.33s/it]
+
+ 1%|▌ | 175/18756 [58:42<89:58:36, 17.43s/it]
+
+ 1%|▌ | 176/18756 [58:59<89:45:15, 17.39s/it]
+
+
+ 1%|▌ | 178/18756 [59:35<90:39:42, 17.57s/it]
+
+ 1%|▌ | 179/18756 [59:52<90:15:29, 17.49s/it]
+
+ 1%|▌ | 180/18756 [1:00:10<90:09:21, 17.47s/it]
+
+ 1%|▌ | 181/18756 [1:00:27<89:13:28, 17.29s/it]
+
+ 1%|▌ | 182/18756 [1:00:45<90:43:21, 17.58s/it]
+{'loss': 1.2176, 'grad_norm': 6.814093447175195, 'learning_rate': 2.109411219286046e-06, 'epoch': 0.03}
+
+ 1%|▌ | 183/18756 [1:01:02<90:09:35, 17.48s/it]
+
+ 1%|▌ | 184/18756 [1:01:20<90:37:28, 17.57s/it]
+
+ 1%|▌ | 185/18756 [1:01:38<90:39:21, 17.57s/it]
+
+
+ 1%|▌ | 187/18756 [1:02:12<89:33:18, 17.36s/it]
+
+ 1%|▌ | 188/18756 [1:02:29<89:14:19, 17.30s/it]
+
+ 1%|▌ | 189/18756 [1:02:47<90:41:46, 17.59s/it]
+
+ 1%|▌ | 190/18756 [1:03:05<90:35:47, 17.57s/it]
+
+ 1%|▌ | 191/18756 [1:03:22<90:32:37, 17.56s/it]
+
+ 1%|▌ | 192/18756 [1:03:40<90:44:03, 17.60s/it]
+
+ 1%|▋ | 193/18756 [1:03:57<89:29:29, 17.36s/it]
+
+ 1%|▋ | 194/18756 [1:04:14<89:50:21, 17.42s/it]
+
+ 1%|▋ | 195/18756 [1:04:32<89:50:01, 17.42s/it]
+{'loss': 1.3032, 'grad_norm': 7.316903117304981, 'learning_rate': 2.260083449235049e-06, 'epoch': 0.03}
+
+
+ 1%|▋ | 197/18756 [1:05:08<91:21:36, 17.72s/it]
+
+ 1%|▋ | 198/18756 [1:05:25<90:32:01, 17.56s/it]
+
+ 1%|▋ | 199/18756 [1:05:42<89:58:20, 17.45s/it]
+{'loss': 1.3222, 'grad_norm': 6.192321748166232, 'learning_rate': 2.3064441353732037e-06, 'epoch': 0.03}
+
+ 1%|▋ | 200/18756 [1:13:36<795:16:30, 154.29s/it]
+
+ 1%|▋ | 201/18756 [1:13:53<582:52:30, 113.09s/it]
+
+ 1%|▋ | 202/18756 [1:14:10<434:51:08, 84.37s/it]
+
+ 1%|▋ | 203/18756 [1:14:28<332:09:05, 64.45s/it]
+
+
+ 1%|▋ | 205/18756 [1:15:04<210:56:48, 40.94s/it]
+{'loss': 1.3144, 'grad_norm': 5.79122048541526, 'learning_rate': 2.375985164580436e-06, 'epoch': 0.03}
+
+ 1%|▋ | 206/18756 [1:15:22<175:07:49, 33.99s/it]
+
+
+ 1%|▋ | 208/18756 [1:15:59<133:09:54, 25.85s/it]
+{'loss': 1.5155, 'grad_norm': 7.782613489271369, 'learning_rate': 2.410755679184052e-06, 'epoch': 0.03}
+
+ 1%|▋ | 209/18756 [1:16:15<118:51:51, 23.07s/it]
+
+
+ 1%|▋ | 211/18756 [1:16:50<103:54:58, 20.17s/it]
+
+ 1%|▋ | 212/18756 [1:17:08<101:26:14, 19.69s/it]
+
+ 1%|▋ | 213/18756 [1:17:26<98:02:42, 19.03s/it]
+
+ 1%|▋ | 214/18756 [1:17:43<95:32:15, 18.55s/it]
+{'loss': 1.3787, 'grad_norm': 7.673422973842772, 'learning_rate': 2.480296708391284e-06, 'epoch': 0.03}
+
+
+ 1%|▋ | 216/18756 [1:18:18<92:11:05, 17.90s/it]
+
+ 1%|▋ | 217/18756 [1:18:35<91:10:22, 17.70s/it]
+
+ 1%|▋ | 218/18756 [1:18:53<91:08:39, 17.70s/it]
+
+ 1%|▋ | 219/18756 [1:19:12<92:55:39, 18.05s/it]
+
+ 1%|▋ | 220/18756 [1:19:29<91:13:14, 17.72s/it]
+
+ 1%|▋ | 221/18756 [1:19:46<90:46:52, 17.63s/it]
+
+ 1%|▋ | 222/18756 [1:20:02<89:04:57, 17.30s/it]
+
+ 1%|▋ | 223/18756 [1:20:20<89:36:19, 17.41s/it]
+
+ 1%|▋ | 224/18756 [1:20:38<90:22:42, 17.56s/it]
+
+ 1%|▋ | 225/18756 [1:20:56<91:28:54, 17.77s/it]
+
+ 1%|▋ | 226/18756 [1:21:15<92:41:44, 18.01s/it]
+
+ 1%|▋ | 227/18756 [1:21:32<91:15:46, 17.73s/it]
+
+ 1%|▋ | 228/18756 [1:21:50<90:58:21, 17.68s/it]
+
+ 1%|▋ | 229/18756 [1:22:06<89:20:46, 17.36s/it]
+
+ 1%|▋ | 230/18756 [1:22:24<90:07:57, 17.51s/it]
+
+ 1%|▊ | 231/18756 [1:22:43<91:38:03, 17.81s/it]
+
+ 1%|▊ | 232/18756 [1:23:00<91:54:17, 17.86s/it]
+
+ 1%|▊ | 233/18756 [1:23:19<92:28:32, 17.97s/it]
+
+ 1%|▊ | 234/18756 [1:23:35<90:13:03, 17.54s/it]
+
+ 1%|▊ | 235/18756 [1:23:53<89:50:30, 17.46s/it]
+
+ 1%|▊ | 236/18756 [1:24:10<89:15:46, 17.35s/it]
+
+ 1%|▊ | 237/18756 [1:24:27<89:05:34, 17.32s/it]
+
+ 1%|▊ | 238/18756 [1:24:46<91:43:47, 17.83s/it]
+
+ 1%|▊ | 239/18756 [1:25:03<90:18:01, 17.56s/it]
+
+ 1%|▊ | 240/18756 [1:25:21<91:26:38, 17.78s/it]
+
+ 1%|▊ | 241/18756 [1:25:38<89:57:43, 17.49s/it]
+
+ 1%|▊ | 242/18756 [1:25:55<89:33:21, 17.41s/it]
+
+ 1%|▊ | 243/18756 [1:26:13<90:46:44, 17.65s/it]
+
+ 1%|▊ | 244/18756 [1:26:31<90:18:18, 17.56s/it]
+
+ 1%|▊ | 245/18756 [1:26:49<91:40:17, 17.83s/it]
+
+ 1%|▊ | 246/18756 [1:27:06<90:22:57, 17.58s/it]
+
+ 1%|▊ | 247/18756 [1:27:24<90:32:58, 17.61s/it]
+
+ 1%|▊ | 248/18756 [1:27:41<90:00:12, 17.51s/it]
+
+ 1%|▊ | 249/18756 [1:27:58<89:09:08, 17.34s/it]
+
+ 1%|▊ | 250/18756 [1:28:16<90:33:38, 17.62s/it]
+
+ 1%|▊ | 251/18756 [1:28:34<90:04:04, 17.52s/it]
+
+ 1%|▊ | 252/18756 [1:28:52<91:08:37, 17.73s/it]
+
+ 1%|▊ | 253/18756 [1:29:09<90:37:18, 17.63s/it]
+
+ 1%|▊ | 254/18756 [1:29:26<89:49:18, 17.48s/it]
+
+ 1%|▊ | 255/18756 [1:29:44<89:19:10, 17.38s/it]
+
+ 1%|▊ | 256/18756 [1:30:01<89:27:23, 17.41s/it]
+
+ 1%|▊ | 257/18756 [1:30:18<89:28:49, 17.41s/it]
+
+ 1%|▊ | 258/18756 [1:30:37<90:59:55, 17.71s/it]
+
+ 1%|▊ | 259/18756 [1:30:54<90:34:24, 17.63s/it]
+
+ 1%|▊ | 260/18756 [1:31:11<89:47:38, 17.48s/it]
+
+ 1%|▊ | 261/18756 [1:31:29<89:52:02, 17.49s/it]
+
+ 1%|▊ | 262/18756 [1:31:46<89:14:49, 17.37s/it]
+
+ 1%|▊ | 263/18756 [1:32:04<90:56:20, 17.70s/it]
+
+ 1%|▊ | 264/18756 [1:32:22<90:42:25, 17.66s/it]
+
+ 1%|▊ | 265/18756 [1:32:39<89:50:10, 17.49s/it]
+
+ 1%|▊ | 266/18756 [1:32:57<90:24:38, 17.60s/it]
+
+ 1%|▊ | 267/18756 [1:33:14<88:54:05, 17.31s/it]
+
+ 1%|▊ | 268/18756 [1:33:31<89:31:45, 17.43s/it]
+
+ 1%|▊ | 269/18756 [1:33:49<89:13:17, 17.37s/it]
+
+ 1%|▉ | 270/18756 [1:34:07<90:48:10, 17.68s/it]
+
+ 1%|▉ | 271/18756 [1:34:24<90:30:21, 17.63s/it]
+
+ 1%|▉ | 272/18756 [1:34:41<89:02:34, 17.34s/it]
+
+ 1%|▉ | 273/18756 [1:35:00<90:36:05, 17.65s/it]
+{'loss': 1.3829, 'grad_norm': 5.782712679199597, 'learning_rate': 3.164116828929068e-06, 'epoch': 0.04}
+
+
+ 1%|▉ | 275/18756 [1:35:34<89:31:00, 17.44s/it]
+
+ 1%|▉ | 276/18756 [1:35:52<90:24:32, 17.61s/it]
+
+ 1%|▉ | 277/18756 [1:36:09<89:17:11, 17.39s/it]
+
+ 1%|▉ | 278/18756 [1:36:27<91:01:09, 17.73s/it]
+
+ 1%|▉ | 279/18756 [1:36:44<89:38:16, 17.46s/it]
+
+ 1%|▉ | 280/18756 [1:37:02<90:13:22, 17.58s/it]
+
+ 1%|▉ | 281/18756 [1:37:20<91:26:07, 17.82s/it]
+
+ 2%|▉ | 282/18756 [1:37:40<94:42:42, 18.46s/it]
+
+ 2%|▉ | 283/18756 [1:38:00<96:49:14, 18.87s/it]
+{'loss': 1.4737, 'grad_norm': 7.42003138080552, 'learning_rate': 3.2800185442744555e-06, 'epoch': 0.05}
+
+
+ 2%|▉ | 285/18756 [1:38:39<98:02:16, 19.11s/it]
+
+ 2%|▉ | 286/18756 [1:38:56<95:35:03, 18.63s/it]
+
+ 2%|▉ | 287/18756 [1:39:14<95:15:16, 18.57s/it]
+
+ 2%|▉ | 288/18756 [1:39:32<92:56:12, 18.12s/it]
+{'loss': 1.3076, 'grad_norm': 8.002206021372862, 'learning_rate': 3.337969401947149e-06, 'epoch': 0.05}
+
+ 2%|▉ | 289/18756 [1:39:51<95:01:23, 18.52s/it]
+
+
+ 2%|▉ | 291/18756 [1:40:27<93:30:13, 18.23s/it]
+{'loss': 1.5006, 'grad_norm': 6.044182860722979, 'learning_rate': 3.372739916550765e-06, 'epoch': 0.05}
+
+ 2%|▉ | 292/18756 [1:40:44<92:25:08, 18.02s/it]
+
+ 2%|▉ | 293/18756 [1:41:03<93:42:23, 18.27s/it]
+
+ 2%|▉ | 294/18756 [1:41:21<92:23:04, 18.01s/it]
+
+
+ 2%|▉ | 296/18756 [1:41:56<91:32:34, 17.85s/it]
+
+ 2%|▉ | 297/18756 [1:42:13<90:55:41, 17.73s/it]
+
+ 2%|▉ | 298/18756 [1:42:31<91:00:56, 17.75s/it]
+
+ 2%|▉ | 299/18756 [1:42:48<90:09:37, 17.59s/it]
+{'loss': 1.6343, 'grad_norm': 6.858650881838593, 'learning_rate': 3.4654612888270746e-06, 'epoch': 0.05}
+
+ 2%|▉ | 300/18756 [1:50:45<796:14:04, 155.31s/it]
+
+ 2%|▉ | 301/18756 [1:51:03<585:03:46, 114.13s/it]
+
+ 2%|▉ | 302/18756 [1:51:20<436:21:51, 85.13s/it]
+
+ 2%|▉ | 303/18756 [1:51:38<332:19:11, 64.83s/it]
+
+ 2%|▉ | 304/18756 [1:51:55<258:23:07, 50.41s/it]
+
+ 2%|▉ | 305/18756 [1:52:12<208:14:57, 40.63s/it]
+
+ 2%|▉ | 306/18756 [1:52:30<172:15:48, 33.61s/it]
+
+ 2%|▉ | 307/18756 [1:52:48<148:15:25, 28.93s/it]
+
+ 2%|▉ | 308/18756 [1:53:04<129:32:59, 25.28s/it]
+
+ 2%|▉ | 309/18756 [1:53:21<117:00:33, 22.83s/it]
+
+ 2%|▉ | 310/18756 [1:53:39<108:27:25, 21.17s/it]
+
+
+ 2%|█ | 312/18756 [1:54:14<99:55:24, 19.50s/it]
+{'loss': 1.5729, 'grad_norm': 6.648585531915439, 'learning_rate': 3.6161335187760785e-06, 'epoch': 0.05}
+
+ 2%|█ | 313/18756 [1:54:31<95:39:01, 18.67s/it]
+
+ 2%|█ | 314/18756 [1:54:49<94:03:29, 18.36s/it]
+
+
+ 2%|█ | 316/18756 [1:55:23<91:10:04, 17.80s/it]
+{'loss': 1.5973, 'grad_norm': 6.90510045491595, 'learning_rate': 3.662494204914233e-06, 'epoch': 0.05}
+
+ 2%|█ | 317/18756 [1:55:41<90:44:14, 17.72s/it]
+
+ 2%|█ | 318/18756 [1:55:58<89:39:44, 17.51s/it]
+
+ 2%|█ | 319/18756 [1:56:16<90:50:47, 17.74s/it]
+
+ 2%|█ | 320/18756 [1:56:32<88:31:04, 17.28s/it]
+
+ 2%|█ | 321/18756 [1:56:50<89:12:24, 17.42s/it]
+
+ 2%|█ | 322/18756 [1:57:07<88:38:15, 17.31s/it]
+
+ 2%|█ | 323/18756 [1:57:25<89:00:39, 17.38s/it]
+
+ 2%|█ | 324/18756 [1:57:43<90:01:55, 17.58s/it]
+
+ 2%|█ | 325/18756 [1:57:59<88:17:48, 17.25s/it]
+
+ 2%|█ | 326/18756 [1:58:18<90:26:29, 17.67s/it]
+
+ 2%|█ | 327/18756 [1:58:34<88:26:09, 17.28s/it]
+
+ 2%|█ | 328/18756 [1:58:51<88:32:09, 17.30s/it]
+
+ 2%|█ | 329/18756 [1:59:09<88:47:56, 17.35s/it]
+
+ 2%|█ | 330/18756 [1:59:26<87:47:31, 17.15s/it]
+
+ 2%|█ | 331/18756 [1:59:45<90:40:52, 17.72s/it]
+
+ 2%|█ | 332/18756 [2:00:01<88:48:09, 17.35s/it]
+
+ 2%|█ | 333/18756 [2:00:19<89:32:10, 17.50s/it]
+
+ 2%|█ | 334/18756 [2:00:35<88:07:45, 17.22s/it]
+{'loss': 1.2344, 'grad_norm': 9.793095347830493, 'learning_rate': 3.882707464070468e-06, 'epoch': 0.05}
+{'loss': 1.5328, 'grad_norm': 5.746766217660957, 'learning_rate': 3.894297635605007e-06, 'epoch': 0.05}
+
+ 2%|█ | 337/18756 [2:01:27<88:17:17, 17.26s/it]
+
+ 2%|█ | 338/18756 [2:01:46<90:14:19, 17.64s/it]
+
+ 2%|█ | 339/18756 [2:02:03<88:49:32, 17.36s/it]
+
+ 2%|█ | 340/18756 [2:02:20<88:33:22, 17.31s/it]
+
+ 2%|█ | 341/18756 [2:02:37<89:04:51, 17.41s/it]
+
+ 2%|█ | 342/18756 [2:02:54<87:51:12, 17.18s/it]
+
+ 2%|█ | 343/18756 [2:03:13<90:01:51, 17.60s/it]
+
+ 2%|█ | 344/18756 [2:03:29<88:37:52, 17.33s/it]
+
+
+ 2%|█▏ | 346/18756 [2:04:05<89:14:29, 17.45s/it]
+
+ 2%|█▏ | 347/18756 [2:04:21<88:14:17, 17.26s/it]
+
+ 2%|█▏ | 348/18756 [2:04:39<89:18:20, 17.47s/it]
+
+ 2%|█▏ | 349/18756 [2:04:56<88:07:44, 17.24s/it]
+
+ 2%|█▏ | 350/18756 [2:05:14<89:48:59, 17.57s/it]
+
+ 2%|█▏ | 351/18756 [2:05:32<89:05:51, 17.43s/it]
+
+ 2%|█▏ | 352/18756 [2:05:49<89:09:52, 17.44s/it]
+
+ 2%|█▏ | 353/18756 [2:06:07<89:27:22, 17.50s/it]
+
+ 2%|█▏ | 354/18756 [2:06:23<88:04:40, 17.23s/it]
+
+ 2%|█▏ | 355/18756 [2:06:41<88:31:09, 17.32s/it]
+
+ 2%|█▏ | 356/18756 [2:06:58<88:04:20, 17.23s/it]
+
+ 2%|█▏ | 357/18756 [2:07:16<89:43:53, 17.56s/it]
+{'loss': 1.2307, 'grad_norm': 6.236343111378016, 'learning_rate': 4.13769123783032e-06, 'epoch': 0.06}
+
+
+ 2%|█▏ | 359/18756 [2:07:51<89:34:53, 17.53s/it]
+
+ 2%|█▏ | 360/18756 [2:08:08<89:15:23, 17.47s/it]
+
+ 2%|█▏ | 361/18756 [2:08:25<87:55:37, 17.21s/it]
+
+ 2%|█▏ | 362/18756 [2:08:43<88:59:22, 17.42s/it]
+
+ 2%|█▏ | 363/18756 [2:09:00<88:40:56, 17.36s/it]
+
+ 2%|█▏ | 364/18756 [2:09:19<90:14:20, 17.66s/it]
+{'loss': 1.2473, 'grad_norm': 6.953897703290024, 'learning_rate': 4.218822438572092e-06, 'epoch': 0.06}
+
+
+ 2%|█▏ | 366/18756 [2:09:53<89:02:38, 17.43s/it]
+{'loss': 1.2149, 'grad_norm': 7.336121838155878, 'learning_rate': 4.242002781641168e-06, 'epoch': 0.06}
+
+
+ 2%|█▏ | 368/18756 [2:10:27<87:55:36, 17.21s/it]
+
+ 2%|█▏ | 369/18756 [2:10:45<89:33:59, 17.54s/it]
+
+ 2%|█▏ | 370/18756 [2:11:03<89:18:34, 17.49s/it]
+
+ 2%|█▏ | 371/18756 [2:11:20<89:51:38, 17.60s/it]
+
+ 2%|█▏ | 372/18756 [2:11:37<88:50:44, 17.40s/it]
+
+ 2%|█▏ | 373/18756 [2:11:55<88:25:44, 17.32s/it]
+
+ 2%|█▏ | 374/18756 [2:12:12<88:07:47, 17.26s/it]
+
+ 2%|█▏ | 375/18756 [2:12:29<88:10:23, 17.27s/it]
+
+ 2%|█▏ | 376/18756 [2:12:47<89:40:37, 17.56s/it]
+{'loss': 1.1685, 'grad_norm': 6.319111902153149, 'learning_rate': 4.357904496986556e-06, 'epoch': 0.06}
+
+
+ 2%|█▏ | 378/18756 [2:13:22<89:25:41, 17.52s/it]
+
+ 2%|█▏ | 379/18756 [2:13:39<88:22:25, 17.31s/it]
+
+ 2%|█▏ | 380/18756 [2:13:56<88:00:18, 17.24s/it]
+
+ 2%|█▏ | 381/18756 [2:14:14<88:43:15, 17.38s/it]
+
+ 2%|█▏ | 382/18756 [2:14:31<88:22:04, 17.31s/it]
+
+ 2%|█▏ | 383/18756 [2:14:49<90:03:25, 17.65s/it]
+
+ 2%|█▏ | 384/18756 [2:15:06<88:55:19, 17.42s/it]
+
+ 2%|█▎ | 385/18756 [2:15:24<89:10:03, 17.47s/it]
+{'loss': 1.2124, 'grad_norm': 7.179233474405836, 'learning_rate': 4.4622160407974036e-06, 'epoch': 0.06}
+
+
+ 2%|█▎ | 387/18756 [2:15:57<87:44:16, 17.20s/it]
+
+ 2%|█▎ | 388/18756 [2:16:16<89:26:42, 17.53s/it]
+
+ 2%|█▎ | 389/18756 [2:16:33<88:52:51, 17.42s/it]
+
+ 2%|█▎ | 390/18756 [2:16:51<90:14:27, 17.69s/it]
+
+ 2%|█▎ | 391/18756 [2:17:08<89:06:07, 17.47s/it]
+
+ 2%|█▎ | 392/18756 [2:17:25<88:03:43, 17.26s/it]
+{'loss': 1.3246, 'grad_norm': 7.136761243972423, 'learning_rate': 4.5433472415391745e-06, 'epoch': 0.06}
+
+
+ 2%|█▎ | 394/18756 [2:18:00<88:13:05, 17.30s/it]
+
+ 2%|█▎ | 395/18756 [2:18:18<89:44:11, 17.59s/it]
+
+ 2%|█▎ | 396/18756 [2:18:35<89:25:37, 17.53s/it]
+
+ 2%|█▎ | 397/18756 [2:18:53<89:44:21, 17.60s/it]
+
+ 2%|█▎ | 398/18756 [2:19:10<88:55:58, 17.44s/it]
+
+ 2%|█▎ | 399/18756 [2:19:27<87:59:22, 17.26s/it]
+
+ 2%|█▎ | 400/18756 [2:27:19<782:52:04, 153.54s/it]
+
+ 2%|█▎ | 401/18756 [2:27:37<575:53:36, 112.95s/it]
+{'loss': 1.4312, 'grad_norm': 7.237488573029084, 'learning_rate': 4.647658785350023e-06, 'epoch': 0.06}
+
+
+ 2%|█▎ | 403/18756 [2:28:11<326:13:16, 63.99s/it]
+
+ 2%|█▎ | 404/18756 [2:28:28<255:38:12, 50.15s/it]
+{'loss': 1.5737, 'grad_norm': 6.097423522144287, 'learning_rate': 4.68242929995364e-06, 'epoch': 0.06}
+
+
+ 2%|█▎ | 406/18756 [2:29:03<170:56:13, 33.54s/it]
+
+ 2%|█▎ | 407/18756 [2:29:20<144:22:39, 28.33s/it]
+
+ 2%|█▎ | 408/18756 [2:29:38<128:44:21, 25.26s/it]
+
+ 2%|█▎ | 409/18756 [2:29:55<116:22:11, 22.83s/it]
+
+ 2%|█▎ | 410/18756 [2:30:12<107:20:37, 21.06s/it]
+
+ 2%|█▎ | 411/18756 [2:30:30<103:13:08, 20.26s/it]
+
+ 2%|█▎ | 412/18756 [2:30:47<98:03:00, 19.24s/it]
+
+ 2%|█▎ | 413/18756 [2:31:05<96:02:02, 18.85s/it]
+
+ 2%|█▎ | 414/18756 [2:31:21<92:30:10, 18.16s/it]
+
+ 2%|█▎ | 415/18756 [2:31:39<91:52:15, 18.03s/it]
+
+ 2%|█▎ | 416/18756 [2:31:57<91:04:49, 17.88s/it]
+
+ 2%|█▎ | 417/18756 [2:32:14<90:14:09, 17.71s/it]
+ 2%|█▎ | 417/18756 [2:32:14<90:14:09, 17.71s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/nn/modules/module.py:1879: UserWarning: Positional args are being deprecated, use kwargs instead. Refer to https://pytorch.org/docs/master/generated/torch.nn.Module.html#torch.nn.Module.state_dict for details.
+ warnings.warn(
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+ 2%|█▎ | 418/18756 [2:33:36<188:59:22, 37.10s/it]
+
+ 2%|█▎ | 419/18756 [2:33:54<159:19:30, 31.28s/it]
+
+ 2%|█▎ | 420/18756 [2:34:11<137:23:08, 26.97s/it]
+
+ 2%|█▎ | 421/18756 [2:34:29<124:03:27, 24.36s/it]
+
+ 2%|█▎ | 422/18756 [2:34:46<112:15:28, 22.04s/it]
+
+ 2%|█▎ | 423/18756 [2:35:04<105:45:36, 20.77s/it]
+
+ 2%|█▎ | 424/18756 [2:35:22<102:35:57, 20.15s/it]
+
+ 2%|█▍ | 425/18756 [2:35:40<98:01:23, 19.25s/it]
+
+ 2%|█▍ | 426/18756 [2:35:57<95:49:44, 18.82s/it]
+
+ 2%|█▍ | 427/18756 [2:36:14<92:19:30, 18.13s/it]
+
+ 2%|█▍ | 428/18756 [2:36:32<91:57:53, 18.06s/it]
+
+ 2%|█▍ | 429/18756 [2:36:49<90:29:35, 17.78s/it]
+
+ 2%|█▍ | 430/18756 [2:37:06<89:58:46, 17.68s/it]
+
+ 2%|█▍ | 431/18756 [2:37:24<90:12:29, 17.72s/it]
+
+ 2%|█▍ | 432/18756 [2:37:41<89:05:55, 17.50s/it]
+
+ 2%|█▍ | 433/18756 [2:37:59<89:07:49, 17.51s/it]
+
+ 2%|█▍ | 434/18756 [2:38:16<87:58:39, 17.29s/it]
+
+ 2%|█▍ | 435/18756 [2:38:33<88:39:26, 17.42s/it]
+
+ 2%|█▍ | 436/18756 [2:38:51<89:08:50, 17.52s/it]
+
+ 2%|█▍ | 437/18756 [2:39:08<88:47:29, 17.45s/it]
+
+ 2%|█▍ | 438/18756 [2:39:26<88:41:15, 17.43s/it]
+
+ 2%|█▍ | 439/18756 [2:39:42<87:40:10, 17.23s/it]
+
+ 2%|█▍ | 440/18756 [2:40:01<89:01:44, 17.50s/it]
+
+ 2%|█▍ | 441/18756 [2:40:17<87:53:28, 17.28s/it]
+
+ 2%|█▍ | 442/18756 [2:40:34<87:41:27, 17.24s/it]
+
+ 2%|█▍ | 443/18756 [2:40:52<88:50:01, 17.46s/it]
+
+ 2%|█▍ | 444/18756 [2:41:09<87:30:14, 17.20s/it]
+
+ 2%|█▍ | 445/18756 [2:41:27<88:52:13, 17.47s/it]
+{'loss': 1.3207, 'grad_norm': 4.997433194300264, 'learning_rate': 5.157626332869727e-06, 'epoch': 0.07}
+
+ 2%|█▍ | 446/18756 [2:41:44<87:21:39, 17.18s/it]
+
+
+ 2%|█▍ | 448/18756 [2:42:19<88:00:56, 17.31s/it]
+{'loss': 1.4542, 'grad_norm': 6.611498047576318, 'learning_rate': 5.192396847473343e-06, 'epoch': 0.07}
+
+ 2%|█▍ | 449/18756 [2:42:36<88:29:59, 17.40s/it]
+
+ 2%|█▍ | 450/18756 [2:42:54<88:45:06, 17.45s/it]
+
+ 2%|█▍ | 451/18756 [2:43:11<88:25:30, 17.39s/it]
+
+ 2%|█▍ | 452/18756 [2:43:29<89:11:13, 17.54s/it]
+
+ 2%|█▍ | 453/18756 [2:43:46<88:17:45, 17.37s/it]
+
+ 2%|█▍ | 454/18756 [2:44:04<89:14:52, 17.56s/it]
+
+ 2%|█▍ | 455/18756 [2:44:22<89:15:28, 17.56s/it]
+
+ 2%|█▍ | 456/18756 [2:44:39<88:31:42, 17.42s/it]
+
+ 2%|█▍ | 457/18756 [2:44:57<89:19:29, 17.57s/it]
+
+ 2%|█▍ | 458/18756 [2:45:13<87:43:19, 17.26s/it]
+
+ 2%|█▍ | 459/18756 [2:45:32<89:47:33, 17.67s/it]
+
+ 2%|█▍ | 460/18756 [2:45:48<88:08:39, 17.34s/it]
+
+ 2%|█▍ | 461/18756 [2:46:07<89:19:30, 17.58s/it]
+
+
+ 2%|█▌ | 463/18756 [2:46:41<88:29:13, 17.41s/it]
+
+ 2%|█▌ | 464/18756 [2:46:59<88:52:49, 17.49s/it]
+
+ 2%|█▌ | 465/18756 [2:47:16<87:50:46, 17.29s/it]
+
+ 2%|█▌ | 466/18756 [2:47:33<88:44:46, 17.47s/it]
+
+ 2%|█▌ | 467/18756 [2:47:51<88:21:53, 17.39s/it]
+{'loss': 1.3666, 'grad_norm': 7.020553387079956, 'learning_rate': 5.412610106629579e-06, 'epoch': 0.07}
+
+ 2%|█▌ | 468/18756 [2:48:08<88:28:48, 17.42s/it]
+
+ 3%|█▌ | 469/18756 [2:48:26<88:36:55, 17.44s/it]
+
+ 3%|█▌ | 470/18756 [2:48:42<87:12:27, 17.17s/it]
+
+ 3%|█▌ | 471/18756 [2:49:00<88:48:55, 17.49s/it]
+
+ 3%|█▌ | 472/18756 [2:49:17<87:27:57, 17.22s/it]
+
+ 3%|█▌ | 473/18756 [2:49:35<88:38:27, 17.45s/it]
+
+ 3%|█▌ | 474/18756 [2:49:52<88:28:37, 17.42s/it]
+
+
+ 3%|█▌ | 476/18756 [2:50:27<88:34:53, 17.44s/it]
+
+ 3%|█▌ | 477/18756 [2:50:44<87:35:31, 17.25s/it]
+
+ 3%|█▌ | 478/18756 [2:51:02<88:40:30, 17.47s/it]
+
+ 3%|█▌ | 479/18756 [2:51:19<87:28:20, 17.23s/it]
+
+ 3%|█▌ | 480/18756 [2:51:37<88:17:50, 17.39s/it]
+{'loss': 1.5254, 'grad_norm': 7.568251259308417, 'learning_rate': 5.563282336578582e-06, 'epoch': 0.08}
+
+
+ 3%|█▌ | 482/18756 [2:52:11<87:49:07, 17.30s/it]
+
+ 3%|█▌ | 483/18756 [2:52:29<88:08:42, 17.37s/it]
+
+ 3%|█▌ | 484/18756 [2:52:45<86:44:58, 17.09s/it]
+
+ 3%|█▌ | 485/18756 [2:53:03<88:39:04, 17.47s/it]
+{'loss': 1.5611, 'grad_norm': 6.546303662700586, 'learning_rate': 5.6212331942512744e-06, 'epoch': 0.08}
+
+ 3%|█▌ | 486/18756 [2:53:20<87:53:06, 17.32s/it]
+
+ 3%|█▌ | 487/18756 [2:53:38<88:09:05, 17.37s/it]
+
+ 3%|█▌ | 488/18756 [2:53:55<88:24:31, 17.42s/it]
+
+ 3%|█▌ | 489/18756 [2:54:12<86:41:56, 17.09s/it]
+
+ 3%|█▌ | 490/18756 [2:54:30<87:52:40, 17.32s/it]
+
+ 3%|█▌ | 491/18756 [2:54:46<86:56:32, 17.14s/it]
+
+ 3%|█▌ | 492/18756 [2:55:04<88:27:20, 17.44s/it]
+
+ 3%|█▌ | 493/18756 [2:55:22<88:48:21, 17.51s/it]
+
+ 3%|█▌ | 494/18756 [2:55:39<88:16:21, 17.40s/it]
+
+ 3%|█▌ | 495/18756 [2:55:56<88:03:16, 17.36s/it]
+
+ 3%|█▌ | 496/18756 [2:56:13<86:41:59, 17.09s/it]
+
+ 3%|█▌ | 497/18756 [2:56:31<88:01:37, 17.36s/it]
+
+ 3%|█▌ | 498/18756 [2:56:48<87:39:17, 17.28s/it]
+
+
+ 3%|█▌ | 500/18756 [3:04:55<776:16:11, 153.08s/it]
+{'loss': 1.1566, 'grad_norm': 6.221577830477056, 'learning_rate': 5.795085767269356e-06, 'epoch': 0.08}
+
+ 3%|█▌ | 501/18756 [3:05:12<568:38:08, 112.14s/it]
+
+ 3%|█▌ | 502/18756 [3:05:29<423:43:46, 83.57s/it]
+
+ 3%|█▌ | 503/18756 [3:05:46<322:56:27, 63.69s/it]
+
+ 3%|█▌ | 504/18756 [3:06:03<252:00:42, 49.71s/it]
+
+ 3%|█▌ | 505/18756 [3:06:22<204:34:38, 40.35s/it]
+
+ 3%|█▌ | 506/18756 [3:06:38<168:33:03, 33.25s/it]
+
+ 3%|█▌ | 507/18756 [3:06:56<144:38:07, 28.53s/it]
+
+ 3%|█▋ | 508/18756 [3:07:13<126:49:28, 25.02s/it]
+
+ 3%|█▋ | 509/18756 [3:07:30<114:35:12, 22.61s/it]
+
+ 3%|█▋ | 510/18756 [3:07:48<107:39:18, 21.24s/it]
+
+ 3%|█▋ | 511/18756 [3:08:04<100:42:01, 19.87s/it]
+
+ 3%|█▋ | 512/18756 [3:08:23<98:39:55, 19.47s/it]
+
+ 3%|█▋ | 513/18756 [3:08:40<94:33:26, 18.66s/it]
+
+ 3%|█▋ | 514/18756 [3:08:57<93:09:21, 18.38s/it]
+
+ 3%|█▋ | 515/18756 [3:09:14<90:36:04, 17.88s/it]
+
+ 3%|█▋ | 516/18756 [3:09:31<88:46:05, 17.52s/it]
+
+ 3%|█▋ | 517/18756 [3:09:49<90:23:08, 17.84s/it]
+
+ 3%|█▋ | 518/18756 [3:10:06<88:23:52, 17.45s/it]
+
+ 3%|█▋ | 519/18756 [3:10:25<90:07:49, 17.79s/it]
+
+ 3%|█▋ | 520/18756 [3:10:42<88:47:46, 17.53s/it]
+
+ 3%|█▋ | 521/18756 [3:10:59<88:18:47, 17.44s/it]
+
+ 3%|█▋ | 522/18756 [3:11:15<87:16:03, 17.23s/it]
+
+ 3%|█▋ | 523/18756 [3:11:32<86:14:26, 17.03s/it]
+
+ 3%|█▋ | 524/18756 [3:11:51<89:22:33, 17.65s/it]
+
+ 3%|█▋ | 525/18756 [3:12:08<88:20:32, 17.44s/it]
+
+ 3%|█▋ | 526/18756 [3:12:26<89:30:05, 17.67s/it]
+
+ 3%|█▋ | 527/18756 [3:12:44<88:45:37, 17.53s/it]
+
+ 3%|█▋ | 528/18756 [3:13:00<87:32:14, 17.29s/it]
+
+ 3%|█▋ | 529/18756 [3:13:17<86:43:54, 17.13s/it]
+
+ 3%|█▋ | 530/18756 [3:13:34<86:12:35, 17.03s/it]
+
+ 3%|█▋ | 531/18756 [3:13:53<88:51:31, 17.55s/it]
+
+ 3%|█▋ | 532/18756 [3:14:10<88:24:21, 17.46s/it]
+
+ 3%|█▋ | 533/18756 [3:14:28<89:11:15, 17.62s/it]
+
+ 3%|█▋ | 534/18756 [3:14:45<88:46:26, 17.54s/it]
+
+ 3%|█▋ | 535/18756 [3:15:02<87:50:16, 17.35s/it]
+
+ 3%|█▋ | 536/18756 [3:15:19<86:45:53, 17.14s/it]
+
+ 3%|█▋ | 537/18756 [3:15:36<86:49:52, 17.16s/it]
+
+ 3%|█▋ | 538/18756 [3:15:55<89:01:54, 17.59s/it]
+
+ 3%|█▊ | 539/18756 [3:16:11<88:05:20, 17.41s/it]
+
+ 3%|█▊ | 540/18756 [3:16:29<88:44:37, 17.54s/it]
+
+ 3%|█▊ | 541/18756 [3:16:46<87:31:39, 17.30s/it]
+
+ 3%|█▊ | 542/18756 [3:17:03<87:36:52, 17.32s/it]
+
+ 3%|█▊ | 543/18756 [3:17:20<86:43:09, 17.14s/it]
+
+ 3%|█▊ | 544/18756 [3:17:38<87:07:06, 17.22s/it]
+
+ 3%|█▊ | 545/18756 [3:17:56<89:05:02, 17.61s/it]
+
+ 3%|█▊ | 546/18756 [3:18:13<87:44:11, 17.34s/it]
+
+ 3%|█▊ | 547/18756 [3:18:31<88:26:14, 17.48s/it]
+
+ 3%|█▊ | 548/18756 [3:18:47<86:44:56, 17.15s/it]
+
+ 3%|█▊ | 549/18756 [3:19:05<87:36:42, 17.32s/it]
+
+ 3%|█▊ | 550/18756 [3:19:22<87:39:31, 17.33s/it]
+
+ 3%|█▊ | 551/18756 [3:19:39<87:24:28, 17.28s/it]
+
+ 3%|█▊ | 552/18756 [3:19:58<89:04:24, 17.62s/it]
+
+ 3%|█▊ | 553/18756 [3:20:14<87:09:30, 17.24s/it]
+
+ 3%|█▊ | 554/18756 [3:20:32<88:39:34, 17.54s/it]
+
+ 3%|█▊ | 555/18756 [3:20:49<86:50:50, 17.18s/it]
+
+ 3%|█▊ | 556/18756 [3:21:06<87:23:34, 17.29s/it]
+
+ 3%|█▊ | 557/18756 [3:21:24<87:36:46, 17.33s/it]
+
+ 3%|█▊ | 558/18756 [3:21:40<86:39:47, 17.14s/it]
+
+ 3%|█▊ | 559/18756 [3:21:59<89:02:29, 17.62s/it]
+
+ 3%|█▊ | 560/18756 [3:22:15<86:58:04, 17.21s/it]
+
+ 3%|█▊ | 561/18756 [3:22:33<88:31:21, 17.51s/it]
+
+ 3%|█▊ | 562/18756 [3:22:50<87:01:02, 17.22s/it]
+
+ 3%|█▊ | 563/18756 [3:23:07<86:57:07, 17.21s/it]
+
+ 3%|█▊ | 564/18756 [3:23:25<88:29:32, 17.51s/it]
+
+ 3%|█▊ | 565/18756 [3:23:42<86:53:29, 17.20s/it]
+
+ 3%|█▊ | 566/18756 [3:24:01<89:19:29, 17.68s/it]
+
+ 3%|█▊ | 567/18756 [3:24:17<87:00:47, 17.22s/it]
+
+ 3%|█▊ | 568/18756 [3:24:35<87:44:09, 17.37s/it]
+
+ 3%|█▊ | 569/18756 [3:24:52<87:24:21, 17.30s/it]
+
+ 3%|█▊ | 570/18756 [3:25:09<86:50:20, 17.19s/it]
+
+ 3%|█▊ | 571/18756 [3:25:27<87:57:03, 17.41s/it]
+
+ 3%|█▊ | 572/18756 [3:25:43<86:26:16, 17.11s/it]
+
+ 3%|█▊ | 573/18756 [3:26:02<88:46:40, 17.58s/it]
+
+ 3%|█▊ | 574/18756 [3:26:18<87:29:45, 17.32s/it]
+
+ 3%|█▊ | 575/18756 [3:26:36<87:42:37, 17.37s/it]
+
+ 3%|█▊ | 576/18756 [3:26:53<87:35:53, 17.35s/it]
+
+ 3%|█▉ | 577/18756 [3:27:10<86:25:47, 17.12s/it]
+
+ 3%|█▉ | 578/18756 [3:27:28<87:53:06, 17.40s/it]
+
+ 3%|█▉ | 579/18756 [3:27:45<87:10:08, 17.26s/it]
+
+ 3%|█▉ | 580/18756 [3:28:03<88:47:17, 17.59s/it]
+
+ 3%|█▉ | 581/18756 [3:28:20<87:41:41, 17.37s/it]
+
+ 3%|█▉ | 582/18756 [3:28:37<87:48:56, 17.40s/it]
+
+ 3%|█▉ | 583/18756 [3:28:54<86:58:01, 17.23s/it]
+
+ 3%|█▉ | 584/18756 [3:29:11<86:51:40, 17.21s/it]
+
+ 3%|█▉ | 585/18756 [3:29:29<88:03:26, 17.45s/it]
+
+ 3%|█▉ | 586/18756 [3:29:46<87:01:44, 17.24s/it]
+
+ 3%|█▉ | 587/18756 [3:30:05<88:49:30, 17.60s/it]
+
+ 3%|█▉ | 588/18756 [3:30:21<87:05:53, 17.26s/it]
+
+
+ 3%|█▉ | 590/18756 [3:30:56<87:05:17, 17.26s/it]
+{'loss': 1.5727, 'grad_norm': 6.233527486840755, 'learning_rate': 6.8382012053778394e-06, 'epoch': 0.09}
+
+ 3%|█▉ | 591/18756 [3:31:13<86:39:49, 17.18s/it]
+
+ 3%|█▉ | 592/18756 [3:31:31<88:10:13, 17.47s/it]
+
+ 3%|█▉ | 593/18756 [3:31:47<86:25:54, 17.13s/it]
+
+ 3%|█▉ | 594/18756 [3:32:06<88:58:32, 17.64s/it]
+
+ 3%|█▉ | 595/18756 [3:32:22<87:08:34, 17.27s/it]
+
+ 3%|█▉ | 596/18756 [3:32:40<87:53:33, 17.42s/it]
+
+ 3%|█▉ | 597/18756 [3:32:57<87:09:15, 17.28s/it]
+
+ 3%|█▉ | 598/18756 [3:33:14<85:59:14, 17.05s/it]
+
+ 3%|█▉ | 599/18756 [3:33:33<88:47:17, 17.60s/it]
+
+ 3%|█▉ | 600/18756 [3:41:23<774:13:55, 153.52s/it]
+
+ 3%|█▉ | 601/18756 [3:41:42<570:00:05, 113.03s/it]
+
+ 3%|█▉ | 602/18756 [3:41:58<423:43:08, 84.02s/it]
+
+ 3%|█▉ | 603/18756 [3:42:15<322:17:37, 63.92s/it]
+
+ 3%|█▉ | 604/18756 [3:42:32<251:10:28, 49.81s/it]
+
+ 3%|█▉ | 605/18756 [3:42:49<202:07:49, 40.09s/it]
+
+ 3%|█▉ | 606/18756 [3:43:07<168:14:01, 33.37s/it]
+
+ 3%|█▉ | 607/18756 [3:43:24<143:42:00, 28.50s/it]
+
+ 3%|█▉ | 608/18756 [3:43:42<127:58:56, 25.39s/it]
+
+ 3%|█▉ | 609/18756 [3:43:59<115:08:21, 22.84s/it]
+
+
+ 3%|█▉ | 611/18756 [3:44:33<99:46:25, 19.80s/it]
+{'loss': 1.2369, 'grad_norm': 7.651243729559977, 'learning_rate': 7.081594807603153e-06, 'epoch': 0.1}
+
+ 3%|█▉ | 612/18756 [3:44:50<96:25:41, 19.13s/it]
+
+ 3%|█▉ | 613/18756 [3:45:08<94:18:36, 18.71s/it]
+
+ 3%|█▉ | 614/18756 [3:45:25<92:00:21, 18.26s/it]
+
+ 3%|██ | 615/18756 [3:45:43<91:34:16, 18.17s/it]
+
+
+ 3%|██ | 617/18756 [3:46:17<88:19:35, 17.53s/it]
+{'loss': 1.0615, 'grad_norm': 6.198271959737265, 'learning_rate': 7.151135836810385e-06, 'epoch': 0.1}
+
+ 3%|██ | 618/18756 [3:46:34<87:53:48, 17.45s/it]
+
+
+ 3%|██ | 620/18756 [3:47:09<87:44:53, 17.42s/it]
+{'loss': 1.1606, 'grad_norm': 6.292416492100519, 'learning_rate': 7.185906351414001e-06, 'epoch': 0.1}
+
+ 3%|██ | 621/18756 [3:47:27<87:31:30, 17.37s/it]
+
+
+ 3%|██ | 623/18756 [3:48:01<87:08:16, 17.30s/it]
+{'loss': 1.2963, 'grad_norm': 15.329806267695561, 'learning_rate': 7.220676866017617e-06, 'epoch': 0.1}
+
+
+ 3%|██ | 625/18756 [3:48:35<86:26:23, 17.16s/it]
+
+ 3%|██ | 626/18756 [3:48:53<87:32:18, 17.38s/it]
+{'loss': 1.1941, 'grad_norm': 5.548577877364761, 'learning_rate': 7.255447380621233e-06, 'epoch': 0.1}
+
+ 3%|██ | 627/18756 [3:49:10<87:18:26, 17.34s/it]
+
+ 3%|██ | 628/18756 [3:49:28<87:13:21, 17.32s/it]
+
+ 3%|██ | 629/18756 [3:49:46<88:17:51, 17.54s/it]
+
+ 3%|██ | 630/18756 [3:50:02<86:14:46, 17.13s/it]
+
+ 3%|██ | 631/18756 [3:50:20<87:06:29, 17.30s/it]
+
+ 3%|██ | 632/18756 [3:50:37<86:29:25, 17.18s/it]
+
+ 3%|██ | 633/18756 [3:50:54<87:28:40, 17.38s/it]
+
+ 3%|██ | 634/18756 [3:51:12<87:48:17, 17.44s/it]
+
+ 3%|██ | 635/18756 [3:51:29<86:41:27, 17.22s/it]
+
+ 3%|██ | 636/18756 [3:51:47<88:28:10, 17.58s/it]
+
+ 3%|██ | 637/18756 [3:52:03<86:19:07, 17.15s/it]
+
+ 3%|██ | 638/18756 [3:52:21<87:18:24, 17.35s/it]
+
+ 3%|██ | 639/18756 [3:52:38<86:41:48, 17.23s/it]
+
+ 3%|██ | 640/18756 [3:52:55<86:50:33, 17.26s/it]
+
+
+ 3%|██ | 642/18756 [3:53:30<86:23:42, 17.17s/it]
+{'loss': 1.4218, 'grad_norm': 5.547897201696847, 'learning_rate': 7.440890125173853e-06, 'epoch': 0.1}
+
+ 3%|██ | 643/18756 [3:53:48<88:07:50, 17.52s/it]
+
+
+ 3%|██ | 645/18756 [3:54:22<86:36:29, 17.22s/it]
+{'loss': 1.3207, 'grad_norm': 10.59564570116928, 'learning_rate': 7.47566063977747e-06, 'epoch': 0.1}
+
+ 3%|██ | 646/18756 [3:54:39<86:52:54, 17.27s/it]
+
+ 3%|██ | 647/18756 [3:54:56<86:34:20, 17.21s/it]
+
+ 3%|██ | 648/18756 [3:55:14<88:03:38, 17.51s/it]
+
+ 3%|██ | 649/18756 [3:55:31<86:36:29, 17.22s/it]
+
+ 3%|██ | 650/18756 [3:55:49<87:37:23, 17.42s/it]
+
+ 3%|██ | 651/18756 [3:56:06<86:37:45, 17.23s/it]
+
+ 3%|██ | 652/18756 [3:56:23<86:26:41, 17.19s/it]
+
+ 3%|██ | 653/18756 [3:56:41<87:23:20, 17.38s/it]
+
+ 3%|██▏ | 654/18756 [3:56:58<86:54:19, 17.28s/it]
+
+
+ 3%|██▏ | 656/18756 [3:57:32<86:24:19, 17.19s/it]
+{'loss': 1.198, 'grad_norm': 5.996343770235312, 'learning_rate': 7.603152526657395e-06, 'epoch': 0.1}
+
+ 4%|██▏ | 657/18756 [3:57:49<86:55:26, 17.29s/it]
+
+
+ 4%|██▏ | 659/18756 [3:58:24<86:45:35, 17.26s/it]
+{'loss': 1.3015, 'grad_norm': 7.11827492807493, 'learning_rate': 7.637923041261011e-06, 'epoch': 0.11}
+
+
+ 4%|██▏ | 661/18756 [3:58:58<86:16:40, 17.16s/it]
+{'loss': 1.1445, 'grad_norm': 5.801614391979413, 'learning_rate': 7.661103384330089e-06, 'epoch': 0.11}
+
+ 4%|██▏ | 662/18756 [3:59:16<86:44:15, 17.26s/it]
+
+
+ 4%|██▏ | 664/18756 [3:59:51<87:13:12, 17.36s/it]
+
+ 4%|██▏ | 665/18756 [4:00:08<87:31:49, 17.42s/it]
+{'loss': 1.3239, 'grad_norm': 6.3516255618410025, 'learning_rate': 7.707464070468242e-06, 'epoch': 0.11}
+
+
+ 4%|██▏ | 667/18756 [4:00:42<86:45:10, 17.27s/it]
+{'loss': 1.1977, 'grad_norm': 6.259196420103739, 'learning_rate': 7.73064441353732e-06, 'epoch': 0.11}
+
+ 4%|██▏ | 668/18756 [4:01:00<86:55:53, 17.30s/it]
+
+
+ 4%|██▏ | 670/18756 [4:01:34<86:46:58, 17.27s/it]
+{'loss': 1.3187, 'grad_norm': 6.821335968096604, 'learning_rate': 7.765414928140936e-06, 'epoch': 0.11}
+
+ 4%|██▏ | 671/18756 [4:01:52<87:20:39, 17.39s/it]
+
+
+ 4%|██▏ | 673/18756 [4:02:26<87:10:21, 17.35s/it]
+{'loss': 1.1219, 'grad_norm': 7.6448758086152315, 'learning_rate': 7.800185442744554e-06, 'epoch': 0.11}
+
+
+ 4%|██▏ | 675/18756 [4:03:01<86:34:10, 17.24s/it]
+{'loss': 1.4499, 'grad_norm': 6.636479634450525, 'learning_rate': 7.82336578581363e-06, 'epoch': 0.11}
+
+ 4%|██▏ | 676/18756 [4:03:18<86:52:23, 17.30s/it]
+
+ 4%|██▏ | 677/18756 [4:03:35<86:48:07, 17.28s/it]
+
+ 4%|██▏ | 678/18756 [4:03:53<87:57:26, 17.52s/it]
+
+ 4%|██▏ | 679/18756 [4:04:10<87:11:06, 17.36s/it]
+
+
+ 4%|██▏ | 681/18756 [4:04:45<86:23:14, 17.21s/it]
+{'loss': 1.2514, 'grad_norm': 5.901579771453414, 'learning_rate': 7.892906815020863e-06, 'epoch': 0.11}
+
+ 4%|██▏ | 682/18756 [4:05:01<85:53:07, 17.11s/it]
+
+ 4%|██▏ | 683/18756 [4:05:20<87:35:57, 17.45s/it]
+
+ 4%|██▏ | 684/18756 [4:05:37<87:14:05, 17.38s/it]
+
+ 4%|██▏ | 685/18756 [4:05:55<88:29:41, 17.63s/it]
+
+ 4%|██▏ | 686/18756 [4:06:12<87:42:57, 17.48s/it]
+
+ 4%|██▏ | 687/18756 [4:06:29<87:11:39, 17.37s/it]
+
+ 4%|██▏ | 688/18756 [4:06:46<86:43:26, 17.28s/it]
+
+ 4%|██▏ | 689/18756 [4:07:03<86:03:22, 17.15s/it]
+
+ 4%|██▏ | 690/18756 [4:07:21<87:34:33, 17.45s/it]
+
+ 4%|██▏ | 691/18756 [4:07:39<87:18:04, 17.40s/it]
+
+ 4%|██▎ | 692/18756 [4:07:57<88:03:08, 17.55s/it]
+
+ 4%|██▎ | 693/18756 [4:08:14<87:24:04, 17.42s/it]
+
+ 4%|██▎ | 694/18756 [4:08:31<86:39:34, 17.27s/it]
+
+
+ 4%|██▎ | 696/18756 [4:09:05<86:24:21, 17.22s/it]
+{'loss': 1.4204, 'grad_norm': 5.954168433511025, 'learning_rate': 8.066759388038942e-06, 'epoch': 0.11}
+
+ 4%|██▎ | 697/18756 [4:09:23<87:28:28, 17.44s/it]
+
+ 4%|██▎ | 698/18756 [4:09:41<87:45:04, 17.49s/it]
+
+ 4%|██▎ | 699/18756 [4:09:58<87:59:48, 17.54s/it]
+
+ 4%|██▏ | 700/18756 [4:17:50<771:19:16, 153.79s/it]
+
+ 4%|██▏ | 701/18756 [4:18:08<566:26:53, 112.94s/it]
+
+ 4%|██▏ | 702/18756 [4:18:25<421:54:45, 84.13s/it]
+
+
+ 4%|██▎ | 704/18756 [4:18:58<249:14:16, 49.70s/it]
+{'loss': 1.3668, 'grad_norm': 6.696074505846078, 'learning_rate': 8.159480760315253e-06, 'epoch': 0.11}
+
+ 4%|██▎ | 705/18756 [4:19:16<202:06:24, 40.31s/it]
+
+ 4%|██▎ | 706/18756 [4:19:33<166:58:06, 33.30s/it]
+
+ 4%|██▎ | 707/18756 [4:19:51<143:35:56, 28.64s/it]
+
+ 4%|██▎ | 708/18756 [4:20:09<127:36:38, 25.45s/it]
+
+ 4%|██▎ | 709/18756 [4:20:25<113:58:07, 22.73s/it]
+
+ 4%|██▎ | 710/18756 [4:20:43<106:04:36, 21.16s/it]
+
+ 4%|██▎ | 711/18756 [4:20:59<98:44:11, 19.70s/it]
+
+ 4%|██▎ | 712/18756 [4:21:18<96:41:20, 19.29s/it]
+
+ 4%|██▎ | 713/18756 [4:21:35<93:12:36, 18.60s/it]
+
+ 4%|██▎ | 714/18756 [4:21:52<90:53:32, 18.14s/it]
+
+ 4%|██▎ | 715/18756 [4:22:10<90:55:29, 18.14s/it]
+
+ 4%|██▎ | 716/18756 [4:22:26<87:58:32, 17.56s/it]
+
+ 4%|██▎ | 717/18756 [4:22:44<88:16:50, 17.62s/it]
+
+ 4%|██▎ | 718/18756 [4:23:00<86:31:17, 17.27s/it]
+
+ 4%|██▎ | 719/18756 [4:23:18<87:19:08, 17.43s/it]
+
+
+ 4%|██▎ | 721/18756 [4:23:53<86:45:09, 17.32s/it]
+{'loss': 1.1239, 'grad_norm': 5.500301872985675, 'learning_rate': 8.356513676402411e-06, 'epoch': 0.12}
+
+ 4%|██▎ | 722/18756 [4:24:11<88:05:34, 17.59s/it]
+
+ 4%|██▎ | 723/18756 [4:24:27<86:03:18, 17.18s/it]
+
+ 4%|██▎ | 724/18756 [4:24:44<86:29:17, 17.27s/it]
+
+ 4%|██▎ | 725/18756 [4:25:01<86:02:39, 17.18s/it]
+
+ 4%|██▎ | 726/18756 [4:25:19<87:02:46, 17.38s/it]
+
+ 4%|██▎ | 727/18756 [4:25:37<87:59:08, 17.57s/it]
+
+
+ 4%|██▎ | 729/18756 [4:26:13<88:37:14, 17.70s/it]
+{'loss': 1.3387, 'grad_norm': 6.221121637707491, 'learning_rate': 8.44923504867872e-06, 'epoch': 0.12}
+
+ 4%|██▎ | 730/18756 [4:26:29<86:49:15, 17.34s/it]
+
+ 4%|██▍ | 731/18756 [4:26:46<86:27:42, 17.27s/it]
+
+ 4%|██▍ | 732/18756 [4:27:04<86:37:55, 17.30s/it]
+
+ 4%|██▍ | 733/18756 [4:27:21<87:01:10, 17.38s/it]
+
+ 4%|██▍ | 734/18756 [4:27:39<86:52:36, 17.35s/it]
+
+ 4%|██▍ | 735/18756 [4:27:56<86:42:25, 17.32s/it]
+
+ 4%|██▍ | 736/18756 [4:28:13<86:58:02, 17.37s/it]
+
+ 4%|██▍ | 737/18756 [4:28:31<86:47:13, 17.34s/it]
+
+
+ 4%|██▍ | 739/18756 [4:29:05<86:23:33, 17.26s/it]
+{'loss': 1.1479, 'grad_norm': 5.150540051337779, 'learning_rate': 8.565136764024109e-06, 'epoch': 0.12}
+
+ 4%|██▍ | 740/18756 [4:29:23<86:47:33, 17.34s/it]
+
+ 4%|██▍ | 741/18756 [4:29:39<86:09:25, 17.22s/it]
+
+ 4%|██▍ | 742/18756 [4:29:57<86:41:41, 17.33s/it]
+
+ 4%|██▍ | 743/18756 [4:30:15<86:54:53, 17.37s/it]
+
+
+ 4%|██▍ | 745/18756 [4:30:49<86:49:12, 17.35s/it]
+{'loss': 1.0737, 'grad_norm': 7.546722600120785, 'learning_rate': 8.63467779323134e-06, 'epoch': 0.12}
+
+ 4%|██▍ | 746/18756 [4:31:06<86:08:07, 17.22s/it]
+
+ 4%|██▍ | 747/18756 [4:31:24<87:17:30, 17.45s/it]
+
+ 4%|██▍ | 748/18756 [4:31:41<86:18:20, 17.25s/it]
+
+ 4%|██▍ | 749/18756 [4:31:58<86:54:27, 17.37s/it]
+
+ 4%|██▍ | 750/18756 [4:32:16<86:58:11, 17.39s/it]
+
+ 4%|██▍ | 751/18756 [4:32:33<86:03:56, 17.21s/it]
+
+ 4%|██▍ | 752/18756 [4:32:51<87:26:38, 17.48s/it]
+
+
+ 4%|██▍ | 754/18756 [4:33:26<87:24:59, 17.48s/it]
+{'loss': 1.4393, 'grad_norm': 5.96999038669956, 'learning_rate': 8.73898933704219e-06, 'epoch': 0.12}
+
+
+ 4%|██▍ | 756/18756 [4:34:00<86:26:38, 17.29s/it]
+{'loss': 1.3484, 'grad_norm': 4.160115903560222, 'learning_rate': 8.762169680111267e-06, 'epoch': 0.12}
+
+ 4%|██▍ | 757/18756 [4:34:17<87:04:50, 17.42s/it]
+
+ 4%|██▍ | 758/18756 [4:34:34<86:04:00, 17.22s/it]
+
+ 4%|██▍ | 759/18756 [4:34:52<87:29:59, 17.50s/it]
+
+ 4%|██▍ | 760/18756 [4:35:09<86:19:45, 17.27s/it]
+
+ 4%|██▍ | 761/18756 [4:35:26<86:23:40, 17.28s/it]
+
+ 4%|██▍ | 762/18756 [4:35:44<86:16:08, 17.26s/it]
+
+ 4%|██▍ | 763/18756 [4:36:01<86:48:41, 17.37s/it]
+
+
+ 4%|██▍ | 765/18756 [4:36:36<86:25:59, 17.30s/it]
+{'loss': 1.1792, 'grad_norm': 4.547002521837189, 'learning_rate': 8.866481223922115e-06, 'epoch': 0.12}
+
+ 4%|██▍ | 766/18756 [4:36:54<87:13:45, 17.46s/it]
+
+ 4%|██▍ | 767/18756 [4:37:10<86:06:02, 17.23s/it]
+
+ 4%|██▍ | 768/18756 [4:37:28<86:16:49, 17.27s/it]
+
+ 4%|██▌ | 769/18756 [4:37:45<86:24:54, 17.30s/it]
+
+
+ 4%|██▌ | 771/18756 [4:38:20<86:29:52, 17.31s/it]
+{'loss': 1.134, 'grad_norm': 6.29224947152171, 'learning_rate': 8.936022253129346e-06, 'epoch': 0.12}
+
+ 4%|██▌ | 772/18756 [4:38:37<86:04:14, 17.23s/it]
+
+ 4%|██▌ | 773/18756 [4:38:54<86:43:42, 17.36s/it]
+
+ 4%|██▌ | 774/18756 [4:39:12<86:25:28, 17.30s/it]
+
+
+ 4%|██▌ | 776/18756 [4:39:46<86:19:20, 17.28s/it]
+{'loss': 1.4973, 'grad_norm': 6.295412106647928, 'learning_rate': 8.99397311080204e-06, 'epoch': 0.12}
+
+ 4%|██▌ | 777/18756 [4:40:04<86:28:51, 17.32s/it]
+
+ 4%|██▌ | 778/18756 [4:40:20<85:49:27, 17.19s/it]
+
+ 4%|██▌ | 779/18756 [4:40:38<86:01:25, 17.23s/it]
+
+
+ 4%|██▌ | 781/18756 [4:41:12<85:58:52, 17.22s/it]
+{'loss': 1.1701, 'grad_norm': 6.765942995558019, 'learning_rate': 9.051923968474734e-06, 'epoch': 0.12}
+
+ 4%|██▌ | 782/18756 [4:41:30<86:35:36, 17.34s/it]
+
+ 4%|██▌ | 783/18756 [4:41:47<86:03:12, 17.24s/it]
+
+ 4%|██▌ | 784/18756 [4:42:05<87:13:27, 17.47s/it]
+
+ 4%|██▌ | 785/18756 [4:42:22<86:15:31, 17.28s/it]
+
+ 4%|██▌ | 786/18756 [4:42:39<86:07:14, 17.25s/it]
+
+ 4%|██▌ | 787/18756 [4:42:57<86:47:49, 17.39s/it]
+
+ 4%|██▌ | 788/18756 [4:43:13<85:46:49, 17.19s/it]
+
+ 4%|██▌ | 789/18756 [4:43:32<87:40:31, 17.57s/it]
+
+
+ 4%|██▌ | 791/18756 [4:44:07<87:31:24, 17.54s/it]
+{'loss': 1.1698, 'grad_norm': 4.474731283804186, 'learning_rate': 9.16782568382012e-06, 'epoch': 0.13}
+
+
+ 4%|██▌ | 793/18756 [4:44:41<86:01:22, 17.24s/it]
+{'loss': 1.2337, 'grad_norm': 5.502741563859594, 'learning_rate': 9.191006026889198e-06, 'epoch': 0.13}
+
+ 4%|██▌ | 794/18756 [4:44:58<86:47:16, 17.39s/it]
+
+ 4%|██▌ | 795/18756 [4:45:15<85:57:07, 17.23s/it]
+
+
+ 4%|██▌ | 797/18756 [4:45:51<87:03:23, 17.45s/it]
+{'loss': 1.3023, 'grad_norm': 5.829256249649029, 'learning_rate': 9.237366713027354e-06, 'epoch': 0.13}
+
+ 4%|██▌ | 798/18756 [4:46:08<87:19:33, 17.51s/it]
+
+ 4%|██▌ | 799/18756 [4:46:25<86:14:28, 17.29s/it]
+
+ 4%|██▌ | 800/18756 [4:54:16<764:37:38, 153.30s/it]
+
+ 4%|██▌ | 801/18756 [4:54:33<560:22:23, 112.36s/it]
+
+ 4%|██▌ | 802/18756 [4:54:50<417:23:19, 83.69s/it]
+
+ 4%|██▌ | 803/18756 [4:55:07<318:11:08, 63.80s/it]
+
+ 4%|██▌ | 804/18756 [4:55:23<247:09:36, 49.56s/it]
+
+ 4%|██▌ | 805/18756 [4:55:42<201:25:41, 40.40s/it]
+
+
+ 4%|██▌ | 807/18756 [4:56:17<143:49:10, 28.85s/it]
+{'loss': 1.1781, 'grad_norm': 7.259762209526369, 'learning_rate': 9.35326842837274e-06, 'epoch': 0.13}
+
+ 4%|██▌ | 808/18756 [4:56:34<126:09:07, 25.30s/it]
+
+ 4%|██▌ | 809/18756 [4:56:51<113:33:54, 22.78s/it]
+
+ 4%|██▌ | 810/18756 [4:57:09<105:21:48, 21.14s/it]
+
+
+ 4%|██▋ | 812/18756 [4:57:44<96:50:30, 19.43s/it]
+{'loss': 1.1288, 'grad_norm': 6.184833607143112, 'learning_rate': 9.411219286045433e-06, 'epoch': 0.13}
+
+ 4%|██▋ | 813/18756 [4:58:00<92:41:45, 18.60s/it]
+
+
+ 4%|██▋ | 815/18756 [4:58:36<90:22:08, 18.13s/it]
+{'loss': 1.3608, 'grad_norm': 15.098716408444037, 'learning_rate': 9.44598980064905e-06, 'epoch': 0.13}
+
+ 4%|██▋ | 816/18756 [4:58:53<88:30:48, 17.76s/it]
+
+ 4%|██▋ | 817/18756 [4:59:10<88:13:41, 17.71s/it]
+
+ 4%|██▋ | 818/18756 [4:59:27<86:01:08, 17.26s/it]
+
+
+ 4%|██▋ | 820/18756 [5:00:02<86:38:24, 17.39s/it]
+
+ 4%|██▋ | 821/18756 [5:00:20<87:52:02, 17.64s/it]
+
+ 4%|██▋ | 822/18756 [5:00:38<88:17:49, 17.72s/it]
+{'loss': 1.4941, 'grad_norm': 7.476813269289154, 'learning_rate': 9.527121001390821e-06, 'epoch': 0.13}
+
+ 4%|██▋ | 823/18756 [5:00:55<87:55:32, 17.65s/it]
+
+ 4%|██▋ | 824/18756 [5:01:13<87:38:22, 17.59s/it]
+
+ 4%|██▋ | 825/18756 [5:01:30<86:23:36, 17.35s/it]
+
+ 4%|██▋ | 826/18756 [5:01:46<85:46:47, 17.22s/it]
+
+
+ 4%|██▋ | 828/18756 [5:02:22<87:32:09, 17.58s/it]
+{'loss': 1.3492, 'grad_norm': 7.071536064707097, 'learning_rate': 9.596662030598054e-06, 'epoch': 0.13}
+
+ 4%|██▋ | 829/18756 [5:02:40<87:13:54, 17.52s/it]
+
+ 4%|██▋ | 830/18756 [5:02:58<88:01:25, 17.68s/it]
+
+
+ 4%|██▋ | 832/18756 [5:03:32<87:03:51, 17.49s/it]
+{'loss': 1.0758, 'grad_norm': 7.139075580667596, 'learning_rate': 9.643022716736208e-06, 'epoch': 0.13}
+
+ 4%|██▋ | 833/18756 [5:03:49<85:48:07, 17.23s/it]
+
+ 4%|██▋ | 834/18756 [5:04:06<85:45:25, 17.23s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/nn/modules/module.py:1879: UserWarning: Positional args are being deprecated, use kwargs instead. Refer to https://pytorch.org/docs/master/generated/torch.nn.Module.html#torch.nn.Module.state_dict for details.
+ warnings.warn(
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+{'loss': 1.0431, 'grad_norm': 4.474962517778496, 'learning_rate': 9.677793231339824e-06, 'epoch': 0.13}
+ 4%|██▋ | 835/18756 [5:05:37<196:24:00, 39.45s/it]
+
+ 4%|██▋ | 836/18756 [5:05:55<164:22:12, 33.02s/it]
+
+ 4%|██▋ | 837/18756 [5:06:12<140:36:40, 28.25s/it]
+
+ 4%|██▋ | 838/18756 [5:06:31<125:44:31, 25.26s/it]
+
+
+ 4%|██▋ | 840/18756 [5:07:06<106:47:38, 21.46s/it]
+
+ 4%|██▋ | 841/18756 [5:07:24<101:43:01, 20.44s/it]
+{'loss': 1.2721, 'grad_norm': 4.763446447490039, 'learning_rate': 9.747334260547057e-06, 'epoch': 0.13}
+
+ 4%|██▋ | 842/18756 [5:07:42<97:33:25, 19.61s/it]
+
+ 4%|██▋ | 843/18756 [5:07:59<93:55:46, 18.88s/it]
+
+ 4%|██▋ | 844/18756 [5:08:16<91:00:02, 18.29s/it]
+
+ 5%|██▋ | 845/18756 [5:08:34<89:45:29, 18.04s/it]
+
+ 5%|██▊ | 846/18756 [5:08:51<88:49:52, 17.86s/it]
+
+ 5%|██▊ | 847/18756 [5:09:09<88:56:12, 17.88s/it]
+[2024-05-15 18:32:27,933] [WARNING] [huggingface_hub.utils._http.http_backoff:293] [PID:20482] '(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/2b/86/2b867da18344f680fe1e7cea9c089573116f66372d72369faf098f79f908cb0e/03683fec6af1410f0b8007d321ed3813b602a33e87f9696703178aa5de9f506d?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240515%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240515T183057Z&X-Amz-Expires=86400&X-Amz-Signature=f87d654a6b010687719dda4ef790db4526bf766c1c77215b3ac82fdfb9839f57&X-Amz-SignedHeaders=host&partNumber=1&uploadId=eErqBG4YiSTsuCKXDQPDt8X1WvT4oO86JccFfLdj9ITOf067ixsHAk2I.SKHu0_4fPn_QbY1kFX764I6njtDaokltoSSjUe3A7D.g8ny8y5mGuWJzDFSvmuCnWko12YC&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: 524c7c89-d6ec-4c58-9d1f-d256cf4ef541)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/2b/86/2b867da18344f680fe1e7cea9c089573116f66372d72369faf098f79f908cb0e/03683fec6af1410f0b8007d321ed3813b602a33e87f9696703178aa5de9f506d?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240515%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240515T183057Z&X-Amz-Expires=86400&X-Amz-Signature=f87d654a6b010687719dda4ef790db4526bf766c1c77215b3ac82fdfb9839f57&X-Amz-SignedHeaders=host&partNumber=1&uploadId=eErqBG4YiSTsuCKXDQPDt8X1WvT4oO86JccFfLdj9ITOf067ixsHAk2I.SKHu0_4fPn_QbY1kFX764I6njtDaokltoSSjUe3A7D.g8ny8y5mGuWJzDFSvmuCnWko12YC&x-id=UploadPart
+ 5%|██▊ | 847/18756 [5:09:09<88:56:12, 17.88s/it]'(MaxRetryError("HTTPSConnectionPool(host='hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com', port=443): Max retries exceeded with url: /repos/2b/86/2b867da18344f680fe1e7cea9c089573116f66372d72369faf098f79f908cb0e/03683fec6af1410f0b8007d321ed3813b602a33e87f9696703178aa5de9f506d?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240515%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240515T183057Z&X-Amz-Expires=86400&X-Amz-Signature=f87d654a6b010687719dda4ef790db4526bf766c1c77215b3ac82fdfb9839f57&X-Amz-SignedHeaders=host&partNumber=1&uploadId=eErqBG4YiSTsuCKXDQPDt8X1WvT4oO86JccFfLdj9ITOf067ixsHAk2I.SKHu0_4fPn_QbY1kFX764I6njtDaokltoSSjUe3A7D.g8ny8y5mGuWJzDFSvmuCnWko12YC&x-id=UploadPart (Caused by SSLError(SSLEOFError(8, '[SSL: UNEXPECTED_EOF_WHILE_READING] EOF occurred in violation of protocol (_ssl.c:1007)')))"), '(Request ID: 524c7c89-d6ec-4c58-9d1f-d256cf4ef541)')' thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/2b/86/2b867da18344f680fe1e7cea9c089573116f66372d72369faf098f79f908cb0e/03683fec6af1410f0b8007d321ed3813b602a33e87f9696703178aa5de9f506d?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240515%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240515T183057Z&X-Amz-Expires=86400&X-Amz-Signature=f87d654a6b010687719dda4ef790db4526bf766c1c77215b3ac82fdfb9839f57&X-Amz-SignedHeaders=host&partNumber=1&uploadId=eErqBG4YiSTsuCKXDQPDt8X1WvT4oO86JccFfLdj9ITOf067ixsHAk2I.SKHu0_4fPn_QbY1kFX764I6njtDaokltoSSjUe3A7D.g8ny8y5mGuWJzDFSvmuCnWko12YC&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+{'loss': 1.3843, 'grad_norm': 5.5248726325220225, 'learning_rate': 9.828465461288827e-06, 'epoch': 0.14}
+ 5%|██▊ | 848/18756 [5:09:26<88:10:16, 17.72s/it]
+
+ 5%|██▊ | 849/18756 [5:09:44<88:16:09, 17.75s/it]
+
+ 5%|██▊ | 850/18756 [5:10:02<88:13:21, 17.74s/it]
+
+ 5%|██▊ | 851/18756 [5:10:19<87:35:51, 17.61s/it]
+
+ 5%|██▊ | 852/18756 [5:10:36<86:58:39, 17.49s/it]
+
+ 5%|██▊ | 853/18756 [5:10:54<86:35:39, 17.41s/it]
+
+ 5%|██▊ | 854/18756 [5:11:12<87:29:08, 17.59s/it]
+
+ 5%|██▊ | 855/18756 [5:11:30<87:57:42, 17.69s/it]
+
+ 5%|██▊ | 856/18756 [5:11:47<87:40:11, 17.63s/it]
+
+ 5%|██▊ | 857/18756 [5:12:05<88:15:58, 17.75s/it]
+
+ 5%|██▊ | 858/18756 [5:12:22<87:08:33, 17.53s/it]
+
+ 5%|██▊ | 859/18756 [5:12:40<87:47:29, 17.66s/it]
+
+ 5%|██▊ | 860/18756 [5:12:57<86:15:16, 17.35s/it]
+
+ 5%|██▊ | 861/18756 [5:13:15<88:07:12, 17.73s/it]
+
+ 5%|██▊ | 862/18756 [5:13:32<87:08:14, 17.53s/it]
+
+ 5%|██▊ | 863/18756 [5:13:50<87:44:07, 17.65s/it]
+
+ 5%|██▊ | 864/18756 [5:14:08<87:52:30, 17.68s/it]
+
+ 5%|██▊ | 865/18756 [5:14:25<87:12:07, 17.55s/it]
+
+ 5%|██▊ | 866/18756 [5:14:44<88:11:01, 17.75s/it]
+
+ 5%|██▊ | 867/18756 [5:15:00<86:50:15, 17.48s/it]
+
+ 5%|██▊ | 868/18756 [5:15:18<87:24:32, 17.59s/it]
+
+
+ 5%|██▊ | 870/18756 [5:15:54<88:03:13, 17.72s/it]
+{'loss': 1.2352, 'grad_norm': 6.922334608400844, 'learning_rate': 1.008344923504868e-05, 'epoch': 0.14}
+
+ 5%|██▊ | 871/18756 [5:16:11<86:48:04, 17.47s/it]
+
+ 5%|██▊ | 872/18756 [5:16:29<88:03:46, 17.73s/it]
+
+ 5%|██▊ | 873/18756 [5:16:46<87:32:15, 17.62s/it]
+
+ 5%|██▊ | 874/18756 [5:17:04<87:36:10, 17.64s/it]
+
+ 5%|██▊ | 875/18756 [5:17:21<87:19:30, 17.58s/it]
+
+ 5%|██▊ | 876/18756 [5:17:38<86:32:49, 17.43s/it]
+
+ 5%|██▊ | 877/18756 [5:17:56<86:40:25, 17.45s/it]
+
+ 5%|██▊ | 878/18756 [5:18:13<85:51:59, 17.29s/it]
+
+ 5%|██▊ | 879/18756 [5:18:31<86:46:39, 17.47s/it]
+
+ 5%|██▊ | 880/18756 [5:18:49<87:27:12, 17.61s/it]
+
+ 5%|██▊ | 881/18756 [5:19:06<87:21:29, 17.59s/it]
+
+ 5%|██▊ | 882/18756 [5:19:24<87:23:16, 17.60s/it]
+
+ 5%|██▊ | 883/18756 [5:19:42<87:34:49, 17.64s/it]
+
+ 5%|██▉ | 884/18756 [5:19:58<86:25:10, 17.41s/it]
+
+ 5%|██▉ | 885/18756 [5:20:16<86:04:07, 17.34s/it]
+
+ 5%|██▉ | 886/18756 [5:20:34<87:17:19, 17.58s/it]
+
+ 5%|██▉ | 887/18756 [5:20:51<86:35:18, 17.44s/it]
+
+ 5%|██▉ | 888/18756 [5:21:09<86:59:49, 17.53s/it]
+
+
+ 5%|██▉ | 890/18756 [5:21:44<86:52:47, 17.51s/it]
+
+ 5%|██▉ | 891/18756 [5:22:01<86:38:57, 17.46s/it]
+
+ 5%|██▉ | 892/18756 [5:22:18<85:40:59, 17.27s/it]
+
+ 5%|██▉ | 893/18756 [5:22:36<87:20:22, 17.60s/it]
+
+ 5%|██▉ | 894/18756 [5:22:53<85:38:01, 17.26s/it]
+{'loss': 1.3411, 'grad_norm': 5.894381430853672, 'learning_rate': 1.0361613351877608e-05, 'epoch': 0.14}
+[2024-05-15 18:46:07,027] [WARNING] [huggingface_hub.utils._http.http_backoff:285] [PID:20482] HTTP Error 500 thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/2b/86/2b867da18344f680fe1e7cea9c089573116f66372d72369faf098f79f908cb0e/69b3ef4401e0194aa541c55062fe326cbee9fb16c0e2ebca0ab96109125f5df0?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240515%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240515T183100Z&X-Amz-Expires=86400&X-Amz-Signature=eb216bfa3555f6718eafc2aec5cb2da28ccef4346612bab7039364754b37d2ad&X-Amz-SignedHeaders=host&partNumber=415&uploadId=EI7x7sAaAia_nuMT1vxN5HhbKekuOOxWO1ZSsIVgJsEqJjU1sxFs80XVkUm9Ep8odDAXtZ3NA8d4b2qbT7uLBSCx.nD2wFcUVxg.i1PUBbFQNkv5XvrpKeeDdWKxyhXw&x-id=UploadPart
+ 5%|██▉ | 894/18756 [5:22:53<85:38:01, 17.26s/it]HTTP Error 500 thrown while requesting PUT https://hf-hub-lfs-us-east-1.s3-accelerate.amazonaws.com/repos/2b/86/2b867da18344f680fe1e7cea9c089573116f66372d72369faf098f79f908cb0e/69b3ef4401e0194aa541c55062fe326cbee9fb16c0e2ebca0ab96109125f5df0?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Content-Sha256=UNSIGNED-PAYLOAD&X-Amz-Credential=AKIA2JU7TKAQFN2FTF47%2F20240515%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240515T183100Z&X-Amz-Expires=86400&X-Amz-Signature=eb216bfa3555f6718eafc2aec5cb2da28ccef4346612bab7039364754b37d2ad&X-Amz-SignedHeaders=host&partNumber=415&uploadId=EI7x7sAaAia_nuMT1vxN5HhbKekuOOxWO1ZSsIVgJsEqJjU1sxFs80XVkUm9Ep8odDAXtZ3NA8d4b2qbT7uLBSCx.nD2wFcUVxg.i1PUBbFQNkv5XvrpKeeDdWKxyhXw&x-id=UploadPart
+Retrying in 1s [Retry 1/5].
+{'loss': 1.3899, 'grad_norm': 8.584023704048128, 'learning_rate': 1.0373203523412147e-05, 'epoch': 0.14}
+ 5%|██▉ | 895/18756 [5:23:11<87:36:02, 17.66s/it]
+
+ 5%|██▉ | 896/18756 [5:23:28<86:33:34, 17.45s/it]
+
+ 5%|██▉ | 897/18756 [5:23:46<86:22:27, 17.41s/it]
+
+ 5%|██▉ | 898/18756 [5:24:03<86:18:24, 17.40s/it]
+
+ 5%|██▉ | 899/18756 [5:24:20<85:36:40, 17.26s/it]
+
+ 5%|██▊ | 900/18756 [5:32:10<759:50:26, 153.19s/it]
+
+ 5%|██▊ | 901/18756 [5:32:28<558:32:19, 112.61s/it]
+
+ 5%|██▉ | 902/18756 [5:32:45<415:59:19, 83.88s/it]
+
+ 5%|██▉ | 903/18756 [5:33:03<317:59:42, 64.12s/it]
+
+ 5%|██▉ | 904/18756 [5:33:20<247:33:42, 49.92s/it]
+
+
+ 5%|██▉ | 906/18756 [5:33:55<166:03:04, 33.49s/it]
+{'loss': 1.2303, 'grad_norm': 6.86733664913756, 'learning_rate': 1.0500695410292072e-05, 'epoch': 0.14}
+
+ 5%|██▉ | 907/18756 [5:34:12<141:52:45, 28.62s/it]
+
+ 5%|██▉ | 908/18756 [5:34:30<125:23:30, 25.29s/it]
+
+ 5%|██▉ | 909/18756 [5:34:47<112:20:58, 22.66s/it]
+
+ 5%|██▉ | 910/18756 [5:35:04<104:52:20, 21.16s/it]
+
+ 5%|██▉ | 911/18756 [5:35:21<98:49:50, 19.94s/it]
+
+ 5%|██▉ | 912/18756 [5:35:39<95:49:28, 19.33s/it]
+
+ 5%|██▉ | 913/18756 [5:35:57<93:17:13, 18.82s/it]
+
+ 5%|██▉ | 914/18756 [5:36:14<91:23:29, 18.44s/it]
+
+ 5%|██▉ | 915/18756 [5:36:31<88:55:55, 17.94s/it]
+
+ 5%|██▉ | 916/18756 [5:36:48<87:07:07, 17.58s/it]
+
+ 5%|██▉ | 917/18756 [5:37:05<86:56:50, 17.55s/it]
+
+ 5%|██▉ | 918/18756 [5:37:23<86:51:26, 17.53s/it]
+
+ 5%|██▉ | 919/18756 [5:37:41<87:51:08, 17.73s/it]
+
+ 5%|██▉ | 920/18756 [5:37:59<87:39:58, 17.69s/it]
+
+ 5%|██▉ | 921/18756 [5:38:16<87:28:57, 17.66s/it]
+
+ 5%|██▉ | 922/18756 [5:38:33<86:11:23, 17.40s/it]
+
+ 5%|███ | 923/18756 [5:38:50<85:31:25, 17.26s/it]
+
+ 5%|███ | 924/18756 [5:39:07<85:22:54, 17.24s/it]
+
+
+ 5%|███ | 926/18756 [5:39:43<87:13:18, 17.61s/it]
+{'loss': 1.2788, 'grad_norm': 4.885647809620882, 'learning_rate': 1.0732498840982847e-05, 'epoch': 0.15}
+
+
+ 5%|███ | 928/18756 [5:40:18<87:05:34, 17.59s/it]
+{'loss': 1.0678, 'grad_norm': 7.683385973216999, 'learning_rate': 1.0755679184051925e-05, 'epoch': 0.15}
+
+
+ 5%|███ | 930/18756 [5:40:52<85:40:19, 17.30s/it]
+{'loss': 1.1063, 'grad_norm': 6.263814688405429, 'learning_rate': 1.0778859527121003e-05, 'epoch': 0.15}
+
+ 5%|███ | 931/18756 [5:41:09<85:33:08, 17.28s/it]
+
+ 5%|███ | 932/18756 [5:41:27<85:45:45, 17.32s/it]
+
+ 5%|███ | 933/18756 [5:41:45<87:06:50, 17.60s/it]
+
+ 5%|███ | 934/18756 [5:42:02<85:50:02, 17.34s/it]
+
+ 5%|███ | 935/18756 [5:42:20<87:15:49, 17.63s/it]
+
+ 5%|███ | 936/18756 [5:42:36<85:32:51, 17.28s/it]
+
+
+ 5%|███ | 938/18756 [5:43:11<85:21:19, 17.25s/it]
+
+ 5%|███ | 939/18756 [5:43:28<85:13:16, 17.22s/it]
+{'loss': 1.2477, 'grad_norm': 6.350093582545069, 'learning_rate': 1.088317107093185e-05, 'epoch': 0.15}
+
+ 5%|███ | 940/18756 [5:43:47<87:04:03, 17.59s/it]
+
+ 5%|███ | 941/18756 [5:44:03<86:01:10, 17.38s/it]
+
+
+ 5%|███ | 943/18756 [5:44:38<85:49:37, 17.35s/it]
+
+ 5%|███ | 944/18756 [5:44:56<86:22:07, 17.46s/it]
+
+ 5%|███ | 945/18756 [5:45:13<85:31:31, 17.29s/it]
+
+ 5%|███ | 946/18756 [5:45:30<85:09:02, 17.21s/it]
+
+ 5%|███ | 947/18756 [5:45:48<86:33:28, 17.50s/it]
+
+ 5%|███ | 948/18756 [5:46:05<85:28:55, 17.28s/it]
+{'loss': 1.0817, 'grad_norm': 7.469528349217614, 'learning_rate': 1.0987482614742698e-05, 'epoch': 0.15}
+
+ 5%|███ | 949/18756 [5:46:24<87:28:36, 17.68s/it]
+
+ 5%|███ | 950/18756 [5:46:40<85:38:56, 17.32s/it]
+
+ 5%|███ | 951/18756 [5:46:58<86:06:13, 17.41s/it]
+
+ 5%|███ | 952/18756 [5:47:15<85:37:21, 17.31s/it]
+
+ 5%|███ | 953/18756 [5:47:32<85:14:01, 17.24s/it]
+
+
+ 5%|███ | 955/18756 [5:48:07<85:23:07, 17.27s/it]
+
+ 5%|███ | 956/18756 [5:48:25<87:04:39, 17.61s/it]
+{'loss': 1.4054, 'grad_norm': 5.269179234153621, 'learning_rate': 1.1080203987019009e-05, 'epoch': 0.15}
+
+ 5%|███ | 957/18756 [5:48:42<86:21:03, 17.47s/it]
+
+ 5%|███ | 958/18756 [5:49:00<87:08:43, 17.63s/it]
+
+ 5%|███ | 959/18756 [5:49:17<86:22:10, 17.47s/it]
+
+ 5%|███ | 960/18756 [5:49:34<85:55:57, 17.38s/it]
+
+ 5%|███▏ | 961/18756 [5:49:52<85:30:22, 17.30s/it]
+
+ 5%|███▏ | 962/18756 [5:50:09<85:38:19, 17.33s/it]
+
+ 5%|███▏ | 963/18756 [5:50:27<86:53:29, 17.58s/it]
+
+ 5%|███▏ | 964/18756 [5:50:44<86:19:20, 17.47s/it]
+
+ 5%|███▏ | 965/18756 [5:51:03<87:44:50, 17.76s/it]
+
+ 5%|███▏ | 966/18756 [5:51:20<86:28:35, 17.50s/it]
+
+ 5%|███▏ | 967/18756 [5:51:37<86:55:02, 17.59s/it]
+
+ 5%|███▏ | 968/18756 [5:51:54<85:29:24, 17.30s/it]
+
+
+ 5%|███▏ | 970/18756 [5:52:29<85:53:31, 17.39s/it]
+
+ 5%|███▏ | 971/18756 [5:52:46<85:18:04, 17.27s/it]
+
+ 5%|███▏ | 972/18756 [5:53:04<87:07:35, 17.64s/it]
+
+ 5%|███▏ | 973/18756 [5:53:21<85:31:05, 17.31s/it]
+
+ 5%|███▏ | 974/18756 [5:53:39<86:55:30, 17.60s/it]
+
+ 5%|███▏ | 975/18756 [5:53:56<85:14:41, 17.26s/it]
+{'loss': 1.5217, 'grad_norm': 5.577844607361089, 'learning_rate': 1.1300417246175243e-05, 'epoch': 0.16}
+
+ 5%|███▏ | 976/18756 [5:54:13<85:31:43, 17.32s/it]
+
+ 5%|███▏ | 977/18756 [5:54:31<85:42:15, 17.35s/it]
+
+ 5%|███▏ | 978/18756 [5:54:47<84:58:36, 17.21s/it]
+
+ 5%|███▏ | 979/18756 [5:55:06<87:25:30, 17.70s/it]
+
+ 5%|███▏ | 980/18756 [5:55:23<85:36:06, 17.34s/it]
+
+ 5%|███▏ | 981/18756 [5:55:41<87:19:22, 17.69s/it]
+
+ 5%|███▏ | 982/18756 [5:55:58<85:14:33, 17.27s/it]
+
+ 5%|███▏ | 983/18756 [5:56:15<84:58:36, 17.21s/it]
+
+ 5%|███▏ | 984/18756 [5:56:32<85:27:06, 17.31s/it]
+
+ 5%|███▏ | 985/18756 [5:56:49<84:50:03, 17.19s/it]
+
+ 5%|███▏ | 986/18756 [5:57:08<87:50:26, 17.80s/it]
+
+ 5%|███▏ | 987/18756 [5:57:25<85:57:45, 17.42s/it]
+
+ 5%|███▏ | 988/18756 [5:57:44<88:03:44, 17.84s/it]
+
+ 5%|███▏ | 989/18756 [5:58:00<85:47:39, 17.38s/it]
+
+ 5%|███▏ | 990/18756 [5:58:17<84:45:33, 17.18s/it]
+
+ 5%|███▏ | 991/18756 [5:58:34<85:21:29, 17.30s/it]
+
+ 5%|███▏ | 992/18756 [5:58:51<84:38:19, 17.15s/it]
+
+ 5%|███▏ | 993/18756 [5:59:10<87:30:37, 17.74s/it]
+
+ 5%|███▏ | 994/18756 [5:59:27<85:57:19, 17.42s/it]
+
+ 5%|███▏ | 995/18756 [5:59:45<87:30:53, 17.74s/it]
+
+ 5%|███▏ | 996/18756 [6:00:02<85:50:13, 17.40s/it]
+
+ 5%|███▏ | 997/18756 [6:00:19<84:57:08, 17.22s/it]
+
+ 5%|███▏ | 998/18756 [6:00:37<85:48:21, 17.40s/it]
+
+ 5%|███▏ | 999/18756 [6:00:54<85:07:19, 17.26s/it]
+
+ 5%|███ | 1000/18756 [6:08:44<755:33:46, 153.19s/it]
+
+ 5%|███ | 1001/18756 [6:09:01<553:58:18, 112.32s/it]
+
+ 5%|███▏ | 1002/18756 [6:09:18<412:41:40, 83.68s/it]
+
+ 5%|███▏ | 1003/18756 [6:09:34<313:26:39, 63.56s/it]
+
+ 5%|███▏ | 1004/18756 [6:09:52<245:00:06, 49.68s/it]
+
+ 5%|███▏ | 1005/18756 [6:10:10<198:18:05, 40.22s/it]
+
+ 5%|███▏ | 1006/18756 [6:10:28<165:02:46, 33.47s/it]
+
+ 5%|███▏ | 1007/18756 [6:10:45<141:48:01, 28.76s/it]
+
+ 5%|███▏ | 1008/18756 [6:11:03<125:09:16, 25.39s/it]
+
+ 5%|███▏ | 1009/18756 [6:11:20<112:50:44, 22.89s/it]
+
+ 5%|███▏ | 1010/18756 [6:11:37<104:09:29, 21.13s/it]
+
+
+ 5%|███▏ | 1012/18756 [6:12:11<94:05:37, 19.09s/it]
+{'loss': 1.1242, 'grad_norm': 6.314806224178679, 'learning_rate': 1.1729253592953176e-05, 'epoch': 0.16}
+
+ 5%|███▏ | 1013/18756 [6:12:29<92:30:04, 18.77s/it]
+
+ 5%|███▏ | 1014/18756 [6:12:47<90:45:41, 18.42s/it]
+
+ 5%|███▏ | 1015/18756 [6:13:04<89:37:41, 18.19s/it]
+
+ 5%|███▎ | 1016/18756 [6:13:21<87:59:00, 17.85s/it]
+
+
+ 5%|███▎ | 1018/18756 [6:13:55<85:30:40, 17.35s/it]
+{'loss': 1.1564, 'grad_norm': 7.603259501931453, 'learning_rate': 1.179879462216041e-05, 'epoch': 0.16}
+
+ 5%|███▎ | 1019/18756 [6:14:13<85:59:04, 17.45s/it]
+
+ 5%|███▎ | 1020/18756 [6:14:31<86:19:39, 17.52s/it]
+
+ 5%|███▎ | 1021/18756 [6:14:48<86:34:31, 17.57s/it]
+
+
+ 5%|███▎ | 1023/18756 [6:15:23<86:19:01, 17.52s/it]
+{'loss': 1.2129, 'grad_norm': 6.624671783954069, 'learning_rate': 1.1856745479833102e-05, 'epoch': 0.16}
+
+
+ 5%|███▎ | 1025/18756 [6:15:58<84:59:48, 17.26s/it]
+{'loss': 1.2441, 'grad_norm': 5.456254558408391, 'learning_rate': 1.187992582290218e-05, 'epoch': 0.16}
+
+ 5%|███▎ | 1026/18756 [6:16:15<84:53:10, 17.24s/it]
+
+ 5%|███▎ | 1027/18756 [6:16:33<86:01:05, 17.47s/it]
+
+ 5%|███▎ | 1028/18756 [6:16:50<85:59:26, 17.46s/it]
+
+ 5%|███▎ | 1029/18756 [6:17:08<86:53:45, 17.65s/it]
+
+ 5%|███▎ | 1030/18756 [6:17:25<86:03:25, 17.48s/it]
+
+ 5%|███▎ | 1031/18756 [6:17:43<86:37:58, 17.60s/it]
+
+ 6%|███▎ | 1032/18756 [6:18:00<85:36:33, 17.39s/it]
+
+ 6%|███▎ | 1033/18756 [6:18:17<85:14:58, 17.32s/it]
+
+ 6%|███▎ | 1034/18756 [6:18:34<85:08:22, 17.30s/it]
+
+ 6%|███▎ | 1035/18756 [6:18:51<84:18:40, 17.13s/it]
+
+ 6%|███▎ | 1036/18756 [6:19:10<86:17:50, 17.53s/it]
+
+ 6%|███▎ | 1037/18756 [6:19:27<85:30:21, 17.37s/it]
+
+
+ 6%|███▎ | 1039/18756 [6:20:02<85:46:56, 17.43s/it]
+{'loss': 1.2359, 'grad_norm': 7.837506122982183, 'learning_rate': 1.2042188224385721e-05, 'epoch': 0.17}
+
+ 6%|███▎ | 1040/18756 [6:20:19<85:36:05, 17.39s/it]
+
+ 6%|███▎ | 1041/18756 [6:20:36<85:08:27, 17.30s/it]
+
+ 6%|███▎ | 1042/18756 [6:20:53<83:49:31, 17.04s/it]
+
+ 6%|███▎ | 1043/18756 [6:21:11<85:44:10, 17.43s/it]
+
+ 6%|███▎ | 1044/18756 [6:21:28<84:38:09, 17.20s/it]
+
+ 6%|███▎ | 1045/18756 [6:21:46<86:04:11, 17.49s/it]
+
+ 6%|███▎ | 1046/18756 [6:22:04<86:16:01, 17.54s/it]
+
+ 6%|███▎ | 1047/18756 [6:22:21<86:04:01, 17.50s/it]
+
+ 6%|███▎ | 1048/18756 [6:22:39<86:23:54, 17.56s/it]
+
+ 6%|███▎ | 1049/18756 [6:22:55<84:54:38, 17.26s/it]
+
+ 6%|███▎ | 1050/18756 [6:23:14<86:25:57, 17.57s/it]
+
+ 6%|███▎ | 1051/18756 [6:23:30<84:29:35, 17.18s/it]
+
+ 6%|███▎ | 1052/18756 [6:23:48<85:40:15, 17.42s/it]
+
+ 6%|███▎ | 1053/18756 [6:24:05<85:24:00, 17.37s/it]
+
+
+ 6%|███▎ | 1055/18756 [6:24:40<86:27:15, 17.58s/it]
+{'loss': 1.2741, 'grad_norm': 6.4611887208376215, 'learning_rate': 1.222763096893834e-05, 'epoch': 0.17}
+
+ 6%|███▍ | 1056/18756 [6:24:57<85:30:00, 17.39s/it]
+
+ 6%|███▍ | 1057/18756 [6:25:15<86:28:59, 17.59s/it]
+
+ 6%|███▍ | 1058/18756 [6:25:32<84:31:36, 17.19s/it]
+
+
+ 6%|███▍ | 1060/18756 [6:26:07<84:55:58, 17.28s/it]
+{'loss': 1.4686, 'grad_norm': 8.144948093349605, 'learning_rate': 1.2285581826611033e-05, 'epoch': 0.17}
+
+ 6%|███▍ | 1061/18756 [6:26:24<85:22:18, 17.37s/it]
+
+ 6%|███▍ | 1062/18756 [6:26:42<86:21:18, 17.57s/it]
+
+ 6%|███▍ | 1063/18756 [6:26:59<85:43:08, 17.44s/it]
+
+ 6%|███▍ | 1064/18756 [6:27:17<86:23:08, 17.58s/it]
+
+ 6%|███▍ | 1065/18756 [6:27:34<84:50:06, 17.26s/it]
+
+ 6%|███▍ | 1066/18756 [6:27:51<85:04:21, 17.31s/it]
+
+ 6%|███▍ | 1067/18756 [6:28:08<84:56:25, 17.29s/it]
+
+ 6%|███▍ | 1068/18756 [6:28:26<85:47:59, 17.46s/it]
+
+ 6%|███▍ | 1069/18756 [6:28:44<86:15:05, 17.56s/it]
+
+ 6%|███▍ | 1070/18756 [6:29:02<86:28:23, 17.60s/it]
+
+ 6%|███▍ | 1071/18756 [6:29:19<86:32:48, 17.62s/it]
+
+ 6%|███▍ | 1072/18756 [6:29:37<85:51:00, 17.48s/it]
+
+ 6%|███▍ | 1073/18756 [6:29:54<85:42:10, 17.45s/it]
+
+ 6%|███▍ | 1074/18756 [6:30:11<85:29:49, 17.41s/it]
+
+ 6%|███▍ | 1075/18756 [6:30:29<85:30:00, 17.41s/it]
+
+ 6%|███▍ | 1076/18756 [6:30:46<85:16:32, 17.36s/it]
+
+ 6%|███▍ | 1077/18756 [6:31:04<85:48:23, 17.47s/it]
+
+ 6%|███▍ | 1078/18756 [6:31:21<85:28:49, 17.41s/it]
+
+ 6%|███▍ | 1079/18756 [6:31:39<86:11:04, 17.55s/it]
+
+ 6%|███▍ | 1080/18756 [6:31:56<85:30:13, 17.41s/it]
+
+ 6%|███▍ | 1081/18756 [6:32:13<85:39:10, 17.45s/it]
+
+ 6%|███▍ | 1082/18756 [6:32:30<85:05:39, 17.33s/it]
+
+ 6%|███▍ | 1083/18756 [6:32:47<84:23:54, 17.19s/it]
+
+ 6%|███▍ | 1084/18756 [6:33:05<85:15:41, 17.37s/it]
+
+ 6%|███▍ | 1085/18756 [6:33:22<84:22:39, 17.19s/it]
+
+
+ 6%|███▍ | 1087/18756 [6:33:57<85:34:08, 17.43s/it]
+
+ 6%|███▍ | 1088/18756 [6:34:15<86:16:54, 17.58s/it]
+
+ 6%|███▍ | 1089/18756 [6:34:33<85:40:58, 17.46s/it]
+
+ 6%|███▍ | 1090/18756 [6:34:49<84:51:36, 17.29s/it]
+{'loss': 1.2002, 'grad_norm': 5.139522724183391, 'learning_rate': 1.2633286972647196e-05, 'epoch': 0.17}
+
+ 6%|███▍ | 1091/18756 [6:35:07<85:13:37, 17.37s/it]
+
+ 6%|███▍ | 1092/18756 [6:35:23<83:51:13, 17.09s/it]
+
+ 6%|███▍ | 1093/18756 [6:35:42<85:51:02, 17.50s/it]
+
+ 6%|███▍ | 1094/18756 [6:35:59<84:49:29, 17.29s/it]
+
+ 6%|███▌ | 1095/18756 [6:36:17<86:01:50, 17.54s/it]
+
+
+ 6%|███▌ | 1097/18756 [6:36:52<85:40:21, 17.47s/it]
+{'loss': 1.1196, 'grad_norm': 6.944680561713984, 'learning_rate': 1.2714418173388968e-05, 'epoch': 0.18}
+
+ 6%|███▌ | 1098/18756 [6:37:09<85:10:37, 17.37s/it]
+
+ 6%|███▌ | 1099/18756 [6:37:25<83:43:20, 17.07s/it]
+
+ 6%|███▍ | 1100/18756 [6:45:17<753:09:37, 153.57s/it]
+
+ 6%|███▍ | 1101/18756 [6:45:34<551:28:10, 112.45s/it]
+
+ 6%|███▍ | 1102/18756 [6:45:51<411:02:38, 83.82s/it]
+
+
+ 6%|███▍ | 1104/18756 [6:46:24<243:10:57, 49.60s/it]
+{'loss': 1.3823, 'grad_norm': 7.583198657953405, 'learning_rate': 1.2795549374130738e-05, 'epoch': 0.18}
+
+ 6%|███▍ | 1105/18756 [6:46:44<199:06:24, 40.61s/it]
+
+ 6%|███▍ | 1106/18756 [6:47:00<163:28:56, 33.34s/it]
+
+ 6%|███▍ | 1107/18756 [6:47:19<142:25:42, 29.05s/it]
+
+ 6%|███▍ | 1108/18756 [6:47:36<123:41:39, 25.23s/it]
+
+ 6%|███▍ | 1109/18756 [6:47:53<111:51:15, 22.82s/it]
+
+ 6%|███▍ | 1110/18756 [6:48:10<103:28:24, 21.11s/it]
+
+ 6%|███▌ | 1111/18756 [6:48:26<96:11:26, 19.63s/it]
+
+ 6%|███▌ | 1112/18756 [6:48:46<95:57:39, 19.58s/it]
+
+ 6%|███▌ | 1113/18756 [6:49:02<91:19:17, 18.63s/it]
+
+ 6%|███▌ | 1114/18756 [6:49:22<92:47:25, 18.93s/it]
+
+ 6%|███▌ | 1115/18756 [6:49:38<89:06:31, 18.18s/it]
+
+ 6%|███▌ | 1116/18756 [6:49:56<88:39:07, 18.09s/it]
+
+ 6%|███▌ | 1117/18756 [6:50:13<86:50:34, 17.72s/it]
+
+ 6%|███▌ | 1118/18756 [6:50:29<85:01:56, 17.36s/it]
+
+ 6%|███▌ | 1119/18756 [6:50:47<85:58:38, 17.55s/it]
+
+ 6%|███▌ | 1120/18756 [6:51:04<84:10:18, 17.18s/it]
+
+ 6%|███▌ | 1121/18756 [6:51:23<86:52:32, 17.73s/it]
+
+ 6%|███▌ | 1122/18756 [6:51:39<85:10:11, 17.39s/it]
+
+ 6%|███▌ | 1123/18756 [6:51:58<86:49:16, 17.73s/it]
+
+ 6%|███▌ | 1124/18756 [6:52:15<85:35:00, 17.47s/it]
+
+ 6%|███▌ | 1125/18756 [6:52:32<85:20:51, 17.43s/it]
+
+ 6%|███▌ | 1126/18756 [6:52:50<85:29:36, 17.46s/it]
+
+ 6%|███▌ | 1127/18756 [6:53:06<83:55:14, 17.14s/it]
+
+ 6%|███▌ | 1128/18756 [6:53:24<85:30:44, 17.46s/it]
+
+ 6%|███▌ | 1129/18756 [6:53:41<84:26:25, 17.25s/it]
+
+ 6%|███▌ | 1130/18756 [6:54:00<86:53:39, 17.75s/it]
+
+ 6%|███▌ | 1131/18756 [6:54:17<85:40:58, 17.50s/it]
+
+ 6%|███▌ | 1132/18756 [6:54:35<86:25:13, 17.65s/it]
+
+ 6%|███▌ | 1133/18756 [6:54:52<85:43:20, 17.51s/it]
+
+ 6%|███▋ | 1134/18756 [6:55:09<84:50:33, 17.33s/it]
+
+ 6%|███▋ | 1135/18756 [6:55:26<85:01:34, 17.37s/it]
+
+ 6%|███▋ | 1136/18756 [6:55:43<84:04:54, 17.18s/it]
+
+ 6%|███▋ | 1137/18756 [6:56:01<85:40:14, 17.50s/it]
+
+ 6%|███▋ | 1138/18756 [6:56:18<84:51:44, 17.34s/it]
+
+ 6%|███▋ | 1139/18756 [6:56:37<86:18:33, 17.64s/it]
+
+
+ 6%|███▋ | 1141/18756 [6:57:11<85:38:36, 17.50s/it]
+{'loss': 1.249, 'grad_norm': 6.736937705729429, 'learning_rate': 1.322438572090867e-05, 'epoch': 0.18}
+
+ 6%|███▋ | 1142/18756 [6:57:28<84:27:34, 17.26s/it]
+
+ 6%|███▋ | 1143/18756 [6:57:45<84:02:15, 17.18s/it]
+
+ 6%|███▋ | 1144/18756 [6:58:02<84:32:56, 17.28s/it]
+
+ 6%|███▋ | 1145/18756 [6:58:20<84:47:06, 17.33s/it]
+
+ 6%|███▋ | 1146/18756 [6:58:39<86:40:01, 17.72s/it]
+
+ 6%|███▋ | 1147/18756 [6:58:56<85:58:43, 17.58s/it]
+
+ 6%|███▋ | 1148/18756 [6:59:14<86:41:55, 17.73s/it]
+
+ 6%|███▋ | 1149/18756 [6:59:30<84:59:36, 17.38s/it]
+
+ 6%|███▋ | 1150/18756 [6:59:48<85:15:58, 17.43s/it]
+
+ 6%|███▋ | 1151/18756 [7:00:05<84:12:18, 17.22s/it]
+
+ 6%|███▋ | 1152/18756 [7:00:22<84:14:52, 17.23s/it]
+
+ 6%|███▋ | 1153/18756 [7:00:40<85:29:07, 17.48s/it]
+
+ 6%|███▋ | 1154/18756 [7:00:57<84:49:09, 17.35s/it]
+
+ 6%|███▋ | 1155/18756 [7:01:16<87:13:34, 17.84s/it]
+
+ 6%|███▋ | 1156/18756 [7:01:33<85:36:43, 17.51s/it]
+
+ 6%|███▋ | 1157/18756 [7:01:51<86:39:37, 17.73s/it]
+
+ 6%|███▋ | 1158/18756 [7:02:08<84:51:02, 17.36s/it]
+
+
+ 6%|███▋ | 1160/18756 [7:02:42<84:19:36, 17.25s/it]
+{'loss': 1.2255, 'grad_norm': 6.111745077059652, 'learning_rate': 1.3444598980064907e-05, 'epoch': 0.19}
+
+ 6%|███▋ | 1161/18756 [7:02:59<83:44:38, 17.13s/it]
+
+ 6%|███▋ | 1162/18756 [7:03:18<86:43:50, 17.75s/it]
+
+ 6%|███▋ | 1163/18756 [7:03:35<85:17:38, 17.45s/it]
+
+ 6%|███▋ | 1164/18756 [7:03:54<87:27:25, 17.90s/it]
+
+ 6%|███▋ | 1165/18756 [7:04:10<84:56:27, 17.38s/it]
+
+ 6%|███▋ | 1166/18756 [7:04:27<85:19:26, 17.46s/it]
+
+ 6%|███▋ | 1167/18756 [7:04:44<84:40:56, 17.33s/it]
+
+ 6%|███▋ | 1168/18756 [7:05:01<84:08:56, 17.22s/it]
+
+
+ 6%|███▋ | 1170/18756 [7:05:36<84:17:50, 17.26s/it]
+{'loss': 1.1473, 'grad_norm': 6.679680646926309, 'learning_rate': 1.3560500695410292e-05, 'epoch': 0.19}
+
+ 6%|███▋ | 1171/18756 [7:05:55<86:42:46, 17.75s/it]
+
+ 6%|███▋ | 1172/18756 [7:06:11<84:28:32, 17.29s/it]
+
+ 6%|███▊ | 1173/18756 [7:06:30<85:59:14, 17.61s/it]
+
+ 6%|███▊ | 1174/18756 [7:06:47<84:59:38, 17.40s/it]
+
+ 6%|███▊ | 1175/18756 [7:07:04<85:08:22, 17.43s/it]
+
+ 6%|███▊ | 1176/18756 [7:07:22<85:19:20, 17.47s/it]
+
+ 6%|███▊ | 1177/18756 [7:07:38<84:14:02, 17.25s/it]
+
+ 6%|███▊ | 1178/18756 [7:07:57<85:38:26, 17.54s/it]
+
+ 6%|███▊ | 1179/18756 [7:08:13<84:23:14, 17.28s/it]
+
+ 6%|███▊ | 1180/18756 [7:08:32<86:32:26, 17.73s/it]
+
+ 6%|███▊ | 1181/18756 [7:08:49<85:06:26, 17.43s/it]
+
+ 6%|███▊ | 1182/18756 [7:09:07<85:43:34, 17.56s/it]
+
+ 6%|███▊ | 1183/18756 [7:09:24<84:49:20, 17.38s/it]
+
+ 6%|███▊ | 1184/18756 [7:09:41<84:17:11, 17.27s/it]
+
+ 6%|███▊ | 1185/18756 [7:09:58<84:07:18, 17.24s/it]
+
+ 6%|███▊ | 1186/18756 [7:10:15<83:50:41, 17.18s/it]
+
+ 6%|███▊ | 1187/18756 [7:10:34<86:08:09, 17.65s/it]
+
+ 6%|███▊ | 1188/18756 [7:10:50<84:59:39, 17.42s/it]
+
+ 6%|███▊ | 1189/18756 [7:11:09<86:44:14, 17.78s/it]
+
+ 6%|███▊ | 1190/18756 [7:11:26<84:46:13, 17.37s/it]
+
+ 6%|███▊ | 1191/18756 [7:11:43<85:37:09, 17.55s/it]
+
+ 6%|███▊ | 1192/18756 [7:12:00<84:50:30, 17.39s/it]
+
+ 6%|███▊ | 1193/18756 [7:12:18<84:37:27, 17.35s/it]
+
+ 6%|███▊ | 1194/18756 [7:12:35<85:06:44, 17.45s/it]
+
+ 6%|███▊ | 1195/18756 [7:12:52<84:24:05, 17.30s/it]
+
+ 6%|███▊ | 1196/18756 [7:13:11<85:57:41, 17.62s/it]
+
+ 6%|███▊ | 1197/18756 [7:13:27<83:50:18, 17.19s/it]
+
+ 6%|███▊ | 1198/18756 [7:13:45<85:26:44, 17.52s/it]
+
+
+ 6%|███▋ | 1200/18756 [7:21:56<752:03:17, 154.21s/it]
+{'loss': 1.2705, 'grad_norm': 6.267798448721754, 'learning_rate': 1.3908205841446454e-05, 'epoch': 0.19}
+
+ 6%|███▋ | 1201/18756 [7:22:13<551:55:17, 113.18s/it]
+
+ 6%|███▊ | 1202/18756 [7:22:31<411:47:02, 84.45s/it]
+
+ 6%|███▊ | 1203/18756 [7:22:47<312:33:23, 64.10s/it]
+
+ 6%|███▊ | 1204/18756 [7:23:05<244:11:36, 50.09s/it]
+
+
+ 6%|███▊ | 1206/18756 [7:23:40<163:40:26, 33.57s/it]
+{'loss': 1.165, 'grad_norm': 9.959865138354974, 'learning_rate': 1.3977746870653685e-05, 'epoch': 0.19}
+
+ 6%|███▊ | 1207/18756 [7:23:57<139:48:04, 28.68s/it]
+
+
+ 6%|███▊ | 1209/18756 [7:24:32<111:44:31, 22.93s/it]
+{'loss': 1.1755, 'grad_norm': 7.393952409000378, 'learning_rate': 1.4012517385257303e-05, 'epoch': 0.19}
+
+ 6%|███▊ | 1210/18756 [7:24:49<103:06:35, 21.16s/it]
+
+ 6%|███▊ | 1211/18756 [7:25:06<97:25:41, 19.99s/it]
+
+ 6%|███▉ | 1212/18756 [7:25:23<92:57:11, 19.07s/it]
+
+ 6%|███▉ | 1213/18756 [7:25:41<90:52:53, 18.65s/it]
+
+ 6%|███▉ | 1214/18756 [7:25:59<89:46:41, 18.42s/it]
+
+ 6%|███▉ | 1215/18756 [7:26:16<88:41:33, 18.20s/it]
+
+ 6%|███▉ | 1216/18756 [7:26:34<87:43:32, 18.01s/it]
+
+ 6%|███▉ | 1217/18756 [7:26:51<86:44:57, 17.81s/it]
+
+ 6%|███▉ | 1218/18756 [7:27:08<85:27:57, 17.54s/it]
+
+ 6%|███▉ | 1219/18756 [7:27:25<84:34:33, 17.36s/it]
+
+ 7%|███▉ | 1220/18756 [7:27:42<83:52:42, 17.22s/it]
+
+ 7%|███▉ | 1221/18756 [7:28:00<85:00:03, 17.45s/it]
+
+ 7%|███▉ | 1222/18756 [7:28:18<85:23:45, 17.53s/it]
+
+ 7%|███▉ | 1223/18756 [7:28:35<85:34:39, 17.57s/it]
+
+ 7%|███▉ | 1224/18756 [7:28:53<85:33:01, 17.57s/it]
+
+ 7%|███▉ | 1225/18756 [7:29:10<84:18:55, 17.31s/it]
+
+ 7%|███▉ | 1226/18756 [7:29:27<84:36:14, 17.37s/it]
+
+ 7%|███▉ | 1227/18756 [7:29:44<83:58:50, 17.25s/it]
+
+ 7%|███▉ | 1228/18756 [7:30:02<84:56:15, 17.44s/it]
+
+ 7%|███▉ | 1229/18756 [7:30:19<84:52:25, 17.43s/it]
+
+ 7%|███▉ | 1230/18756 [7:30:37<85:17:44, 17.52s/it]
+
+ 7%|███▉ | 1231/18756 [7:30:55<85:19:29, 17.53s/it]
+
+ 7%|███▉ | 1232/18756 [7:31:11<84:05:16, 17.27s/it]
+
+ 7%|███▉ | 1233/18756 [7:31:29<85:09:58, 17.50s/it]
+
+ 7%|███▉ | 1234/18756 [7:31:46<84:05:39, 17.28s/it]
+
+ 7%|███▉ | 1235/18756 [7:32:04<84:45:57, 17.42s/it]
+
+ 7%|███▉ | 1236/18756 [7:32:21<84:20:19, 17.33s/it]
+
+ 7%|███▉ | 1237/18756 [7:32:39<85:02:36, 17.48s/it]
+
+ 7%|███▉ | 1238/18756 [7:32:56<85:13:53, 17.52s/it]
+
+ 7%|███▉ | 1239/18756 [7:33:14<84:40:36, 17.40s/it]
+
+ 7%|███▉ | 1240/18756 [7:33:32<85:42:38, 17.62s/it]
+
+ 7%|███▉ | 1241/18756 [7:33:48<84:04:12, 17.28s/it]
+
+ 7%|███▉ | 1242/18756 [7:34:06<84:58:20, 17.47s/it]
+
+ 7%|███▉ | 1243/18756 [7:34:23<84:14:58, 17.32s/it]
+
+ 7%|███▉ | 1244/18756 [7:34:41<84:29:49, 17.37s/it]
+
+ 7%|███▉ | 1245/18756 [7:34:58<84:25:58, 17.36s/it]
+
+ 7%|███▉ | 1246/18756 [7:35:15<84:32:52, 17.38s/it]
+
+ 7%|███▉ | 1247/18756 [7:35:33<85:16:37, 17.53s/it]
+
+ 7%|███▉ | 1248/18756 [7:35:50<84:47:02, 17.43s/it]
+
+ 7%|███▉ | 1249/18756 [7:36:08<85:23:26, 17.56s/it]
+
+ 7%|███▉ | 1250/18756 [7:36:25<84:06:10, 17.30s/it]
+
+ 7%|████ | 1251/18756 [7:36:43<84:31:39, 17.38s/it]/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/nn/modules/module.py:1879: UserWarning: Positional args are being deprecated, use kwargs instead. Refer to https://pytorch.org/docs/master/generated/torch.nn.Module.html#torch.nn.Module.state_dict for details.
diff --git a/wandb/run-20240515_132302-ky6h7dv9/files/requirements.txt b/wandb/run-20240515_132302-ky6h7dv9/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..dbb544609f8c928733a9a2d5f4b71da3839a900a
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/files/requirements.txt
@@ -0,0 +1,192 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.28.0
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.0
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.15.0
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.13.1
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.15.0
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.0.dev0
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240515_132302-ky6h7dv9/files/tmp/axolotl_config_dayfyzi6.yml b/wandb/run-20240515_132302-ky6h7dv9/files/tmp/axolotl_config_dayfyzi6.yml
new file mode 100644
index 0000000000000000000000000000000000000000..e3004e1b8fd5ea2aa4d083a368cbb089f28479c4
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/files/tmp/axolotl_config_dayfyzi6.yml
@@ -0,0 +1,62 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+push_to_hub: true
+hf_use_auth_token: true
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 1
+gradient_accumulation_steps: 1
+num_epochs: 3
+learning_rate: 0.0001
+optimizer: adamw_bnb_8bit
+optim_args:
+ amsgrad: true
+lr_scheduler: cosine
+train_on_inputs: true
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.5
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero2.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240515_132302-ky6h7dv9/files/wandb-metadata.json b/wandb/run-20240515_132302-ky6h7dv9/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..af541d940db0e4b121de8fd300498b7c073f9921
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-15T13:23:03.174078",
+ "startedAt": "2024-05-15T13:23:02.544142",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "81da7d2531f19a6b443b05b627f9d692f7d71546"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1341.2250833333335,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1700.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2100.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.41,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3449.587,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1600.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3450.866,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2199.998,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.40616607666016
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json b/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..df79e3ce4713ea886c10a4d88f5663a8d7c290a9
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
@@ -0,0 +1 @@
+{"train/loss": 1.3403, "train/grad_norm": 6.2325761154919235, "train/learning_rate": 1.4499304589707929e-05, "train/epoch": 0.2, "train/global_step": 1251, "_timestamp": 1715806787.1312647, "_runtime": 27404.57999563217, "_step": 1250, "parameters/model.embed_tokens.weight": {"_type": "histogram", "values": [3.0, 5.0, 13.0, 16.0, 28.0, 100.0, 384.0, 2500.0, 14306.0, 85819.0, 421628.0, 1815276.0, 6478117.0, 16777216.0, 16777216.0, 16777216.0, 16777216.0, 16777216.0, 16777216.0, 16777216.0, 16777216.0, 9742432.0, 3191018.0, 756701.0, 172918.0, 34336.0, 6441.0, 1660.0, 437.0, 154.0, 92.0, 48.0, 25.0, 15.0, 9.0, 9.0, 5.0, 2.0, 4.0, 0.0, 1.0, 3.0, 2.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 0.0, 0.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.08251953125, -0.07772064208984375, -0.0729217529296875, -0.06812286376953125, -0.063323974609375, -0.05852508544921875, -0.0537261962890625, -0.04892730712890625, -0.04412841796875, -0.03932952880859375, -0.0345306396484375, -0.02973175048828125, -0.024932861328125, -0.02013397216796875, -0.0153350830078125, -0.01053619384765625, -0.0057373046875, -0.00093841552734375, 0.0038604736328125, 0.00865936279296875, 0.013458251953125, 0.01825714111328125, 0.0230560302734375, 0.02785491943359375, 0.03265380859375, 0.03745269775390625, 0.0422515869140625, 0.04705047607421875, 0.051849365234375, 0.05664825439453125, 0.0614471435546875, 0.06624603271484375, 0.071044921875, 0.07584381103515625, 0.0806427001953125, 0.08544158935546875, 0.090240478515625, 0.09503936767578125, 0.0998382568359375, 0.10463714599609375, 0.10943603515625, 0.11423492431640625, 0.1190338134765625, 0.12383270263671875, 0.128631591796875, 0.13343048095703125, 0.1382293701171875, 0.14302825927734375, 0.1478271484375, 0.15262603759765625, 0.1574249267578125, 0.16222381591796875, 0.167022705078125, 0.17182159423828125, 0.1766204833984375, 0.18141937255859375, 0.18621826171875, 0.19101715087890625, 0.1958160400390625, 0.20061492919921875, 0.205413818359375, 0.21021270751953125, 0.2150115966796875, 0.21981048583984375, 0.224609375]}, "parameters/model.layers.0.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 2.0, 1.0, 0.0, 3.0, 5.0, 3.0, 3.0, 13.0, 13.0, 19.0, 25.0, 33.0, 34.0, 58.0, 88.0, 141.0, 275.0, 464.0, 869.0, 1442.0, 2923.0, 6521.0, 16338.0, 49045.0, 187968.0, 897675.0, 9390899.0, 5151728.0, 820253.0, 178290.0, 44034.0, 15982.0, 5963.0, 2724.0, 1483.0, 795.0, 424.0, 220.0, 148.0, 92.0, 71.0, 50.0, 24.0, 18.0, 14.0, 14.0, 7.0, 2.0, 3.0, 5.0, 2.0, 0.0, 2.0, 1.0, 1.0, 0.0, 2.0], "bins": [-0.7734375, -0.75067138671875, -0.7279052734375, -0.70513916015625, -0.682373046875, -0.65960693359375, -0.6368408203125, -0.61407470703125, -0.59130859375, -0.56854248046875, -0.5457763671875, -0.52301025390625, -0.500244140625, -0.47747802734375, -0.4547119140625, -0.43194580078125, -0.4091796875, -0.38641357421875, -0.3636474609375, -0.34088134765625, -0.318115234375, -0.29534912109375, -0.2725830078125, -0.24981689453125, -0.22705078125, -0.20428466796875, -0.1815185546875, -0.15875244140625, -0.135986328125, -0.11322021484375, -0.0904541015625, -0.06768798828125, -0.044921875, -0.02215576171875, 0.0006103515625, 0.02337646484375, 0.046142578125, 0.06890869140625, 0.0916748046875, 0.11444091796875, 0.13720703125, 0.15997314453125, 0.1827392578125, 0.20550537109375, 0.228271484375, 0.25103759765625, 0.2738037109375, 0.29656982421875, 0.3193359375, 0.34210205078125, 0.3648681640625, 0.38763427734375, 0.410400390625, 0.43316650390625, 0.4559326171875, 0.47869873046875, 0.50146484375, 0.52423095703125, 0.5469970703125, 0.56976318359375, 0.592529296875, 0.61529541015625, 0.6380615234375, 0.66082763671875, 0.68359375]}, "parameters/model.layers.0.self_attn.k_proj.weight": {"_type": "histogram", "values": [3.0, 5.0, 1.0, 2.0, 1.0, 5.0, 3.0, 4.0, 9.0, 20.0, 23.0, 18.0, 76.0, 69.0, 92.0, 135.0, 277.0, 433.0, 734.0, 1233.0, 2158.0, 3717.0, 7561.0, 16023.0, 38330.0, 110357.0, 344886.0, 2237575.0, 989117.0, 286516.0, 91968.0, 33255.0, 14798.0, 6745.0, 3378.0, 1895.0, 1093.0, 704.0, 392.0, 245.0, 139.0, 103.0, 61.0, 51.0, 31.0, 20.0, 11.0, 12.0, 5.0, 0.0, 2.0, 3.0, 1.0, 1.0, 4.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.6015625, -0.58001708984375, -0.5584716796875, -0.53692626953125, -0.515380859375, -0.49383544921875, -0.4722900390625, -0.45074462890625, -0.42919921875, -0.40765380859375, -0.3861083984375, -0.36456298828125, -0.343017578125, -0.32147216796875, -0.2999267578125, -0.27838134765625, -0.2568359375, -0.23529052734375, -0.2137451171875, -0.19219970703125, -0.170654296875, -0.14910888671875, -0.1275634765625, -0.10601806640625, -0.08447265625, -0.06292724609375, -0.0413818359375, -0.01983642578125, 0.001708984375, 0.02325439453125, 0.0447998046875, 0.06634521484375, 0.087890625, 0.10943603515625, 0.1309814453125, 0.15252685546875, 0.174072265625, 0.19561767578125, 0.2171630859375, 0.23870849609375, 0.26025390625, 0.28179931640625, 0.3033447265625, 0.32489013671875, 0.346435546875, 0.36798095703125, 0.3895263671875, 0.41107177734375, 0.4326171875, 0.45416259765625, 0.4757080078125, 0.49725341796875, 0.518798828125, 0.54034423828125, 0.5618896484375, 0.58343505859375, 0.60498046875, 0.62652587890625, 0.6480712890625, 0.66961669921875, 0.691162109375, 0.71270751953125, 0.7342529296875, 0.75579833984375, 0.77734375]}, "parameters/model.layers.0.self_attn.v_proj.weight": {"_type": "histogram", "values": [4.0, 3.0, 7.0, 9.0, 13.0, 17.0, 32.0, 63.0, 75.0, 150.0, 205.0, 377.0, 643.0, 1021.0, 1560.0, 2523.0, 4172.0, 5934.0, 8505.0, 12752.0, 18002.0, 25298.0, 36895.0, 51359.0, 71915.0, 92514.0, 122896.0, 167566.0, 230048.0, 315386.0, 487020.0, 824268.0, 510943.0, 327178.0, 231629.0, 173361.0, 128339.0, 94379.0, 71048.0, 53511.0, 39149.0, 25981.0, 18474.0, 13029.0, 8843.0, 6009.0, 4019.0, 2689.0, 1779.0, 996.0, 677.0, 392.0, 264.0, 159.0, 93.0, 47.0, 35.0, 21.0, 13.0, 5.0, 5.0, 2.0, 0.0, 3.0], "bins": [-0.05029296875, -0.04869842529296875, -0.0471038818359375, -0.04550933837890625, -0.043914794921875, -0.04232025146484375, -0.0407257080078125, -0.03913116455078125, -0.03753662109375, -0.03594207763671875, -0.0343475341796875, -0.03275299072265625, -0.031158447265625, -0.02956390380859375, -0.0279693603515625, -0.02637481689453125, -0.0247802734375, -0.02318572998046875, -0.0215911865234375, -0.01999664306640625, -0.018402099609375, -0.01680755615234375, -0.0152130126953125, -0.01361846923828125, -0.01202392578125, -0.01042938232421875, -0.0088348388671875, -0.00724029541015625, -0.005645751953125, -0.00405120849609375, -0.0024566650390625, -0.00086212158203125, 0.000732421875, 0.00232696533203125, 0.0039215087890625, 0.00551605224609375, 0.007110595703125, 0.00870513916015625, 0.0102996826171875, 0.01189422607421875, 0.01348876953125, 0.01508331298828125, 0.0166778564453125, 0.01827239990234375, 0.019866943359375, 0.02146148681640625, 0.0230560302734375, 0.02465057373046875, 0.0262451171875, 0.02783966064453125, 0.0294342041015625, 0.03102874755859375, 0.032623291015625, 0.03421783447265625, 0.0358123779296875, 0.03740692138671875, 0.03900146484375, 0.04059600830078125, 0.0421905517578125, 0.04378509521484375, 0.045379638671875, 0.04697418212890625, 0.0485687255859375, 0.05016326904296875, 0.0517578125]}, "parameters/model.layers.0.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 2.0, 1.0, 2.0, 3.0, 2.0, 4.0, 1.0, 0.0, 3.0, 5.0, 11.0, 12.0, 13.0, 18.0, 21.0, 27.0, 52.0, 71.0, 144.0, 299.0, 803.0, 2768.0, 23291.0, 506559.0, 6045526.0, 8902882.0, 1228754.0, 59368.0, 4604.0, 1146.0, 381.0, 174.0, 97.0, 46.0, 31.0, 20.0, 9.0, 10.0, 12.0, 6.0, 6.0, 4.0, 6.0, 2.0, 0.0, 4.0, 0.0, 2.0, 2.0, 0.0, 2.0, 2.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.349609375, -0.337615966796875, -0.32562255859375, -0.313629150390625, -0.3016357421875, -0.289642333984375, -0.27764892578125, -0.265655517578125, -0.253662109375, -0.241668701171875, -0.22967529296875, -0.217681884765625, -0.2056884765625, -0.193695068359375, -0.18170166015625, -0.169708251953125, -0.15771484375, -0.145721435546875, -0.13372802734375, -0.121734619140625, -0.1097412109375, -0.097747802734375, -0.08575439453125, -0.073760986328125, -0.061767578125, -0.049774169921875, -0.03778076171875, -0.025787353515625, -0.0137939453125, -0.001800537109375, 0.01019287109375, 0.022186279296875, 0.0341796875, 0.046173095703125, 0.05816650390625, 0.070159912109375, 0.0821533203125, 0.094146728515625, 0.10614013671875, 0.118133544921875, 0.130126953125, 0.142120361328125, 0.15411376953125, 0.166107177734375, 0.1781005859375, 0.190093994140625, 0.20208740234375, 0.214080810546875, 0.22607421875, 0.238067626953125, 0.25006103515625, 0.262054443359375, 0.2740478515625, 0.286041259765625, 0.29803466796875, 0.310028076171875, 0.322021484375, 0.334014892578125, 0.34600830078125, 0.358001708984375, 0.3699951171875, 0.381988525390625, 0.39398193359375, 0.405975341796875, 0.41796875]}, "parameters/model.layers.0.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 1.0, 2.0, 0.0, 0.0, 3.0, 0.0, 8.0, 3.0, 4.0, 12.0, 11.0, 11.0, 14.0, 30.0, 58.0, 105.0, 264.0, 710.0, 2217.0, 10341.0, 138158.0, 3670799.0, 16777216.0, 16777216.0, 16777216.0, 1634928.0, 61529.0, 6141.0, 1438.0, 525.0, 187.0, 93.0, 41.0, 28.0, 21.0, 9.0, 5.0, 9.0, 5.0, 4.0, 3.0, 2.0, 1.0, 0.0, 1.0, 0.0, 2.0, 1.0], "bins": [-0.58984375, -0.574859619140625, -0.55987548828125, -0.544891357421875, -0.5299072265625, -0.514923095703125, -0.49993896484375, -0.484954833984375, -0.469970703125, -0.454986572265625, -0.44000244140625, -0.425018310546875, -0.4100341796875, -0.395050048828125, -0.38006591796875, -0.365081787109375, -0.35009765625, -0.335113525390625, -0.32012939453125, -0.305145263671875, -0.2901611328125, -0.275177001953125, -0.26019287109375, -0.245208740234375, -0.230224609375, -0.215240478515625, -0.20025634765625, -0.185272216796875, -0.1702880859375, -0.155303955078125, -0.14031982421875, -0.125335693359375, -0.1103515625, -0.095367431640625, -0.08038330078125, -0.065399169921875, -0.0504150390625, -0.035430908203125, -0.02044677734375, -0.005462646484375, 0.009521484375, 0.024505615234375, 0.03948974609375, 0.054473876953125, 0.0694580078125, 0.084442138671875, 0.09942626953125, 0.114410400390625, 0.12939453125, 0.144378662109375, 0.15936279296875, 0.174346923828125, 0.1893310546875, 0.204315185546875, 0.21929931640625, 0.234283447265625, 0.249267578125, 0.264251708984375, 0.27923583984375, 0.294219970703125, 0.3092041015625, 0.324188232421875, 0.33917236328125, 0.354156494140625, 0.369140625]}, "parameters/model.layers.0.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 3.0, 2.0, 1.0, 1.0, 1.0, 1.0, 3.0, 1.0, 0.0, 4.0, 2.0, 4.0, 7.0, 6.0, 7.0, 9.0, 14.0, 18.0, 35.0, 47.0, 68.0, 117.0, 240.0, 681.0, 2591.0, 31234.0, 1818535.0, 16777216.0, 16777216.0, 3862612.0, 74484.0, 3874.0, 796.0, 329.0, 130.0, 83.0, 44.0, 27.0, 27.0, 19.0, 10.0, 9.0, 11.0, 6.0, 4.0, 1.0, 2.0, 3.0, 2.0, 2.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.5625, -0.54498291015625, -0.5274658203125, -0.50994873046875, -0.492431640625, -0.47491455078125, -0.4573974609375, -0.43988037109375, -0.42236328125, -0.40484619140625, -0.3873291015625, -0.36981201171875, -0.352294921875, -0.33477783203125, -0.3172607421875, -0.29974365234375, -0.2822265625, -0.26470947265625, -0.2471923828125, -0.22967529296875, -0.212158203125, -0.19464111328125, -0.1771240234375, -0.15960693359375, -0.14208984375, -0.12457275390625, -0.1070556640625, -0.08953857421875, -0.072021484375, -0.05450439453125, -0.0369873046875, -0.01947021484375, -0.001953125, 0.01556396484375, 0.0330810546875, 0.05059814453125, 0.068115234375, 0.08563232421875, 0.1031494140625, 0.12066650390625, 0.13818359375, 0.15570068359375, 0.1732177734375, 0.19073486328125, 0.208251953125, 0.22576904296875, 0.2432861328125, 0.26080322265625, 0.2783203125, 0.29583740234375, 0.3133544921875, 0.33087158203125, 0.348388671875, 0.36590576171875, 0.3834228515625, 0.40093994140625, 0.41845703125, 0.43597412109375, 0.4534912109375, 0.47100830078125, 0.488525390625, 0.50604248046875, 0.5235595703125, 0.54107666015625, 0.55859375]}, "parameters/model.layers.0.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 301.0, 1188.0, 1024.0, 441.0, 148.0, 91.0, 72.0, 107.0, 111.0, 100.0, 66.0, 48.0, 35.0, 35.0, 37.0, 21.0, 19.0, 24.0, 19.0, 24.0, 17.0, 21.0, 24.0, 29.0, 15.0, 11.0, 10.0, 8.0, 7.0, 7.0, 5.0, 6.0, 1.0, 4.0, 2.0, 5.0, 3.0, 1.0, 1.0, 1.0, 1.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.2021484375, -0.1801910400390625, -0.158233642578125, -0.1362762451171875, -0.11431884765625, -0.0923614501953125, -0.070404052734375, -0.0484466552734375, -0.0264892578125, -0.0045318603515625, 0.017425537109375, 0.0393829345703125, 0.06134033203125, 0.0832977294921875, 0.105255126953125, 0.1272125244140625, 0.149169921875, 0.1711273193359375, 0.193084716796875, 0.2150421142578125, 0.23699951171875, 0.2589569091796875, 0.280914306640625, 0.3028717041015625, 0.3248291015625, 0.3467864990234375, 0.368743896484375, 0.3907012939453125, 0.41265869140625, 0.4346160888671875, 0.456573486328125, 0.4785308837890625, 0.50048828125, 0.5224456787109375, 0.544403076171875, 0.5663604736328125, 0.58831787109375, 0.6102752685546875, 0.632232666015625, 0.6541900634765625, 0.6761474609375, 0.6981048583984375, 0.720062255859375, 0.7420196533203125, 0.76397705078125, 0.7859344482421875, 0.807891845703125, 0.8298492431640625, 0.851806640625, 0.8737640380859375, 0.895721435546875, 0.9176788330078125, 0.93963623046875, 0.9615936279296875, 0.983551025390625, 1.0055084228515625, 1.0274658203125, 1.0494232177734375, 1.071380615234375, 1.0933380126953125, 1.11529541015625, 1.1372528076171875, 1.159210205078125, 1.1811676025390625, 1.203125]}, "parameters/model.layers.0.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 18.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 3.0, 1.0, 1.0, 5.0, 3.0, 2.0, 2.0, 2.0, 3.0, 7.0, 3.0, 7.0, 8.0, 12.0, 23.0, 59.0, 112.0, 578.0, 2063.0, 870.0, 180.0, 48.0, 32.0, 20.0, 8.0, 4.0, 5.0, 0.0, 3.0, 0.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00439453125, -0.00038909912109375, 0.0036163330078125, 0.00762176513671875, 0.011627197265625, 0.01563262939453125, 0.0196380615234375, 0.02364349365234375, 0.02764892578125, 0.03165435791015625, 0.0356597900390625, 0.03966522216796875, 0.043670654296875, 0.04767608642578125, 0.0516815185546875, 0.05568695068359375, 0.0596923828125, 0.06369781494140625, 0.0677032470703125, 0.07170867919921875, 0.075714111328125, 0.07971954345703125, 0.0837249755859375, 0.08773040771484375, 0.09173583984375, 0.09574127197265625, 0.0997467041015625, 0.10375213623046875, 0.107757568359375, 0.11176300048828125, 0.1157684326171875, 0.11977386474609375, 0.123779296875, 0.12778472900390625, 0.1317901611328125, 0.13579559326171875, 0.139801025390625, 0.14380645751953125, 0.1478118896484375, 0.15181732177734375, 0.15582275390625, 0.15982818603515625, 0.1638336181640625, 0.16783905029296875, 0.171844482421875, 0.17584991455078125, 0.1798553466796875, 0.18386077880859375, 0.1878662109375, 0.19187164306640625, 0.1958770751953125, 0.19988250732421875, 0.203887939453125, 0.20789337158203125, 0.2118988037109375, 0.21590423583984375, 0.21990966796875, 0.22391510009765625, 0.2279205322265625, 0.23192596435546875, 0.235931396484375, 0.23993682861328125, 0.2439422607421875, 0.24794769287109375, 0.251953125]}, "parameters/model.layers.1.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 2.0, 1.0, 1.0, 4.0, 6.0, 14.0, 18.0, 31.0, 51.0, 105.0, 185.0, 285.0, 688.0, 1697.0, 4903.0, 16032.0, 54988.0, 201430.0, 660005.0, 2166544.0, 6704846.0, 4816994.0, 1500864.0, 458838.0, 134995.0, 36988.0, 11036.0, 3300.0, 1278.0, 491.0, 262.0, 133.0, 84.0, 44.0, 28.0, 10.0, 10.0, 3.0, 4.0, 3.0, 2.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.5390625, -0.523590087890625, -0.50811767578125, -0.492645263671875, -0.4771728515625, -0.461700439453125, -0.44622802734375, -0.430755615234375, -0.415283203125, -0.399810791015625, -0.38433837890625, -0.368865966796875, -0.3533935546875, -0.337921142578125, -0.32244873046875, -0.306976318359375, -0.29150390625, -0.276031494140625, -0.26055908203125, -0.245086669921875, -0.2296142578125, -0.214141845703125, -0.19866943359375, -0.183197021484375, -0.167724609375, -0.152252197265625, -0.13677978515625, -0.121307373046875, -0.1058349609375, -0.090362548828125, -0.07489013671875, -0.059417724609375, -0.0439453125, -0.028472900390625, -0.01300048828125, 0.002471923828125, 0.0179443359375, 0.033416748046875, 0.04888916015625, 0.064361572265625, 0.079833984375, 0.095306396484375, 0.11077880859375, 0.126251220703125, 0.1417236328125, 0.157196044921875, 0.17266845703125, 0.188140869140625, 0.20361328125, 0.219085693359375, 0.23455810546875, 0.250030517578125, 0.2655029296875, 0.280975341796875, 0.29644775390625, 0.311920166015625, 0.327392578125, 0.342864990234375, 0.35833740234375, 0.373809814453125, 0.3892822265625, 0.404754638671875, 0.42022705078125, 0.435699462890625, 0.451171875]}, "parameters/model.layers.1.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 3.0, 3.0, 6.0, 2.0, 4.0, 12.0, 13.0, 16.0, 42.0, 50.0, 82.0, 130.0, 174.0, 284.0, 437.0, 715.0, 978.0, 1741.0, 2974.0, 4932.0, 8389.0, 15362.0, 24938.0, 48276.0, 86501.0, 147220.0, 262423.0, 439878.0, 817931.0, 1042554.0, 541190.0, 320235.0, 182248.0, 106330.0, 61548.0, 32821.0, 18033.0, 10610.0, 6029.0, 3607.0, 1989.0, 1264.0, 822.0, 524.0, 312.0, 239.0, 146.0, 91.0, 58.0, 45.0, 31.0, 22.0, 19.0, 6.0, 3.0, 4.0, 4.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.337890625, -0.327392578125, -0.31689453125, -0.306396484375, -0.2958984375, -0.285400390625, -0.27490234375, -0.264404296875, -0.25390625, -0.243408203125, -0.23291015625, -0.222412109375, -0.2119140625, -0.201416015625, -0.19091796875, -0.180419921875, -0.169921875, -0.159423828125, -0.14892578125, -0.138427734375, -0.1279296875, -0.117431640625, -0.10693359375, -0.096435546875, -0.0859375, -0.075439453125, -0.06494140625, -0.054443359375, -0.0439453125, -0.033447265625, -0.02294921875, -0.012451171875, -0.001953125, 0.008544921875, 0.01904296875, 0.029541015625, 0.0400390625, 0.050537109375, 0.06103515625, 0.071533203125, 0.08203125, 0.092529296875, 0.10302734375, 0.113525390625, 0.1240234375, 0.134521484375, 0.14501953125, 0.155517578125, 0.166015625, 0.176513671875, 0.18701171875, 0.197509765625, 0.2080078125, 0.218505859375, 0.22900390625, 0.239501953125, 0.25, 0.260498046875, 0.27099609375, 0.281494140625, 0.2919921875, 0.302490234375, 0.31298828125, 0.323486328125, 0.333984375]}, "parameters/model.layers.1.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 0.0, 2.0, 0.0, 6.0, 10.0, 28.0, 51.0, 165.0, 396.0, 1050.0, 2404.0, 5550.0, 11762.0, 23938.0, 45448.0, 82808.0, 144376.0, 218689.0, 332869.0, 446934.0, 556954.0, 630178.0, 524134.0, 412240.0, 296888.0, 191895.0, 122491.0, 70416.0, 37109.0, 18558.0, 9339.0, 4482.0, 1863.0, 732.0, 310.0, 128.0, 41.0, 30.0, 10.0, 3.0, 3.0, 2.0, 1.0, 2.0, 2.0, 0.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.09033203125, -0.08777618408203125, -0.0852203369140625, -0.08266448974609375, -0.080108642578125, -0.07755279541015625, -0.0749969482421875, -0.07244110107421875, -0.06988525390625, -0.06732940673828125, -0.0647735595703125, -0.06221771240234375, -0.059661865234375, -0.05710601806640625, -0.0545501708984375, -0.05199432373046875, -0.0494384765625, -0.04688262939453125, -0.0443267822265625, -0.04177093505859375, -0.039215087890625, -0.03665924072265625, -0.0341033935546875, -0.03154754638671875, -0.02899169921875, -0.02643585205078125, -0.0238800048828125, -0.02132415771484375, -0.018768310546875, -0.01621246337890625, -0.0136566162109375, -0.01110076904296875, -0.008544921875, -0.00598907470703125, -0.0034332275390625, -0.00087738037109375, 0.001678466796875, 0.00423431396484375, 0.0067901611328125, 0.00934600830078125, 0.01190185546875, 0.01445770263671875, 0.0170135498046875, 0.01956939697265625, 0.022125244140625, 0.02468109130859375, 0.0272369384765625, 0.02979278564453125, 0.0323486328125, 0.03490447998046875, 0.0374603271484375, 0.04001617431640625, 0.042572021484375, 0.04512786865234375, 0.0476837158203125, 0.05023956298828125, 0.05279541015625, 0.05535125732421875, 0.0579071044921875, 0.06046295166015625, 0.063018798828125, 0.06557464599609375, 0.0681304931640625, 0.07068634033203125, 0.0732421875]}, "parameters/model.layers.1.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 3.0, 8.0, 5.0, 6.0, 17.0, 17.0, 30.0, 44.0, 58.0, 84.0, 133.0, 204.0, 351.0, 596.0, 1148.0, 2535.0, 6922.0, 53474.0, 878887.0, 6063216.0, 7899210.0, 1730621.0, 122445.0, 10779.0, 3275.0, 1382.0, 695.0, 395.0, 257.0, 144.0, 80.0, 53.0, 40.0, 29.0, 12.0, 16.0, 9.0, 8.0, 7.0, 3.0, 4.0, 1.0, 2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.353515625, -0.3421630859375, -0.330810546875, -0.3194580078125, -0.30810546875, -0.2967529296875, -0.285400390625, -0.2740478515625, -0.2626953125, -0.2513427734375, -0.239990234375, -0.2286376953125, -0.21728515625, -0.2059326171875, -0.194580078125, -0.1832275390625, -0.171875, -0.1605224609375, -0.149169921875, -0.1378173828125, -0.12646484375, -0.1151123046875, -0.103759765625, -0.0924072265625, -0.0810546875, -0.0697021484375, -0.058349609375, -0.0469970703125, -0.03564453125, -0.0242919921875, -0.012939453125, -0.0015869140625, 0.009765625, 0.0211181640625, 0.032470703125, 0.0438232421875, 0.05517578125, 0.0665283203125, 0.077880859375, 0.0892333984375, 0.1005859375, 0.1119384765625, 0.123291015625, 0.1346435546875, 0.14599609375, 0.1573486328125, 0.168701171875, 0.1800537109375, 0.19140625, 0.2027587890625, 0.214111328125, 0.2254638671875, 0.23681640625, 0.2481689453125, 0.259521484375, 0.2708740234375, 0.2822265625, 0.2935791015625, 0.304931640625, 0.3162841796875, 0.32763671875, 0.3389892578125, 0.350341796875, 0.3616943359375, 0.373046875]}, "parameters/model.layers.1.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 1.0, 2.0, 4.0, 4.0, 1.0, 3.0, 2.0, 2.0, 4.0, 6.0, 9.0, 10.0, 7.0, 23.0, 14.0, 13.0, 15.0, 33.0, 40.0, 64.0, 148.0, 360.0, 896.0, 2640.0, 17962.0, 452582.0, 7950900.0, 16777216.0, 16777216.0, 15363595.0, 1258421.0, 46072.0, 4177.0, 1265.0, 479.0, 204.0, 99.0, 50.0, 27.0, 20.0, 20.0, 15.0, 11.0, 11.0, 4.0, 6.0, 4.0, 6.0, 2.0, 6.0, 2.0, 5.0, 6.0, 5.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.41015625, -0.396087646484375, -0.38201904296875, -0.367950439453125, -0.3538818359375, -0.339813232421875, -0.32574462890625, -0.311676025390625, -0.297607421875, -0.283538818359375, -0.26947021484375, -0.255401611328125, -0.2413330078125, -0.227264404296875, -0.21319580078125, -0.199127197265625, -0.18505859375, -0.170989990234375, -0.15692138671875, -0.142852783203125, -0.1287841796875, -0.114715576171875, -0.10064697265625, -0.086578369140625, -0.072509765625, -0.058441162109375, -0.04437255859375, -0.030303955078125, -0.0162353515625, -0.002166748046875, 0.01190185546875, 0.025970458984375, 0.0400390625, 0.054107666015625, 0.06817626953125, 0.082244873046875, 0.0963134765625, 0.110382080078125, 0.12445068359375, 0.138519287109375, 0.152587890625, 0.166656494140625, 0.18072509765625, 0.194793701171875, 0.2088623046875, 0.222930908203125, 0.23699951171875, 0.251068115234375, 0.26513671875, 0.279205322265625, 0.29327392578125, 0.307342529296875, 0.3214111328125, 0.335479736328125, 0.34954833984375, 0.363616943359375, 0.377685546875, 0.391754150390625, 0.40582275390625, 0.419891357421875, 0.4339599609375, 0.448028564453125, 0.46209716796875, 0.476165771484375, 0.490234375]}, "parameters/model.layers.1.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 2.0, 1.0, 1.0, 5.0, 4.0, 3.0, 5.0, 8.0, 9.0, 16.0, 28.0, 52.0, 175.0, 424.0, 1612.0, 60251.0, 4633515.0, 16777216.0, 16777216.0, 1064751.0, 8541.0, 912.0, 276.0, 103.0, 48.0, 20.0, 9.0, 13.0, 8.0, 10.0, 4.0, 5.0, 1.0, 0.0, 3.0, 2.0, 0.0, 2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 4.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.5546875, -0.53607177734375, -0.5174560546875, -0.49884033203125, -0.480224609375, -0.46160888671875, -0.4429931640625, -0.42437744140625, -0.40576171875, -0.38714599609375, -0.3685302734375, -0.34991455078125, -0.331298828125, -0.31268310546875, -0.2940673828125, -0.27545166015625, -0.2568359375, -0.23822021484375, -0.2196044921875, -0.20098876953125, -0.182373046875, -0.16375732421875, -0.1451416015625, -0.12652587890625, -0.10791015625, -0.08929443359375, -0.0706787109375, -0.05206298828125, -0.033447265625, -0.01483154296875, 0.0037841796875, 0.02239990234375, 0.041015625, 0.05963134765625, 0.0782470703125, 0.09686279296875, 0.115478515625, 0.13409423828125, 0.1527099609375, 0.17132568359375, 0.18994140625, 0.20855712890625, 0.2271728515625, 0.24578857421875, 0.264404296875, 0.28302001953125, 0.3016357421875, 0.32025146484375, 0.3388671875, 0.35748291015625, 0.3760986328125, 0.39471435546875, 0.413330078125, 0.43194580078125, 0.4505615234375, 0.46917724609375, 0.48779296875, 0.50640869140625, 0.5250244140625, 0.54364013671875, 0.562255859375, 0.58087158203125, 0.5994873046875, 0.61810302734375, 0.63671875]}, "parameters/model.layers.1.input_layernorm.weight": {"_type": "histogram", "values": [38.0, 0.0, 1.0, 23.0, 770.0, 1247.0, 686.0, 346.0, 196.0, 154.0, 89.0, 48.0, 64.0, 53.0, 35.0, 28.0, 28.0, 18.0, 24.0, 20.0, 23.0, 14.0, 20.0, 22.0, 20.0, 18.0, 19.0, 7.0, 11.0, 13.0, 13.0, 5.0, 4.0, 3.0, 2.0, 3.0, 6.0, 5.0, 3.0, 5.0, 0.0, 0.0, 2.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00018787384033203125, 0.014219358563423157, 0.028626590967178345, 0.04303382337093353, 0.05744105577468872, 0.07184828817844391, 0.0862555205821991, 0.10066275298595428, 0.11506998538970947, 0.12947721779346466, 0.14388445019721985, 0.15829168260097504, 0.17269891500473022, 0.1871061474084854, 0.2015133798122406, 0.2159206122159958, 0.23032784461975098, 0.24473507702350616, 0.25914230942726135, 0.27354955673217773, 0.28795677423477173, 0.3023639917373657, 0.3167712390422821, 0.3311784863471985, 0.3455857038497925, 0.3599929213523865, 0.37440016865730286, 0.38880741596221924, 0.40321463346481323, 0.4176218509674072, 0.4320290982723236, 0.44643634557724, 0.460843563079834, 0.475250780582428, 0.48965802788734436, 0.5040652751922607, 0.5184724926948547, 0.5328797101974487, 0.5472869873046875, 0.5616942048072815, 0.5761014223098755, 0.5905086398124695, 0.6049158573150635, 0.6193231344223022, 0.6337303519248962, 0.6481375694274902, 0.662544846534729, 0.676952064037323, 0.691359281539917, 0.705766499042511, 0.720173716545105, 0.7345809936523438, 0.7489882111549377, 0.7633954286575317, 0.7778027057647705, 0.7922099232673645, 0.8066171407699585, 0.8210243582725525, 0.8354315757751465, 0.8498388528823853, 0.8642460703849792, 0.8786532878875732, 0.893060564994812, 0.907467782497406, 0.921875]}, "parameters/model.layers.1.post_attention_layernorm.weight": {"_type": "histogram", "values": [15.0, 0.0, 1.0, 0.0, 1.0, 4.0, 3.0, 0.0, 3.0, 1.0, 2.0, 2.0, 3.0, 0.0, 1.0, 2.0, 2.0, 3.0, 4.0, 1.0, 2.0, 3.0, 4.0, 14.0, 13.0, 19.0, 26.0, 42.0, 25.0, 25.0, 23.0, 18.0, 22.0, 17.0, 24.0, 20.0, 31.0, 21.0, 34.0, 75.0, 127.0, 433.0, 912.0, 1862.0, 218.0, 15.0, 3.0, 2.0, 6.0, 2.0, 2.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.000263214111328125, 0.004074394702911377, 0.008412003517150879, 0.01274961233139038, 0.017087221145629883, 0.021424829959869385, 0.025762438774108887, 0.03010004758834839, 0.03443765640258789, 0.03877526521682739, 0.043112874031066895, 0.047450482845306396, 0.0517880916595459, 0.0561257004737854, 0.0604633092880249, 0.0648009181022644, 0.0691385269165039, 0.07347613573074341, 0.07781374454498291, 0.08215135335922241, 0.08648896217346191, 0.09082657098770142, 0.09516417980194092, 0.09950178861618042, 0.10383939743041992, 0.10817700624465942, 0.11251461505889893, 0.11685222387313843, 0.12118983268737793, 0.12552744150161743, 0.12986505031585693, 0.13420265913009644, 0.13854026794433594, 0.14287787675857544, 0.14721548557281494, 0.15155309438705444, 0.15589070320129395, 0.16022831201553345, 0.16456592082977295, 0.16890352964401245, 0.17324113845825195, 0.17757874727249146, 0.18191635608673096, 0.18625396490097046, 0.19059157371520996, 0.19492918252944946, 0.19926679134368896, 0.20360440015792847, 0.20794200897216797, 0.21227961778640747, 0.21661722660064697, 0.22095483541488647, 0.22529244422912598, 0.22963005304336548, 0.23396766185760498, 0.23830527067184448, 0.24264287948608398, 0.2469804883003235, 0.251318097114563, 0.2556557059288025, 0.259993314743042, 0.2643309235572815, 0.268668532371521, 0.2730061411857605, 0.27734375]}, "parameters/model.layers.2.self_attn.q_proj.weight": {"_type": "histogram", "values": [4.0, 0.0, 3.0, 0.0, 3.0, 5.0, 15.0, 14.0, 31.0, 24.0, 51.0, 60.0, 79.0, 134.0, 146.0, 208.0, 372.0, 483.0, 779.0, 1290.0, 2573.0, 6069.0, 12286.0, 27782.0, 66085.0, 131134.0, 265090.0, 490443.0, 929845.0, 1622084.0, 2686614.0, 3573550.0, 2929002.0, 1826777.0, 1041089.0, 575979.0, 292749.0, 156261.0, 74671.0, 34157.0, 15016.0, 6821.0, 3288.0, 1540.0, 911.0, 536.0, 355.0, 229.0, 160.0, 92.0, 113.0, 57.0, 52.0, 33.0, 24.0, 17.0, 8.0, 8.0, 8.0, 0.0, 2.0, 0.0, 2.0, 3.0], "bins": [-0.234375, -0.226959228515625, -0.21954345703125, -0.212127685546875, -0.2047119140625, -0.197296142578125, -0.18988037109375, -0.182464599609375, -0.175048828125, -0.167633056640625, -0.16021728515625, -0.152801513671875, -0.1453857421875, -0.137969970703125, -0.13055419921875, -0.123138427734375, -0.11572265625, -0.108306884765625, -0.10089111328125, -0.093475341796875, -0.0860595703125, -0.078643798828125, -0.07122802734375, -0.063812255859375, -0.056396484375, -0.048980712890625, -0.04156494140625, -0.034149169921875, -0.0267333984375, -0.019317626953125, -0.01190185546875, -0.004486083984375, 0.0029296875, 0.010345458984375, 0.01776123046875, 0.025177001953125, 0.0325927734375, 0.040008544921875, 0.04742431640625, 0.054840087890625, 0.062255859375, 0.069671630859375, 0.07708740234375, 0.084503173828125, 0.0919189453125, 0.099334716796875, 0.10675048828125, 0.114166259765625, 0.12158203125, 0.128997802734375, 0.13641357421875, 0.143829345703125, 0.1512451171875, 0.158660888671875, 0.16607666015625, 0.173492431640625, 0.180908203125, 0.188323974609375, 0.19573974609375, 0.203155517578125, 0.2105712890625, 0.217987060546875, 0.22540283203125, 0.232818603515625, 0.240234375]}, "parameters/model.layers.2.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 2.0, 1.0, 4.0, 7.0, 12.0, 16.0, 19.0, 35.0, 41.0, 51.0, 52.0, 80.0, 144.0, 238.0, 349.0, 625.0, 1173.0, 2382.0, 4220.0, 8748.0, 15095.0, 27479.0, 49521.0, 86603.0, 141130.0, 236603.0, 401421.0, 659103.0, 855096.0, 682674.0, 419960.0, 250489.0, 145869.0, 89633.0, 51320.0, 29135.0, 15807.0, 9136.0, 4711.0, 2266.0, 1334.0, 579.0, 403.0, 223.0, 158.0, 105.0, 69.0, 63.0, 27.0, 33.0, 19.0, 16.0, 6.0, 5.0, 5.0, 2.0, 3.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.306640625, -0.296600341796875, -0.28656005859375, -0.276519775390625, -0.2664794921875, -0.256439208984375, -0.24639892578125, -0.236358642578125, -0.226318359375, -0.216278076171875, -0.20623779296875, -0.196197509765625, -0.1861572265625, -0.176116943359375, -0.16607666015625, -0.156036376953125, -0.14599609375, -0.135955810546875, -0.12591552734375, -0.115875244140625, -0.1058349609375, -0.095794677734375, -0.08575439453125, -0.075714111328125, -0.065673828125, -0.055633544921875, -0.04559326171875, -0.035552978515625, -0.0255126953125, -0.015472412109375, -0.00543212890625, 0.004608154296875, 0.0146484375, 0.024688720703125, 0.03472900390625, 0.044769287109375, 0.0548095703125, 0.064849853515625, 0.07489013671875, 0.084930419921875, 0.094970703125, 0.105010986328125, 0.11505126953125, 0.125091552734375, 0.1351318359375, 0.145172119140625, 0.15521240234375, 0.165252685546875, 0.17529296875, 0.185333251953125, 0.19537353515625, 0.205413818359375, 0.2154541015625, 0.225494384765625, 0.23553466796875, 0.245574951171875, 0.255615234375, 0.265655517578125, 0.27569580078125, 0.285736083984375, 0.2957763671875, 0.305816650390625, 0.31585693359375, 0.325897216796875, 0.3359375]}, "parameters/model.layers.2.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 1.0, 3.0, 3.0, 1.0, 6.0, 15.0, 26.0, 70.0, 121.0, 252.0, 373.0, 888.0, 1647.0, 2796.0, 4911.0, 8484.0, 14184.0, 21427.0, 35723.0, 55680.0, 78634.0, 116185.0, 160185.0, 217067.0, 286574.0, 347685.0, 412458.0, 447013.0, 433589.0, 381085.0, 317921.0, 248798.0, 191208.0, 131782.0, 97268.0, 66655.0, 42864.0, 28081.0, 17570.0, 10887.0, 6341.0, 3554.0, 1912.0, 1175.0, 605.0, 302.0, 128.0, 76.0, 35.0, 23.0, 13.0, 4.0, 6.0, 1.0, 3.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.046630859375, -0.045063018798828125, -0.04349517822265625, -0.041927337646484375, -0.0403594970703125, -0.038791656494140625, -0.03722381591796875, -0.035655975341796875, -0.034088134765625, -0.032520294189453125, -0.03095245361328125, -0.029384613037109375, -0.0278167724609375, -0.026248931884765625, -0.02468109130859375, -0.023113250732421875, -0.02154541015625, -0.019977569580078125, -0.01840972900390625, -0.016841888427734375, -0.0152740478515625, -0.013706207275390625, -0.01213836669921875, -0.010570526123046875, -0.009002685546875, -0.007434844970703125, -0.00586700439453125, -0.004299163818359375, -0.0027313232421875, -0.001163482666015625, 0.00040435791015625, 0.001972198486328125, 0.0035400390625, 0.005107879638671875, 0.00667572021484375, 0.008243560791015625, 0.0098114013671875, 0.011379241943359375, 0.01294708251953125, 0.014514923095703125, 0.016082763671875, 0.017650604248046875, 0.01921844482421875, 0.020786285400390625, 0.0223541259765625, 0.023921966552734375, 0.02548980712890625, 0.027057647705078125, 0.02862548828125, 0.030193328857421875, 0.03176116943359375, 0.033329010009765625, 0.0348968505859375, 0.036464691162109375, 0.03803253173828125, 0.039600372314453125, 0.041168212890625, 0.042736053466796875, 0.04430389404296875, 0.045871734619140625, 0.0474395751953125, 0.049007415771484375, 0.05057525634765625, 0.052143096923828125, 0.0537109375]}, "parameters/model.layers.2.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 2.0, 2.0, 2.0, 4.0, 6.0, 3.0, 6.0, 7.0, 8.0, 22.0, 28.0, 32.0, 52.0, 77.0, 122.0, 208.0, 430.0, 971.0, 3317.0, 25851.0, 345463.0, 2723727.0, 7990577.0, 4773346.0, 833163.0, 70848.0, 6160.0, 1489.0, 600.0, 278.0, 144.0, 71.0, 46.0, 37.0, 34.0, 17.0, 9.0, 10.0, 12.0, 8.0, 4.0, 3.0, 5.0, 2.0, 2.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.298828125, -0.289398193359375, -0.27996826171875, -0.270538330078125, -0.2611083984375, -0.251678466796875, -0.24224853515625, -0.232818603515625, -0.223388671875, -0.213958740234375, -0.20452880859375, -0.195098876953125, -0.1856689453125, -0.176239013671875, -0.16680908203125, -0.157379150390625, -0.14794921875, -0.138519287109375, -0.12908935546875, -0.119659423828125, -0.1102294921875, -0.100799560546875, -0.09136962890625, -0.081939697265625, -0.072509765625, -0.063079833984375, -0.05364990234375, -0.044219970703125, -0.0347900390625, -0.025360107421875, -0.01593017578125, -0.006500244140625, 0.0029296875, 0.012359619140625, 0.02178955078125, 0.031219482421875, 0.0406494140625, 0.050079345703125, 0.05950927734375, 0.068939208984375, 0.078369140625, 0.087799072265625, 0.09722900390625, 0.106658935546875, 0.1160888671875, 0.125518798828125, 0.13494873046875, 0.144378662109375, 0.15380859375, 0.163238525390625, 0.17266845703125, 0.182098388671875, 0.1915283203125, 0.200958251953125, 0.21038818359375, 0.219818115234375, 0.229248046875, 0.238677978515625, 0.24810791015625, 0.257537841796875, 0.2669677734375, 0.276397705078125, 0.28582763671875, 0.295257568359375, 0.3046875]}, "parameters/model.layers.2.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 6.0, 1.0, 3.0, 11.0, 4.0, 10.0, 4.0, 14.0, 12.0, 18.0, 28.0, 43.0, 53.0, 69.0, 120.0, 198.0, 365.0, 677.0, 1391.0, 3214.0, 12580.0, 86934.0, 619741.0, 3543388.0, 12760880.0, 16777216.0, 16777216.0, 16777216.0, 9762481.0, 2499249.0, 387780.0, 51357.0, 8420.0, 2490.0, 1120.0, 509.0, 340.0, 197.0, 99.0, 79.0, 59.0, 26.0, 29.0, 25.0, 12.0, 11.0, 6.0, 3.0, 9.0, 2.0, 3.0, 4.0, 3.0, 0.0, 0.0, 3.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.251953125, -0.243377685546875, -0.23480224609375, -0.226226806640625, -0.2176513671875, -0.209075927734375, -0.20050048828125, -0.191925048828125, -0.183349609375, -0.174774169921875, -0.16619873046875, -0.157623291015625, -0.1490478515625, -0.140472412109375, -0.13189697265625, -0.123321533203125, -0.11474609375, -0.106170654296875, -0.09759521484375, -0.089019775390625, -0.0804443359375, -0.071868896484375, -0.06329345703125, -0.054718017578125, -0.046142578125, -0.037567138671875, -0.02899169921875, -0.020416259765625, -0.0118408203125, -0.003265380859375, 0.00531005859375, 0.013885498046875, 0.0224609375, 0.031036376953125, 0.03961181640625, 0.048187255859375, 0.0567626953125, 0.065338134765625, 0.07391357421875, 0.082489013671875, 0.091064453125, 0.099639892578125, 0.10821533203125, 0.116790771484375, 0.1253662109375, 0.133941650390625, 0.14251708984375, 0.151092529296875, 0.15966796875, 0.168243408203125, 0.17681884765625, 0.185394287109375, 0.1939697265625, 0.202545166015625, 0.21112060546875, 0.219696044921875, 0.228271484375, 0.236846923828125, 0.24542236328125, 0.253997802734375, 0.2625732421875, 0.271148681640625, 0.27972412109375, 0.288299560546875, 0.296875]}, "parameters/model.layers.2.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 3.0, 5.0, 1.0, 2.0, 5.0, 5.0, 6.0, 9.0, 29.0, 28.0, 49.0, 92.0, 160.0, 340.0, 806.0, 9206.0, 441864.0, 7390728.0, 16777216.0, 16777216.0, 3237766.0, 122430.0, 2811.0, 576.0, 276.0, 133.0, 72.0, 39.0, 26.0, 13.0, 2.0, 9.0, 6.0, 4.0, 3.0, 2.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.5390625, -0.524810791015625, -0.51055908203125, -0.496307373046875, -0.4820556640625, -0.467803955078125, -0.45355224609375, -0.439300537109375, -0.425048828125, -0.410797119140625, -0.39654541015625, -0.382293701171875, -0.3680419921875, -0.353790283203125, -0.33953857421875, -0.325286865234375, -0.31103515625, -0.296783447265625, -0.28253173828125, -0.268280029296875, -0.2540283203125, -0.239776611328125, -0.22552490234375, -0.211273193359375, -0.197021484375, -0.182769775390625, -0.16851806640625, -0.154266357421875, -0.1400146484375, -0.125762939453125, -0.11151123046875, -0.097259521484375, -0.0830078125, -0.068756103515625, -0.05450439453125, -0.040252685546875, -0.0260009765625, -0.011749267578125, 0.00250244140625, 0.016754150390625, 0.031005859375, 0.045257568359375, 0.05950927734375, 0.073760986328125, 0.0880126953125, 0.102264404296875, 0.11651611328125, 0.130767822265625, 0.14501953125, 0.159271240234375, 0.17352294921875, 0.187774658203125, 0.2020263671875, 0.216278076171875, 0.23052978515625, 0.244781494140625, 0.259033203125, 0.273284912109375, 0.28753662109375, 0.301788330078125, 0.3160400390625, 0.330291748046875, 0.34454345703125, 0.358795166015625, 0.373046875]}, "parameters/model.layers.2.input_layernorm.weight": {"_type": "histogram", "values": [7.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 1.0, 2.0, 1.0, 2.0, 5.0, 9.0, 5.0, 21.0, 123.0, 489.0, 695.0, 642.0, 512.0, 349.0, 245.0, 159.0, 93.0, 86.0, 83.0, 68.0, 47.0, 41.0, 34.0, 27.0, 18.0, 11.0, 18.0, 13.0, 13.0, 9.0, 12.0, 23.0, 10.0, 13.0, 18.0, 18.0, 12.0, 14.0, 19.0, 17.0, 18.0, 21.0, 20.0, 13.0, 8.0, 3.0, 7.0, 9.0, 2.0, 2.0, 1.0, 3.0, 0.0, 2.0], "bins": [-0.000637054443359375, 0.012617528438568115, 0.025872111320495605, 0.039126694202423096, 0.052381277084350586, 0.06563585996627808, 0.07889044284820557, 0.09214502573013306, 0.10539960861206055, 0.11865419149398804, 0.13190877437591553, 0.14516335725784302, 0.1584179401397705, 0.171672523021698, 0.1849271059036255, 0.19818168878555298, 0.21143627166748047, 0.22469085454940796, 0.23794543743133545, 0.25120002031326294, 0.26445460319519043, 0.2777091860771179, 0.2909637689590454, 0.3042183518409729, 0.3174729347229004, 0.3307275176048279, 0.34398210048675537, 0.35723668336868286, 0.37049126625061035, 0.38374584913253784, 0.39700043201446533, 0.4102550148963928, 0.4235095977783203, 0.4367641806602478, 0.4500187635421753, 0.4632733464241028, 0.4765279293060303, 0.48978251218795776, 0.5030370950698853, 0.5162916779518127, 0.5295462608337402, 0.5428008437156677, 0.5560554265975952, 0.5693100094795227, 0.5825645923614502, 0.5958191752433777, 0.6090737581253052, 0.6223283410072327, 0.6355829238891602, 0.6488375067710876, 0.6620920896530151, 0.6753466725349426, 0.6886012554168701, 0.7018558382987976, 0.7151104211807251, 0.7283650040626526, 0.7416195869445801, 0.7548741698265076, 0.7681287527084351, 0.7813833355903625, 0.79463791847229, 0.8078925013542175, 0.821147084236145, 0.8344016671180725, 0.84765625]}, "parameters/model.layers.2.post_attention_layernorm.weight": {"_type": "histogram", "values": [17.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 2.0, 1.0, 1.0, 1.0, 0.0, 2.0, 0.0, 4.0, 3.0, 1.0, 7.0, 3.0, 4.0, 2.0, 12.0, 19.0, 23.0, 16.0, 19.0, 13.0, 17.0, 13.0, 12.0, 11.0, 13.0, 10.0, 11.0, 9.0, 12.0, 8.0, 11.0, 11.0, 8.0, 10.0, 13.0, 16.0, 18.0, 19.0, 27.0, 26.0, 31.0, 39.0, 68.0, 112.0, 194.0, 432.0, 868.0, 1416.0, 483.0, 26.0], "bins": [-0.000713348388671875, 0.003081977367401123, 0.006877303123474121, 0.01067262887954712, 0.014467954635620117, 0.018263280391693115, 0.022058606147766113, 0.02585393190383911, 0.02964925765991211, 0.03344458341598511, 0.037239909172058105, 0.041035234928131104, 0.0448305606842041, 0.0486258864402771, 0.0524212121963501, 0.056216537952423096, 0.060011863708496094, 0.06380718946456909, 0.06760251522064209, 0.07139784097671509, 0.07519316673278809, 0.07898849248886108, 0.08278381824493408, 0.08657914400100708, 0.09037446975708008, 0.09416979551315308, 0.09796512126922607, 0.10176044702529907, 0.10555577278137207, 0.10935109853744507, 0.11314642429351807, 0.11694175004959106, 0.12073707580566406, 0.12453240156173706, 0.12832772731781006, 0.13212305307388306, 0.13591837882995605, 0.13971370458602905, 0.14350903034210205, 0.14730435609817505, 0.15109968185424805, 0.15489500761032104, 0.15869033336639404, 0.16248565912246704, 0.16628098487854004, 0.17007631063461304, 0.17387163639068604, 0.17766696214675903, 0.18146228790283203, 0.18525761365890503, 0.18905293941497803, 0.19284826517105103, 0.19664359092712402, 0.20043891668319702, 0.20423424243927002, 0.20802956819534302, 0.21182489395141602, 0.215620219707489, 0.219415545463562, 0.223210871219635, 0.227006196975708, 0.230801522731781, 0.234596848487854, 0.238392174243927, 0.2421875]}, "parameters/model.layers.3.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 2.0, 1.0, 1.0, 5.0, 2.0, 3.0, 6.0, 10.0, 10.0, 14.0, 28.0, 50.0, 59.0, 147.0, 260.0, 490.0, 916.0, 1588.0, 3582.0, 7044.0, 14751.0, 29817.0, 63734.0, 119211.0, 218924.0, 398241.0, 723500.0, 1132900.0, 1811138.0, 2529729.0, 2895886.0, 2457049.0, 1762624.0, 1092887.0, 694723.0, 379644.0, 208664.0, 113601.0, 60807.0, 28102.0, 13794.0, 6525.0, 3269.0, 1545.0, 859.0, 457.0, 284.0, 128.0, 72.0, 49.0, 25.0, 20.0, 14.0, 7.0, 5.0, 2.0, 2.0, 1.0, 2.0, 0.0, 0.0, 2.0], "bins": [-0.2080078125, -0.20159912109375, -0.1951904296875, -0.18878173828125, -0.182373046875, -0.17596435546875, -0.1695556640625, -0.16314697265625, -0.15673828125, -0.15032958984375, -0.1439208984375, -0.13751220703125, -0.131103515625, -0.12469482421875, -0.1182861328125, -0.11187744140625, -0.10546875, -0.09906005859375, -0.0926513671875, -0.08624267578125, -0.079833984375, -0.07342529296875, -0.0670166015625, -0.06060791015625, -0.05419921875, -0.04779052734375, -0.0413818359375, -0.03497314453125, -0.028564453125, -0.02215576171875, -0.0157470703125, -0.00933837890625, -0.0029296875, 0.00347900390625, 0.0098876953125, 0.01629638671875, 0.022705078125, 0.02911376953125, 0.0355224609375, 0.04193115234375, 0.04833984375, 0.05474853515625, 0.0611572265625, 0.06756591796875, 0.073974609375, 0.08038330078125, 0.0867919921875, 0.09320068359375, 0.099609375, 0.10601806640625, 0.1124267578125, 0.11883544921875, 0.125244140625, 0.13165283203125, 0.1380615234375, 0.14447021484375, 0.15087890625, 0.15728759765625, 0.1636962890625, 0.17010498046875, 0.176513671875, 0.18292236328125, 0.1893310546875, 0.19573974609375, 0.2021484375]}, "parameters/model.layers.3.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 5.0, 16.0, 24.0, 33.0, 88.0, 133.0, 222.0, 403.0, 741.0, 1426.0, 2477.0, 4330.0, 7889.0, 12989.0, 21440.0, 38495.0, 65444.0, 106395.0, 160375.0, 264862.0, 376762.0, 532856.0, 654469.0, 607111.0, 459318.0, 323790.0, 208054.0, 135784.0, 86267.0, 51563.0, 28341.0, 18278.0, 10093.0, 6284.0, 3254.0, 1905.0, 1088.0, 598.0, 281.0, 186.0, 113.0, 61.0, 21.0, 15.0, 8.0, 3.0, 0.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.26953125, -0.26104736328125, -0.2525634765625, -0.24407958984375, -0.235595703125, -0.22711181640625, -0.2186279296875, -0.21014404296875, -0.20166015625, -0.19317626953125, -0.1846923828125, -0.17620849609375, -0.167724609375, -0.15924072265625, -0.1507568359375, -0.14227294921875, -0.1337890625, -0.12530517578125, -0.1168212890625, -0.10833740234375, -0.099853515625, -0.09136962890625, -0.0828857421875, -0.07440185546875, -0.06591796875, -0.05743408203125, -0.0489501953125, -0.04046630859375, -0.031982421875, -0.02349853515625, -0.0150146484375, -0.00653076171875, 0.001953125, 0.01043701171875, 0.0189208984375, 0.02740478515625, 0.035888671875, 0.04437255859375, 0.0528564453125, 0.06134033203125, 0.06982421875, 0.07830810546875, 0.0867919921875, 0.09527587890625, 0.103759765625, 0.11224365234375, 0.1207275390625, 0.12921142578125, 0.1376953125, 0.14617919921875, 0.1546630859375, 0.16314697265625, 0.171630859375, 0.18011474609375, 0.1885986328125, 0.19708251953125, 0.20556640625, 0.21405029296875, 0.2225341796875, 0.23101806640625, 0.239501953125, 0.24798583984375, 0.2564697265625, 0.26495361328125, 0.2734375]}, "parameters/model.layers.3.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 4.0, 3.0, 14.0, 22.0, 43.0, 124.0, 245.0, 622.0, 1866.0, 5063.0, 12636.0, 32585.0, 75908.0, 166593.0, 298651.0, 489342.0, 664521.0, 746566.0, 648734.0, 477864.0, 289968.0, 159885.0, 71500.0, 31736.0, 12538.0, 4651.0, 1626.0, 586.0, 247.0, 74.0, 35.0, 22.0, 3.0, 9.0, 5.0, 1.0, 3.0, 1.0, 2.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.08642578125, -0.083160400390625, -0.07989501953125, -0.076629638671875, -0.0733642578125, -0.070098876953125, -0.06683349609375, -0.063568115234375, -0.060302734375, -0.057037353515625, -0.05377197265625, -0.050506591796875, -0.0472412109375, -0.043975830078125, -0.04071044921875, -0.037445068359375, -0.0341796875, -0.030914306640625, -0.02764892578125, -0.024383544921875, -0.0211181640625, -0.017852783203125, -0.01458740234375, -0.011322021484375, -0.008056640625, -0.004791259765625, -0.00152587890625, 0.001739501953125, 0.0050048828125, 0.008270263671875, 0.01153564453125, 0.014801025390625, 0.01806640625, 0.021331787109375, 0.02459716796875, 0.027862548828125, 0.0311279296875, 0.034393310546875, 0.03765869140625, 0.040924072265625, 0.044189453125, 0.047454833984375, 0.05072021484375, 0.053985595703125, 0.0572509765625, 0.060516357421875, 0.06378173828125, 0.067047119140625, 0.0703125, 0.073577880859375, 0.07684326171875, 0.080108642578125, 0.0833740234375, 0.086639404296875, 0.08990478515625, 0.093170166015625, 0.096435546875, 0.099700927734375, 0.10296630859375, 0.106231689453125, 0.1094970703125, 0.112762451171875, 0.11602783203125, 0.119293212890625, 0.12255859375]}, "parameters/model.layers.3.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 3.0, 2.0, 2.0, 1.0, 5.0, 7.0, 9.0, 14.0, 9.0, 17.0, 11.0, 23.0, 38.0, 51.0, 80.0, 142.0, 249.0, 618.0, 3124.0, 39548.0, 468607.0, 2817701.0, 6558337.0, 5199287.0, 1490942.0, 182691.0, 13249.0, 1434.0, 478.0, 182.0, 104.0, 73.0, 44.0, 23.0, 28.0, 26.0, 13.0, 6.0, 4.0, 8.0, 5.0, 2.0, 6.0, 3.0, 1.0, 2.0, 1.0, 0.0, 1.0, 0.0, 2.0], "bins": [-0.33203125, -0.32275390625, -0.3134765625, -0.30419921875, -0.294921875, -0.28564453125, -0.2763671875, -0.26708984375, -0.2578125, -0.24853515625, -0.2392578125, -0.22998046875, -0.220703125, -0.21142578125, -0.2021484375, -0.19287109375, -0.18359375, -0.17431640625, -0.1650390625, -0.15576171875, -0.146484375, -0.13720703125, -0.1279296875, -0.11865234375, -0.109375, -0.10009765625, -0.0908203125, -0.08154296875, -0.072265625, -0.06298828125, -0.0537109375, -0.04443359375, -0.03515625, -0.02587890625, -0.0166015625, -0.00732421875, 0.001953125, 0.01123046875, 0.0205078125, 0.02978515625, 0.0390625, 0.04833984375, 0.0576171875, 0.06689453125, 0.076171875, 0.08544921875, 0.0947265625, 0.10400390625, 0.11328125, 0.12255859375, 0.1318359375, 0.14111328125, 0.150390625, 0.15966796875, 0.1689453125, 0.17822265625, 0.1875, 0.19677734375, 0.2060546875, 0.21533203125, 0.224609375, 0.23388671875, 0.2431640625, 0.25244140625, 0.26171875]}, "parameters/model.layers.3.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 2.0, 3.0, 2.0, 1.0, 8.0, 7.0, 12.0, 21.0, 27.0, 30.0, 59.0, 79.0, 148.0, 245.0, 531.0, 1532.0, 6644.0, 47529.0, 562646.0, 5996211.0, 16777216.0, 16777216.0, 16777216.0, 4927244.0, 433269.0, 37973.0, 5435.0, 1474.0, 595.0, 254.0, 137.0, 74.0, 49.0, 33.0, 19.0, 16.0, 14.0, 5.0, 6.0, 2.0, 3.0, 3.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.34765625, -0.334991455078125, -0.32232666015625, -0.309661865234375, -0.2969970703125, -0.284332275390625, -0.27166748046875, -0.259002685546875, -0.246337890625, -0.233673095703125, -0.22100830078125, -0.208343505859375, -0.1956787109375, -0.183013916015625, -0.17034912109375, -0.157684326171875, -0.14501953125, -0.132354736328125, -0.11968994140625, -0.107025146484375, -0.0943603515625, -0.081695556640625, -0.06903076171875, -0.056365966796875, -0.043701171875, -0.031036376953125, -0.01837158203125, -0.005706787109375, 0.0069580078125, 0.019622802734375, 0.03228759765625, 0.044952392578125, 0.0576171875, 0.070281982421875, 0.08294677734375, 0.095611572265625, 0.1082763671875, 0.120941162109375, 0.13360595703125, 0.146270751953125, 0.158935546875, 0.171600341796875, 0.18426513671875, 0.196929931640625, 0.2095947265625, 0.222259521484375, 0.23492431640625, 0.247589111328125, 0.26025390625, 0.272918701171875, 0.28558349609375, 0.298248291015625, 0.3109130859375, 0.323577880859375, 0.33624267578125, 0.348907470703125, 0.361572265625, 0.374237060546875, 0.38690185546875, 0.399566650390625, 0.4122314453125, 0.424896240234375, 0.43756103515625, 0.450225830078125, 0.462890625]}, "parameters/model.layers.3.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 1.0, 6.0, 6.0, 8.0, 8.0, 14.0, 34.0, 54.0, 113.0, 259.0, 747.0, 12260.0, 494757.0, 6717228.0, 16777216.0, 16777216.0, 4675165.0, 270184.0, 6250.0, 581.0, 231.0, 101.0, 56.0, 18.0, 15.0, 8.0, 6.0, 3.0, 3.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.484375, -0.47125244140625, -0.4581298828125, -0.44500732421875, -0.431884765625, -0.41876220703125, -0.4056396484375, -0.39251708984375, -0.37939453125, -0.36627197265625, -0.3531494140625, -0.34002685546875, -0.326904296875, -0.31378173828125, -0.3006591796875, -0.28753662109375, -0.2744140625, -0.26129150390625, -0.2481689453125, -0.23504638671875, -0.221923828125, -0.20880126953125, -0.1956787109375, -0.18255615234375, -0.16943359375, -0.15631103515625, -0.1431884765625, -0.13006591796875, -0.116943359375, -0.10382080078125, -0.0906982421875, -0.07757568359375, -0.064453125, -0.05133056640625, -0.0382080078125, -0.02508544921875, -0.011962890625, 0.00115966796875, 0.0142822265625, 0.02740478515625, 0.04052734375, 0.05364990234375, 0.0667724609375, 0.07989501953125, 0.093017578125, 0.10614013671875, 0.1192626953125, 0.13238525390625, 0.1455078125, 0.15863037109375, 0.1717529296875, 0.18487548828125, 0.197998046875, 0.21112060546875, 0.2242431640625, 0.23736572265625, 0.25048828125, 0.26361083984375, 0.2767333984375, 0.28985595703125, 0.302978515625, 0.31610107421875, 0.3292236328125, 0.34234619140625, 0.35546875]}, "parameters/model.layers.3.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 2.0, 4.0, 0.0, 3.0, 6.0, 2.0, 10.0, 23.0, 112.0, 458.0, 587.0, 701.0, 557.0, 577.0, 331.0, 210.0, 137.0, 127.0, 80.0, 39.0, 29.0, 19.0, 16.0, 14.0, 10.0, 6.0, 5.0, 1.0, 3.0, 4.0, 0.0, 0.0, 5.0, 5.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0], "bins": [-1.800060272216797e-05, 0.010175151750445366, 0.0203683041036129, 0.030561456456780434, 0.04075460880994797, 0.05094776302576065, 0.061140913516283035, 0.07133406400680542, 0.0815272182226181, 0.09172037243843079, 0.10191352665424347, 0.11210667341947556, 0.12229982763528824, 0.13249297440052032, 0.142686128616333, 0.1528792828321457, 0.16307243704795837, 0.17326559126377106, 0.18345874547958374, 0.19365189969539642, 0.2038450539112091, 0.2140381932258606, 0.22423134744167328, 0.23442450165748596, 0.24461765587329865, 0.25481081008911133, 0.2650039494037628, 0.2751971185207367, 0.2853902578353882, 0.29558342695236206, 0.30577656626701355, 0.3159697353839874, 0.3261628746986389, 0.3363560140132904, 0.3465491831302643, 0.35674232244491577, 0.36693549156188965, 0.37712863087654114, 0.387321799993515, 0.3975149393081665, 0.4077081084251404, 0.41790124773979187, 0.42809438705444336, 0.43828755617141724, 0.4484806954860687, 0.4586738646030426, 0.4688670039176941, 0.47906017303466797, 0.48925331234931946, 0.49944645166397095, 0.5096396207809448, 0.5198327898979187, 0.5300258994102478, 0.5402190685272217, 0.5504122376441956, 0.5606054067611694, 0.5707985162734985, 0.5809916853904724, 0.5911848545074463, 0.6013779640197754, 0.6115711331367493, 0.6217643022537231, 0.631957471370697, 0.6421505808830261, 0.65234375]}, "parameters/model.layers.3.post_attention_layernorm.weight": {"_type": "histogram", "values": [10.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 1.0, 2.0, 3.0, 0.0, 1.0, 1.0, 0.0, 2.0, 3.0, 4.0, 0.0, 3.0, 5.0, 2.0, 4.0, 3.0, 9.0, 13.0, 16.0, 10.0, 14.0, 21.0, 18.0, 26.0, 15.0, 13.0, 19.0, 10.0, 17.0, 10.0, 12.0, 16.0, 12.0, 19.0, 21.0, 25.0, 28.0, 40.0, 62.0, 69.0, 138.0, 219.0, 457.0, 1613.0, 925.0, 166.0, 13.0, 2.0], "bins": [-0.000385284423828125, 0.0039237141609191895, 0.008232712745666504, 0.012541711330413818, 0.016850709915161133, 0.021159708499908447, 0.02546870708465576, 0.029777705669403076, 0.03408670425415039, 0.038395702838897705, 0.04270470142364502, 0.047013700008392334, 0.05132269859313965, 0.05563169717788696, 0.05994069576263428, 0.06424969434738159, 0.0685586929321289, 0.07286769151687622, 0.07717669010162354, 0.08148568868637085, 0.08579468727111816, 0.09010368585586548, 0.09441268444061279, 0.09872168302536011, 0.10303068161010742, 0.10733968019485474, 0.11164867877960205, 0.11595767736434937, 0.12026667594909668, 0.124575674533844, 0.1288846731185913, 0.13319367170333862, 0.13750267028808594, 0.14181166887283325, 0.14612066745758057, 0.15042966604232788, 0.1547386646270752, 0.1590476632118225, 0.16335666179656982, 0.16766566038131714, 0.17197465896606445, 0.17628365755081177, 0.18059265613555908, 0.1849016547203064, 0.1892106533050537, 0.19351965188980103, 0.19782865047454834, 0.20213764905929565, 0.20644664764404297, 0.21075564622879028, 0.2150646448135376, 0.2193736433982849, 0.22368264198303223, 0.22799164056777954, 0.23230063915252686, 0.23660963773727417, 0.24091863632202148, 0.2452276349067688, 0.2495366334915161, 0.2538456320762634, 0.25815463066101074, 0.26246362924575806, 0.26677262783050537, 0.2710816264152527, 0.275390625]}, "parameters/model.layers.4.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 3.0, 2.0, 2.0, 8.0, 9.0, 11.0, 21.0, 15.0, 23.0, 47.0, 46.0, 56.0, 81.0, 106.0, 138.0, 215.0, 377.0, 857.0, 2089.0, 5303.0, 14457.0, 44971.0, 125362.0, 322134.0, 820782.0, 1796966.0, 3446269.0, 4438530.0, 3087563.0, 1543678.0, 688347.0, 275699.0, 105919.0, 36425.0, 12387.0, 4589.0, 1795.0, 783.0, 368.0, 206.0, 141.0, 99.0, 78.0, 64.0, 53.0, 34.0, 23.0, 20.0, 18.0, 14.0, 7.0, 8.0, 5.0, 3.0, 3.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.306640625, -0.296875, -0.287109375, -0.27734375, -0.267578125, -0.2578125, -0.248046875, -0.23828125, -0.228515625, -0.21875, -0.208984375, -0.19921875, -0.189453125, -0.1796875, -0.169921875, -0.16015625, -0.150390625, -0.140625, -0.130859375, -0.12109375, -0.111328125, -0.1015625, -0.091796875, -0.08203125, -0.072265625, -0.0625, -0.052734375, -0.04296875, -0.033203125, -0.0234375, -0.013671875, -0.00390625, 0.005859375, 0.015625, 0.025390625, 0.03515625, 0.044921875, 0.0546875, 0.064453125, 0.07421875, 0.083984375, 0.09375, 0.103515625, 0.11328125, 0.123046875, 0.1328125, 0.142578125, 0.15234375, 0.162109375, 0.171875, 0.181640625, 0.19140625, 0.201171875, 0.2109375, 0.220703125, 0.23046875, 0.240234375, 0.25, 0.259765625, 0.26953125, 0.279296875, 0.2890625, 0.298828125, 0.30859375, 0.318359375]}, "parameters/model.layers.4.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 3.0, 3.0, 3.0, 6.0, 10.0, 15.0, 13.0, 27.0, 29.0, 45.0, 37.0, 44.0, 62.0, 112.0, 124.0, 196.0, 397.0, 802.0, 1669.0, 3744.0, 6972.0, 15506.0, 29170.0, 59808.0, 110131.0, 186903.0, 327333.0, 501467.0, 701089.0, 751154.0, 580311.0, 389128.0, 234105.0, 138686.0, 76938.0, 39348.0, 19237.0, 10176.0, 4678.0, 2330.0, 1052.0, 537.0, 250.0, 173.0, 130.0, 90.0, 63.0, 43.0, 33.0, 32.0, 20.0, 19.0, 14.0, 8.0, 9.0, 7.0, 3.0, 2.0, 1.0, 3.0, 1.0], "bins": [-0.322265625, -0.312255859375, -0.30224609375, -0.292236328125, -0.2822265625, -0.272216796875, -0.26220703125, -0.252197265625, -0.2421875, -0.232177734375, -0.22216796875, -0.212158203125, -0.2021484375, -0.192138671875, -0.18212890625, -0.172119140625, -0.162109375, -0.152099609375, -0.14208984375, -0.132080078125, -0.1220703125, -0.112060546875, -0.10205078125, -0.092041015625, -0.08203125, -0.072021484375, -0.06201171875, -0.052001953125, -0.0419921875, -0.031982421875, -0.02197265625, -0.011962890625, -0.001953125, 0.008056640625, 0.01806640625, 0.028076171875, 0.0380859375, 0.048095703125, 0.05810546875, 0.068115234375, 0.078125, 0.088134765625, 0.09814453125, 0.108154296875, 0.1181640625, 0.128173828125, 0.13818359375, 0.148193359375, 0.158203125, 0.168212890625, 0.17822265625, 0.188232421875, 0.1982421875, 0.208251953125, 0.21826171875, 0.228271484375, 0.23828125, 0.248291015625, 0.25830078125, 0.268310546875, 0.2783203125, 0.288330078125, 0.29833984375, 0.308349609375, 0.318359375]}, "parameters/model.layers.4.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 3.0, 1.0, 4.0, 9.0, 23.0, 37.0, 82.0, 148.0, 312.0, 540.0, 1148.0, 2214.0, 4077.0, 7672.0, 12861.0, 20246.0, 34900.0, 56854.0, 88870.0, 130677.0, 190600.0, 262021.0, 345455.0, 413642.0, 467087.0, 473009.0, 431330.0, 358484.0, 282036.0, 207647.0, 146648.0, 96162.0, 62202.0, 40010.0, 24308.0, 14678.0, 8410.0, 4602.0, 2534.0, 1286.0, 738.0, 348.0, 197.0, 101.0, 40.0, 26.0, 8.0, 1.0, 7.0, 3.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0], "bins": [-0.072265625, -0.0700836181640625, -0.067901611328125, -0.0657196044921875, -0.06353759765625, -0.0613555908203125, -0.059173583984375, -0.0569915771484375, -0.0548095703125, -0.0526275634765625, -0.050445556640625, -0.0482635498046875, -0.04608154296875, -0.0438995361328125, -0.041717529296875, -0.0395355224609375, -0.037353515625, -0.0351715087890625, -0.032989501953125, -0.0308074951171875, -0.02862548828125, -0.0264434814453125, -0.024261474609375, -0.0220794677734375, -0.0198974609375, -0.0177154541015625, -0.015533447265625, -0.0133514404296875, -0.01116943359375, -0.0089874267578125, -0.006805419921875, -0.0046234130859375, -0.00244140625, -0.0002593994140625, 0.001922607421875, 0.0041046142578125, 0.00628662109375, 0.0084686279296875, 0.010650634765625, 0.0128326416015625, 0.0150146484375, 0.0171966552734375, 0.019378662109375, 0.0215606689453125, 0.02374267578125, 0.0259246826171875, 0.028106689453125, 0.0302886962890625, 0.032470703125, 0.0346527099609375, 0.036834716796875, 0.0390167236328125, 0.04119873046875, 0.0433807373046875, 0.045562744140625, 0.0477447509765625, 0.0499267578125, 0.0521087646484375, 0.054290771484375, 0.0564727783203125, 0.05865478515625, 0.0608367919921875, 0.063018798828125, 0.0652008056640625, 0.0673828125]}, "parameters/model.layers.4.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 3.0, 0.0, 4.0, 2.0, 5.0, 4.0, 5.0, 6.0, 17.0, 10.0, 33.0, 16.0, 45.0, 64.0, 90.0, 104.0, 198.0, 293.0, 554.0, 885.0, 2228.0, 12008.0, 123999.0, 911356.0, 3724793.0, 6551120.0, 4150937.0, 1113653.0, 163608.0, 16176.0, 2502.0, 991.0, 524.0, 337.0, 211.0, 118.0, 85.0, 63.0, 47.0, 24.0, 30.0, 21.0, 11.0, 7.0, 6.0, 2.0, 6.0, 4.0, 4.0, 1.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.3125, -0.3037109375, -0.294921875, -0.2861328125, -0.27734375, -0.2685546875, -0.259765625, -0.2509765625, -0.2421875, -0.2333984375, -0.224609375, -0.2158203125, -0.20703125, -0.1982421875, -0.189453125, -0.1806640625, -0.171875, -0.1630859375, -0.154296875, -0.1455078125, -0.13671875, -0.1279296875, -0.119140625, -0.1103515625, -0.1015625, -0.0927734375, -0.083984375, -0.0751953125, -0.06640625, -0.0576171875, -0.048828125, -0.0400390625, -0.03125, -0.0224609375, -0.013671875, -0.0048828125, 0.00390625, 0.0126953125, 0.021484375, 0.0302734375, 0.0390625, 0.0478515625, 0.056640625, 0.0654296875, 0.07421875, 0.0830078125, 0.091796875, 0.1005859375, 0.109375, 0.1181640625, 0.126953125, 0.1357421875, 0.14453125, 0.1533203125, 0.162109375, 0.1708984375, 0.1796875, 0.1884765625, 0.197265625, 0.2060546875, 0.21484375, 0.2236328125, 0.232421875, 0.2412109375, 0.25]}, "parameters/model.layers.4.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 4.0, 5.0, 3.0, 8.0, 17.0, 13.0, 25.0, 27.0, 46.0, 46.0, 101.0, 118.0, 195.0, 362.0, 719.0, 1891.0, 7590.0, 42508.0, 266657.0, 1706742.0, 8824668.0, 16777216.0, 16777216.0, 16777216.0, 8059877.0, 1522800.0, 244261.0, 38945.0, 7157.0, 1878.0, 774.0, 355.0, 204.0, 129.0, 103.0, 53.0, 45.0, 28.0, 17.0, 18.0, 15.0, 8.0, 7.0, 5.0, 3.0, 2.0, 1.0, 4.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.298828125, -0.288330078125, -0.27783203125, -0.267333984375, -0.2568359375, -0.246337890625, -0.23583984375, -0.225341796875, -0.21484375, -0.204345703125, -0.19384765625, -0.183349609375, -0.1728515625, -0.162353515625, -0.15185546875, -0.141357421875, -0.130859375, -0.120361328125, -0.10986328125, -0.099365234375, -0.0888671875, -0.078369140625, -0.06787109375, -0.057373046875, -0.046875, -0.036376953125, -0.02587890625, -0.015380859375, -0.0048828125, 0.005615234375, 0.01611328125, 0.026611328125, 0.037109375, 0.047607421875, 0.05810546875, 0.068603515625, 0.0791015625, 0.089599609375, 0.10009765625, 0.110595703125, 0.12109375, 0.131591796875, 0.14208984375, 0.152587890625, 0.1630859375, 0.173583984375, 0.18408203125, 0.194580078125, 0.205078125, 0.215576171875, 0.22607421875, 0.236572265625, 0.2470703125, 0.257568359375, 0.26806640625, 0.278564453125, 0.2890625, 0.299560546875, 0.31005859375, 0.320556640625, 0.3310546875, 0.341552734375, 0.35205078125, 0.362548828125, 0.373046875]}, "parameters/model.layers.4.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 4.0, 1.0, 2.0, 2.0, 5.0, 5.0, 10.0, 8.0, 15.0, 25.0, 51.0, 115.0, 260.0, 738.0, 8242.0, 796885.0, 14291394.0, 16777216.0, 9380477.0, 344622.0, 3770.0, 603.0, 207.0, 101.0, 47.0, 17.0, 16.0, 8.0, 5.0, 11.0, 4.0, 1.0, 4.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.58984375, -0.573638916015625, -0.55743408203125, -0.541229248046875, -0.5250244140625, -0.508819580078125, -0.49261474609375, -0.476409912109375, -0.460205078125, -0.444000244140625, -0.42779541015625, -0.411590576171875, -0.3953857421875, -0.379180908203125, -0.36297607421875, -0.346771240234375, -0.33056640625, -0.314361572265625, -0.29815673828125, -0.281951904296875, -0.2657470703125, -0.249542236328125, -0.23333740234375, -0.217132568359375, -0.200927734375, -0.184722900390625, -0.16851806640625, -0.152313232421875, -0.1361083984375, -0.119903564453125, -0.10369873046875, -0.087493896484375, -0.0712890625, -0.055084228515625, -0.03887939453125, -0.022674560546875, -0.0064697265625, 0.009735107421875, 0.02593994140625, 0.042144775390625, 0.058349609375, 0.074554443359375, 0.09075927734375, 0.106964111328125, 0.1231689453125, 0.139373779296875, 0.15557861328125, 0.171783447265625, 0.18798828125, 0.204193115234375, 0.22039794921875, 0.236602783203125, 0.2528076171875, 0.269012451171875, 0.28521728515625, 0.301422119140625, 0.317626953125, 0.333831787109375, 0.35003662109375, 0.366241455078125, 0.3824462890625, 0.398651123046875, 0.41485595703125, 0.431060791015625, 0.447265625]}, "parameters/model.layers.4.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 11.0, 85.0, 418.0, 754.0, 675.0, 702.0, 341.0, 259.0, 161.0, 119.0, 95.0, 56.0, 62.0, 45.0, 30.0, 48.0, 30.0, 35.0, 22.0, 30.0, 18.0, 11.0, 9.0, 18.0, 5.0, 5.0, 9.0, 8.0, 4.0, 5.0, 7.0, 1.0, 1.0, 0.0, 1.0, 2.0, 1.0, 2.0, 2.0, 0.0, 1.0], "bins": [0.0002880096435546875, 0.011208802461624146, 0.022129595279693604, 0.03305038809776306, 0.04397118091583252, 0.05489197373390198, 0.06581276655197144, 0.0767335593700409, 0.08765435218811035, 0.09857514500617981, 0.10949593782424927, 0.12041673064231873, 0.13133752346038818, 0.14225831627845764, 0.1531791090965271, 0.16409990191459656, 0.17502069473266602, 0.18594148755073547, 0.19686228036880493, 0.2077830731868744, 0.21870386600494385, 0.2296246588230133, 0.24054545164108276, 0.2514662444591522, 0.2623870372772217, 0.27330783009529114, 0.2842286229133606, 0.29514941573143005, 0.3060702085494995, 0.31699100136756897, 0.3279117941856384, 0.3388325870037079, 0.34975337982177734, 0.3606741726398468, 0.37159496545791626, 0.3825157582759857, 0.3934365510940552, 0.40435734391212463, 0.4152781367301941, 0.42619892954826355, 0.437119722366333, 0.44804051518440247, 0.4589613080024719, 0.4698821008205414, 0.48080289363861084, 0.4917236864566803, 0.5026444792747498, 0.5135653018951416, 0.5244860649108887, 0.5354068279266357, 0.5463276505470276, 0.5572484731674194, 0.5681692361831665, 0.5790899991989136, 0.5900108218193054, 0.6009316444396973, 0.6118524074554443, 0.6227731704711914, 0.6336939930915833, 0.6446148157119751, 0.6555355787277222, 0.6664563417434692, 0.6773771643638611, 0.6882979869842529, 0.69921875]}, "parameters/model.layers.4.post_attention_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 2.0, 1.0, 1.0, 3.0, 4.0, 2.0, 5.0, 2.0, 4.0, 3.0, 14.0, 10.0, 10.0, 15.0, 25.0, 28.0, 23.0, 21.0, 20.0, 20.0, 21.0, 13.0, 12.0, 19.0, 25.0, 18.0, 28.0, 49.0, 38.0, 50.0, 113.0, 130.0, 359.0, 621.0, 1161.0, 1157.0, 53.0, 2.0], "bins": [-0.00078582763671875, 0.0038651227951049805, 0.008516073226928711, 0.013167023658752441, 0.017817974090576172, 0.022468924522399902, 0.027119874954223633, 0.03177082538604736, 0.036421775817871094, 0.041072726249694824, 0.045723676681518555, 0.050374627113342285, 0.055025577545166016, 0.059676527976989746, 0.06432747840881348, 0.06897842884063721, 0.07362937927246094, 0.07828032970428467, 0.0829312801361084, 0.08758223056793213, 0.09223318099975586, 0.09688413143157959, 0.10153508186340332, 0.10618603229522705, 0.11083698272705078, 0.11548793315887451, 0.12013888359069824, 0.12478983402252197, 0.1294407844543457, 0.13409173488616943, 0.13874268531799316, 0.1433936357498169, 0.14804458618164062, 0.15269553661346436, 0.15734648704528809, 0.16199743747711182, 0.16664838790893555, 0.17129933834075928, 0.175950288772583, 0.18060123920440674, 0.18525218963623047, 0.1899031400680542, 0.19455409049987793, 0.19920504093170166, 0.2038559913635254, 0.20850694179534912, 0.21315789222717285, 0.21780884265899658, 0.2224597930908203, 0.22711074352264404, 0.23176169395446777, 0.2364126443862915, 0.24106359481811523, 0.24571454524993896, 0.2503654956817627, 0.2550164461135864, 0.25966739654541016, 0.2643183469772339, 0.2689692974090576, 0.27362024784088135, 0.2782711982727051, 0.2829221487045288, 0.28757309913635254, 0.29222404956817627, 0.296875]}, "parameters/model.layers.5.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 0.0, 2.0, 3.0, 4.0, 11.0, 18.0, 22.0, 25.0, 36.0, 57.0, 74.0, 89.0, 155.0, 214.0, 305.0, 448.0, 740.0, 1355.0, 2257.0, 4874.0, 10220.0, 21049.0, 49752.0, 98689.0, 196058.0, 390647.0, 757703.0, 1299412.0, 2212652.0, 3196312.0, 3236184.0, 2300164.0, 1377614.0, 797371.0, 408620.0, 209217.0, 108544.0, 50872.0, 23454.0, 10626.0, 5099.0, 2639.0, 1345.0, 694.0, 469.0, 342.0, 224.0, 150.0, 149.0, 85.0, 55.0, 45.0, 17.0, 19.0, 13.0, 7.0, 6.0, 1.0, 2.0, 3.0], "bins": [-0.244140625, -0.236968994140625, -0.22979736328125, -0.222625732421875, -0.2154541015625, -0.208282470703125, -0.20111083984375, -0.193939208984375, -0.186767578125, -0.179595947265625, -0.17242431640625, -0.165252685546875, -0.1580810546875, -0.150909423828125, -0.14373779296875, -0.136566162109375, -0.12939453125, -0.122222900390625, -0.11505126953125, -0.107879638671875, -0.1007080078125, -0.093536376953125, -0.08636474609375, -0.079193115234375, -0.072021484375, -0.064849853515625, -0.05767822265625, -0.050506591796875, -0.0433349609375, -0.036163330078125, -0.02899169921875, -0.021820068359375, -0.0146484375, -0.007476806640625, -0.00030517578125, 0.006866455078125, 0.0140380859375, 0.021209716796875, 0.02838134765625, 0.035552978515625, 0.042724609375, 0.049896240234375, 0.05706787109375, 0.064239501953125, 0.0714111328125, 0.078582763671875, 0.08575439453125, 0.092926025390625, 0.10009765625, 0.107269287109375, 0.11444091796875, 0.121612548828125, 0.1287841796875, 0.135955810546875, 0.14312744140625, 0.150299072265625, 0.157470703125, 0.164642333984375, 0.17181396484375, 0.178985595703125, 0.1861572265625, 0.193328857421875, 0.20050048828125, 0.207672119140625, 0.21484375]}, "parameters/model.layers.5.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 1.0, 4.0, 2.0, 2.0, 5.0, 3.0, 10.0, 13.0, 14.0, 14.0, 25.0, 29.0, 39.0, 59.0, 61.0, 75.0, 130.0, 129.0, 217.0, 420.0, 1218.0, 3722.0, 11250.0, 33356.0, 93414.0, 225569.0, 495287.0, 953969.0, 1128961.0, 690855.0, 326052.0, 146233.0, 53530.0, 19671.0, 6130.0, 2075.0, 734.0, 335.0, 178.0, 116.0, 97.0, 64.0, 48.0, 29.0, 33.0, 26.0, 17.0, 24.0, 14.0, 12.0, 7.0, 8.0, 1.0, 1.0, 0.0, 3.0, 3.0, 0.0, 5.0, 0.0, 2.0], "bins": [-0.4765625, -0.4613037109375, -0.446044921875, -0.4307861328125, -0.41552734375, -0.4002685546875, -0.385009765625, -0.3697509765625, -0.3544921875, -0.3392333984375, -0.323974609375, -0.3087158203125, -0.29345703125, -0.2781982421875, -0.262939453125, -0.2476806640625, -0.232421875, -0.2171630859375, -0.201904296875, -0.1866455078125, -0.17138671875, -0.1561279296875, -0.140869140625, -0.1256103515625, -0.1103515625, -0.0950927734375, -0.079833984375, -0.0645751953125, -0.04931640625, -0.0340576171875, -0.018798828125, -0.0035400390625, 0.01171875, 0.0269775390625, 0.042236328125, 0.0574951171875, 0.07275390625, 0.0880126953125, 0.103271484375, 0.1185302734375, 0.1337890625, 0.1490478515625, 0.164306640625, 0.1795654296875, 0.19482421875, 0.2100830078125, 0.225341796875, 0.2406005859375, 0.255859375, 0.2711181640625, 0.286376953125, 0.3016357421875, 0.31689453125, 0.3321533203125, 0.347412109375, 0.3626708984375, 0.3779296875, 0.3931884765625, 0.408447265625, 0.4237060546875, 0.43896484375, 0.4542236328125, 0.469482421875, 0.4847412109375, 0.5]}, "parameters/model.layers.5.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 3.0, 1.0, 6.0, 3.0, 6.0, 15.0, 80.0, 361.0, 1527.0, 6110.0, 19551.0, 58270.0, 154935.0, 321855.0, 577451.0, 824512.0, 865981.0, 657963.0, 390854.0, 195293.0, 79779.0, 28178.0, 8418.0, 2433.0, 522.0, 114.0, 35.0, 12.0, 6.0, 5.0, 4.0, 2.0, 1.0, 1.0, 1.0, 2.0, 1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.09814453125, -0.09465789794921875, -0.0911712646484375, -0.08768463134765625, -0.084197998046875, -0.08071136474609375, -0.0772247314453125, -0.07373809814453125, -0.07025146484375, -0.06676483154296875, -0.0632781982421875, -0.05979156494140625, -0.056304931640625, -0.05281829833984375, -0.0493316650390625, -0.04584503173828125, -0.0423583984375, -0.03887176513671875, -0.0353851318359375, -0.03189849853515625, -0.028411865234375, -0.02492523193359375, -0.0214385986328125, -0.01795196533203125, -0.01446533203125, -0.01097869873046875, -0.0074920654296875, -0.00400543212890625, -0.000518798828125, 0.00296783447265625, 0.0064544677734375, 0.00994110107421875, 0.013427734375, 0.01691436767578125, 0.0204010009765625, 0.02388763427734375, 0.027374267578125, 0.03086090087890625, 0.0343475341796875, 0.03783416748046875, 0.04132080078125, 0.04480743408203125, 0.0482940673828125, 0.05178070068359375, 0.055267333984375, 0.05875396728515625, 0.0622406005859375, 0.06572723388671875, 0.0692138671875, 0.07270050048828125, 0.0761871337890625, 0.07967376708984375, 0.083160400390625, 0.08664703369140625, 0.0901336669921875, 0.09362030029296875, 0.09710693359375, 0.10059356689453125, 0.1040802001953125, 0.10756683349609375, 0.111053466796875, 0.11454010009765625, 0.1180267333984375, 0.12151336669921875, 0.125]}, "parameters/model.layers.5.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 0.0, 1.0, 4.0, 3.0, 5.0, 6.0, 12.0, 8.0, 16.0, 14.0, 35.0, 58.0, 122.0, 286.0, 1157.0, 11237.0, 141113.0, 1126335.0, 4734950.0, 7062495.0, 3083001.0, 552495.0, 58074.0, 4740.0, 619.0, 205.0, 91.0, 52.0, 25.0, 16.0, 11.0, 7.0, 2.0, 2.0, 0.0, 5.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.3359375, -0.326934814453125, -0.31793212890625, -0.308929443359375, -0.2999267578125, -0.290924072265625, -0.28192138671875, -0.272918701171875, -0.263916015625, -0.254913330078125, -0.24591064453125, -0.236907958984375, -0.2279052734375, -0.218902587890625, -0.20989990234375, -0.200897216796875, -0.19189453125, -0.182891845703125, -0.17388916015625, -0.164886474609375, -0.1558837890625, -0.146881103515625, -0.13787841796875, -0.128875732421875, -0.119873046875, -0.110870361328125, -0.10186767578125, -0.092864990234375, -0.0838623046875, -0.074859619140625, -0.06585693359375, -0.056854248046875, -0.0478515625, -0.038848876953125, -0.02984619140625, -0.020843505859375, -0.0118408203125, -0.002838134765625, 0.00616455078125, 0.015167236328125, 0.024169921875, 0.033172607421875, 0.04217529296875, 0.051177978515625, 0.0601806640625, 0.069183349609375, 0.07818603515625, 0.087188720703125, 0.09619140625, 0.105194091796875, 0.11419677734375, 0.123199462890625, 0.1322021484375, 0.141204833984375, 0.15020751953125, 0.159210205078125, 0.168212890625, 0.177215576171875, 0.18621826171875, 0.195220947265625, 0.2042236328125, 0.213226318359375, 0.22222900390625, 0.231231689453125, 0.240234375]}, "parameters/model.layers.5.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 2.0, 1.0, 2.0, 4.0, 7.0, 5.0, 11.0, 17.0, 21.0, 61.0, 92.0, 181.0, 321.0, 727.0, 1922.0, 10296.0, 90920.0, 777842.0, 5811197.0, 16777216.0, 16777216.0, 16777216.0, 7602065.0, 1088648.0, 124265.0, 13850.0, 2489.0, 751.0, 369.0, 186.0, 106.0, 60.0, 47.0, 22.0, 12.0, 9.0, 6.0, 5.0, 3.0, 2.0, 3.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.38671875, -0.374847412109375, -0.36297607421875, -0.351104736328125, -0.3392333984375, -0.327362060546875, -0.31549072265625, -0.303619384765625, -0.291748046875, -0.279876708984375, -0.26800537109375, -0.256134033203125, -0.2442626953125, -0.232391357421875, -0.22052001953125, -0.208648681640625, -0.19677734375, -0.184906005859375, -0.17303466796875, -0.161163330078125, -0.1492919921875, -0.137420654296875, -0.12554931640625, -0.113677978515625, -0.101806640625, -0.089935302734375, -0.07806396484375, -0.066192626953125, -0.0543212890625, -0.042449951171875, -0.03057861328125, -0.018707275390625, -0.0068359375, 0.005035400390625, 0.01690673828125, 0.028778076171875, 0.0406494140625, 0.052520751953125, 0.06439208984375, 0.076263427734375, 0.088134765625, 0.100006103515625, 0.11187744140625, 0.123748779296875, 0.1356201171875, 0.147491455078125, 0.15936279296875, 0.171234130859375, 0.18310546875, 0.194976806640625, 0.20684814453125, 0.218719482421875, 0.2305908203125, 0.242462158203125, 0.25433349609375, 0.266204833984375, 0.278076171875, 0.289947509765625, 0.30181884765625, 0.313690185546875, 0.3255615234375, 0.337432861328125, 0.34930419921875, 0.361175537109375, 0.373046875]}, "parameters/model.layers.5.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 3.0, 1.0, 2.0, 0.0, 6.0, 4.0, 13.0, 16.0, 20.0, 32.0, 52.0, 81.0, 171.0, 433.0, 2371.0, 213039.0, 6991310.0, 16777216.0, 16777216.0, 1423060.0, 18104.0, 773.0, 257.0, 115.0, 49.0, 38.0, 21.0, 15.0, 10.0, 8.0, 3.0, 5.0, 4.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.5234375, -0.5074462890625, -0.491455078125, -0.4754638671875, -0.45947265625, -0.4434814453125, -0.427490234375, -0.4114990234375, -0.3955078125, -0.3795166015625, -0.363525390625, -0.3475341796875, -0.33154296875, -0.3155517578125, -0.299560546875, -0.2835693359375, -0.267578125, -0.2515869140625, -0.235595703125, -0.2196044921875, -0.20361328125, -0.1876220703125, -0.171630859375, -0.1556396484375, -0.1396484375, -0.1236572265625, -0.107666015625, -0.0916748046875, -0.07568359375, -0.0596923828125, -0.043701171875, -0.0277099609375, -0.01171875, 0.0042724609375, 0.020263671875, 0.0362548828125, 0.05224609375, 0.0682373046875, 0.084228515625, 0.1002197265625, 0.1162109375, 0.1322021484375, 0.148193359375, 0.1641845703125, 0.18017578125, 0.1961669921875, 0.212158203125, 0.2281494140625, 0.244140625, 0.2601318359375, 0.276123046875, 0.2921142578125, 0.30810546875, 0.3240966796875, 0.340087890625, 0.3560791015625, 0.3720703125, 0.3880615234375, 0.404052734375, 0.4200439453125, 0.43603515625, 0.4520263671875, 0.468017578125, 0.4840087890625, 0.5]}, "parameters/model.layers.5.input_layernorm.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 3.0, 0.0, 2.0, 2.0, 21.0, 106.0, 352.0, 572.0, 532.0, 662.0, 411.0, 351.0, 208.0, 142.0, 111.0, 103.0, 52.0, 52.0, 34.0, 28.0, 32.0, 28.0, 38.0, 15.0, 21.0, 15.0, 25.0, 21.0, 15.0, 16.0, 12.0, 10.0, 11.0, 16.0, 9.0, 12.0, 12.0, 5.0, 6.0, 4.0, 7.0, 3.0, 5.0, 3.0, 2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [0.087890625, 0.099517822265625, 0.11114501953125, 0.122772216796875, 0.1343994140625, 0.146026611328125, 0.15765380859375, 0.169281005859375, 0.180908203125, 0.192535400390625, 0.20416259765625, 0.215789794921875, 0.2274169921875, 0.239044189453125, 0.25067138671875, 0.262298583984375, 0.27392578125, 0.285552978515625, 0.29718017578125, 0.308807373046875, 0.3204345703125, 0.332061767578125, 0.34368896484375, 0.355316162109375, 0.366943359375, 0.378570556640625, 0.39019775390625, 0.401824951171875, 0.4134521484375, 0.425079345703125, 0.43670654296875, 0.448333740234375, 0.4599609375, 0.471588134765625, 0.48321533203125, 0.494842529296875, 0.5064697265625, 0.518096923828125, 0.52972412109375, 0.541351318359375, 0.552978515625, 0.564605712890625, 0.57623291015625, 0.587860107421875, 0.5994873046875, 0.611114501953125, 0.62274169921875, 0.634368896484375, 0.64599609375, 0.657623291015625, 0.66925048828125, 0.680877685546875, 0.6925048828125, 0.704132080078125, 0.71575927734375, 0.727386474609375, 0.739013671875, 0.750640869140625, 0.76226806640625, 0.773895263671875, 0.7855224609375, 0.797149658203125, 0.80877685546875, 0.820404052734375, 0.83203125]}, "parameters/model.layers.5.post_attention_layernorm.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 5.0, 1.0, 6.0, 9.0, 8.0, 9.0, 18.0, 19.0, 29.0, 20.0, 22.0, 14.0, 21.0, 21.0, 15.0, 15.0, 15.0, 18.0, 15.0, 27.0, 28.0, 37.0, 30.0, 33.0, 51.0, 96.0, 145.0, 183.0, 467.0, 651.0, 1415.0, 622.0, 20.0, 1.0, 2.0], "bins": [2.956390380859375e-05, 0.0051255375146865845, 0.010221511125564575, 0.015317484736442566, 0.020413458347320557, 0.025509431958198547, 0.030605405569076538, 0.03570137917995453, 0.04079735279083252, 0.04589332640171051, 0.0509893000125885, 0.05608527362346649, 0.06118124723434448, 0.06627722084522247, 0.07137319445610046, 0.07646916806697845, 0.08156514167785645, 0.08666111528873444, 0.09175708889961243, 0.09685306251049042, 0.10194903612136841, 0.1070450097322464, 0.11214098334312439, 0.11723695695400238, 0.12233293056488037, 0.12742890417575836, 0.13252487778663635, 0.13762085139751434, 0.14271682500839233, 0.14781279861927032, 0.15290877223014832, 0.1580047458410263, 0.1631007194519043, 0.1681966930627823, 0.17329266667366028, 0.17838864028453827, 0.18348461389541626, 0.18858058750629425, 0.19367656111717224, 0.19877253472805023, 0.20386850833892822, 0.2089644819498062, 0.2140604555606842, 0.2191564291715622, 0.22425240278244019, 0.22934837639331818, 0.23444435000419617, 0.23954032361507416, 0.24463629722595215, 0.24973227083683014, 0.25482824444770813, 0.2599242329597473, 0.2650201916694641, 0.2701161503791809, 0.2752121388912201, 0.2803081274032593, 0.2854040861129761, 0.29050004482269287, 0.29559603333473206, 0.30069202184677124, 0.30578798055648804, 0.31088393926620483, 0.315979927778244, 0.3210759162902832, 0.326171875]}, "parameters/model.layers.6.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 0.0, 0.0, 2.0, 7.0, 4.0, 3.0, 9.0, 5.0, 16.0, 29.0, 52.0, 76.0, 143.0, 313.0, 490.0, 969.0, 1993.0, 4264.0, 8021.0, 17812.0, 38304.0, 87370.0, 177235.0, 369583.0, 761285.0, 1400709.0, 2361620.0, 3236553.0, 3201021.0, 2312794.0, 1360815.0, 752777.0, 352639.0, 177640.0, 81937.0, 37493.0, 16896.0, 8329.0, 4091.0, 1788.0, 1020.0, 515.0, 255.0, 141.0, 75.0, 38.0, 26.0, 18.0, 8.0, 9.0, 9.0, 4.0, 2.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.236328125, -0.22869873046875, -0.2210693359375, -0.21343994140625, -0.205810546875, -0.19818115234375, -0.1905517578125, -0.18292236328125, -0.17529296875, -0.16766357421875, -0.1600341796875, -0.15240478515625, -0.144775390625, -0.13714599609375, -0.1295166015625, -0.12188720703125, -0.1142578125, -0.10662841796875, -0.0989990234375, -0.09136962890625, -0.083740234375, -0.07611083984375, -0.0684814453125, -0.06085205078125, -0.05322265625, -0.04559326171875, -0.0379638671875, -0.03033447265625, -0.022705078125, -0.01507568359375, -0.0074462890625, 0.00018310546875, 0.0078125, 0.01544189453125, 0.0230712890625, 0.03070068359375, 0.038330078125, 0.04595947265625, 0.0535888671875, 0.06121826171875, 0.06884765625, 0.07647705078125, 0.0841064453125, 0.09173583984375, 0.099365234375, 0.10699462890625, 0.1146240234375, 0.12225341796875, 0.1298828125, 0.13751220703125, 0.1451416015625, 0.15277099609375, 0.160400390625, 0.16802978515625, 0.1756591796875, 0.18328857421875, 0.19091796875, 0.19854736328125, 0.2061767578125, 0.21380615234375, 0.221435546875, 0.22906494140625, 0.2366943359375, 0.24432373046875, 0.251953125]}, "parameters/model.layers.6.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 1.0, 1.0, 1.0, 1.0, 6.0, 12.0, 7.0, 23.0, 27.0, 31.0, 42.0, 77.0, 99.0, 156.0, 288.0, 489.0, 859.0, 1616.0, 2961.0, 4945.0, 9512.0, 17052.0, 28655.0, 54578.0, 88670.0, 140423.0, 227001.0, 334314.0, 448261.0, 550822.0, 579482.0, 521448.0, 402203.0, 296749.0, 191502.0, 119158.0, 74919.0, 44723.0, 22410.0, 13761.0, 7641.0, 4117.0, 2268.0, 1263.0, 683.0, 414.0, 238.0, 141.0, 83.0, 65.0, 33.0, 28.0, 10.0, 12.0, 7.0, 4.0, 5.0, 2.0, 1.0, 0.0, 1.0], "bins": [-0.271484375, -0.263336181640625, -0.25518798828125, -0.247039794921875, -0.2388916015625, -0.230743408203125, -0.22259521484375, -0.214447021484375, -0.206298828125, -0.198150634765625, -0.19000244140625, -0.181854248046875, -0.1737060546875, -0.165557861328125, -0.15740966796875, -0.149261474609375, -0.14111328125, -0.132965087890625, -0.12481689453125, -0.116668701171875, -0.1085205078125, -0.100372314453125, -0.09222412109375, -0.084075927734375, -0.075927734375, -0.067779541015625, -0.05963134765625, -0.051483154296875, -0.0433349609375, -0.035186767578125, -0.02703857421875, -0.018890380859375, -0.0107421875, -0.002593994140625, 0.00555419921875, 0.013702392578125, 0.0218505859375, 0.029998779296875, 0.03814697265625, 0.046295166015625, 0.054443359375, 0.062591552734375, 0.07073974609375, 0.078887939453125, 0.0870361328125, 0.095184326171875, 0.10333251953125, 0.111480712890625, 0.11962890625, 0.127777099609375, 0.13592529296875, 0.144073486328125, 0.1522216796875, 0.160369873046875, 0.16851806640625, 0.176666259765625, 0.184814453125, 0.192962646484375, 0.20111083984375, 0.209259033203125, 0.2174072265625, 0.225555419921875, 0.23370361328125, 0.241851806640625, 0.25]}, "parameters/model.layers.6.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 3.0, 5.0, 7.0, 12.0, 19.0, 49.0, 73.0, 162.0, 271.0, 485.0, 799.0, 1485.0, 2537.0, 4358.0, 6684.0, 10640.0, 17358.0, 26041.0, 40916.0, 61922.0, 86695.0, 118639.0, 156576.0, 204728.0, 256405.0, 294823.0, 341861.0, 369593.0, 376717.0, 354506.0, 324923.0, 274410.0, 232043.0, 176442.0, 136803.0, 105224.0, 71237.0, 49232.0, 32036.0, 21402.0, 14737.0, 8645.0, 5325.0, 3065.0, 1885.0, 1100.0, 617.0, 384.0, 195.0, 119.0, 47.0, 29.0, 17.0, 9.0, 4.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.04833984375, -0.046741485595703125, -0.04514312744140625, -0.043544769287109375, -0.0419464111328125, -0.040348052978515625, -0.03874969482421875, -0.037151336669921875, -0.035552978515625, -0.033954620361328125, -0.03235626220703125, -0.030757904052734375, -0.0291595458984375, -0.027561187744140625, -0.02596282958984375, -0.024364471435546875, -0.02276611328125, -0.021167755126953125, -0.01956939697265625, -0.017971038818359375, -0.0163726806640625, -0.014774322509765625, -0.01317596435546875, -0.011577606201171875, -0.009979248046875, -0.008380889892578125, -0.00678253173828125, -0.005184173583984375, -0.0035858154296875, -0.001987457275390625, -0.00038909912109375, 0.001209259033203125, 0.0028076171875, 0.004405975341796875, 0.00600433349609375, 0.007602691650390625, 0.0092010498046875, 0.010799407958984375, 0.01239776611328125, 0.013996124267578125, 0.015594482421875, 0.017192840576171875, 0.01879119873046875, 0.020389556884765625, 0.0219879150390625, 0.023586273193359375, 0.02518463134765625, 0.026782989501953125, 0.02838134765625, 0.029979705810546875, 0.03157806396484375, 0.033176422119140625, 0.0347747802734375, 0.036373138427734375, 0.03797149658203125, 0.039569854736328125, 0.041168212890625, 0.042766571044921875, 0.04436492919921875, 0.045963287353515625, 0.0475616455078125, 0.049160003662109375, 0.05075836181640625, 0.052356719970703125, 0.053955078125]}, "parameters/model.layers.6.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 0.0, 3.0, 7.0, 9.0, 20.0, 47.0, 82.0, 287.0, 4145.0, 360167.0, 6781102.0, 8864835.0, 756420.0, 9543.0, 348.0, 106.0, 42.0, 22.0, 12.0, 2.0, 5.0, 1.0, 1.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.345703125, -0.329315185546875, -0.31292724609375, -0.296539306640625, -0.2801513671875, -0.263763427734375, -0.24737548828125, -0.230987548828125, -0.214599609375, -0.198211669921875, -0.18182373046875, -0.165435791015625, -0.1490478515625, -0.132659912109375, -0.11627197265625, -0.099884033203125, -0.08349609375, -0.067108154296875, -0.05072021484375, -0.034332275390625, -0.0179443359375, -0.001556396484375, 0.01483154296875, 0.031219482421875, 0.047607421875, 0.063995361328125, 0.08038330078125, 0.096771240234375, 0.1131591796875, 0.129547119140625, 0.14593505859375, 0.162322998046875, 0.1787109375, 0.195098876953125, 0.21148681640625, 0.227874755859375, 0.2442626953125, 0.260650634765625, 0.27703857421875, 0.293426513671875, 0.309814453125, 0.326202392578125, 0.34259033203125, 0.358978271484375, 0.3753662109375, 0.391754150390625, 0.40814208984375, 0.424530029296875, 0.44091796875, 0.457305908203125, 0.47369384765625, 0.490081787109375, 0.5064697265625, 0.522857666015625, 0.53924560546875, 0.555633544921875, 0.572021484375, 0.588409423828125, 0.60479736328125, 0.621185302734375, 0.6375732421875, 0.653961181640625, 0.67034912109375, 0.686737060546875, 0.703125]}, "parameters/model.layers.6.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 2.0, 0.0, 0.0, 3.0, 5.0, 2.0, 5.0, 7.0, 16.0, 31.0, 50.0, 71.0, 153.0, 266.0, 541.0, 1261.0, 4577.0, 28077.0, 209702.0, 1496793.0, 9126712.0, 16777216.0, 16777216.0, 16777216.0, 5748336.0, 850792.0, 116349.0, 15679.0, 3118.0, 931.0, 401.0, 221.0, 108.0, 63.0, 45.0, 23.0, 11.0, 11.0, 5.0, 8.0, 3.0, 5.0, 1.0, 3.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.30078125, -0.28936767578125, -0.2779541015625, -0.26654052734375, -0.255126953125, -0.24371337890625, -0.2322998046875, -0.22088623046875, -0.20947265625, -0.19805908203125, -0.1866455078125, -0.17523193359375, -0.163818359375, -0.15240478515625, -0.1409912109375, -0.12957763671875, -0.1181640625, -0.10675048828125, -0.0953369140625, -0.08392333984375, -0.072509765625, -0.06109619140625, -0.0496826171875, -0.03826904296875, -0.02685546875, -0.01544189453125, -0.0040283203125, 0.00738525390625, 0.018798828125, 0.03021240234375, 0.0416259765625, 0.05303955078125, 0.064453125, 0.07586669921875, 0.0872802734375, 0.09869384765625, 0.110107421875, 0.12152099609375, 0.1329345703125, 0.14434814453125, 0.15576171875, 0.16717529296875, 0.1785888671875, 0.19000244140625, 0.201416015625, 0.21282958984375, 0.2242431640625, 0.23565673828125, 0.2470703125, 0.25848388671875, 0.2698974609375, 0.28131103515625, 0.292724609375, 0.30413818359375, 0.3155517578125, 0.32696533203125, 0.33837890625, 0.34979248046875, 0.3612060546875, 0.37261962890625, 0.384033203125, 0.39544677734375, 0.4068603515625, 0.41827392578125, 0.4296875]}, "parameters/model.layers.6.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 2.0, 0.0, 1.0, 2.0, 0.0, 1.0, 3.0, 0.0, 2.0, 3.0, 6.0, 9.0, 8.0, 18.0, 25.0, 47.0, 73.0, 129.0, 287.0, 1095.0, 59234.0, 3462465.0, 16777216.0, 16777216.0, 2617079.0, 38403.0, 947.0, 291.0, 126.0, 66.0, 32.0, 20.0, 15.0, 6.0, 0.0, 3.0, 4.0, 5.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.578125, -0.56158447265625, -0.5450439453125, -0.52850341796875, -0.511962890625, -0.49542236328125, -0.4788818359375, -0.46234130859375, -0.44580078125, -0.42926025390625, -0.4127197265625, -0.39617919921875, -0.379638671875, -0.36309814453125, -0.3465576171875, -0.33001708984375, -0.3134765625, -0.29693603515625, -0.2803955078125, -0.26385498046875, -0.247314453125, -0.23077392578125, -0.2142333984375, -0.19769287109375, -0.18115234375, -0.16461181640625, -0.1480712890625, -0.13153076171875, -0.114990234375, -0.09844970703125, -0.0819091796875, -0.06536865234375, -0.048828125, -0.03228759765625, -0.0157470703125, 0.00079345703125, 0.017333984375, 0.03387451171875, 0.0504150390625, 0.06695556640625, 0.08349609375, 0.10003662109375, 0.1165771484375, 0.13311767578125, 0.149658203125, 0.16619873046875, 0.1827392578125, 0.19927978515625, 0.2158203125, 0.23236083984375, 0.2489013671875, 0.26544189453125, 0.281982421875, 0.29852294921875, 0.3150634765625, 0.33160400390625, 0.34814453125, 0.36468505859375, 0.3812255859375, 0.39776611328125, 0.414306640625, 0.43084716796875, 0.4473876953125, 0.46392822265625, 0.48046875]}, "parameters/model.layers.6.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 5.0, 9.0, 54.0, 196.0, 377.0, 618.0, 726.0, 586.0, 382.0, 267.0, 191.0, 150.0, 104.0, 75.0, 67.0, 48.0, 41.0, 28.0, 27.0, 36.0, 11.0, 6.0, 11.0, 10.0, 7.0, 11.0, 7.0, 6.0, 8.0, 3.0, 1.0, 3.0, 4.0, 2.0, 4.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [0.087890625, 0.099151611328125, 0.11041259765625, 0.121673583984375, 0.1329345703125, 0.144195556640625, 0.15545654296875, 0.166717529296875, 0.177978515625, 0.189239501953125, 0.20050048828125, 0.211761474609375, 0.2230224609375, 0.234283447265625, 0.24554443359375, 0.256805419921875, 0.26806640625, 0.279327392578125, 0.29058837890625, 0.301849365234375, 0.3131103515625, 0.324371337890625, 0.33563232421875, 0.346893310546875, 0.358154296875, 0.369415283203125, 0.38067626953125, 0.391937255859375, 0.4031982421875, 0.414459228515625, 0.42572021484375, 0.436981201171875, 0.4482421875, 0.459503173828125, 0.47076416015625, 0.482025146484375, 0.4932861328125, 0.504547119140625, 0.51580810546875, 0.527069091796875, 0.538330078125, 0.549591064453125, 0.56085205078125, 0.572113037109375, 0.5833740234375, 0.594635009765625, 0.60589599609375, 0.617156982421875, 0.62841796875, 0.639678955078125, 0.65093994140625, 0.662200927734375, 0.6734619140625, 0.684722900390625, 0.69598388671875, 0.707244873046875, 0.718505859375, 0.729766845703125, 0.74102783203125, 0.752288818359375, 0.7635498046875, 0.774810791015625, 0.78607177734375, 0.797332763671875, 0.80859375]}, "parameters/model.layers.6.post_attention_layernorm.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 2.0, 6.0, 7.0, 13.0, 21.0, 30.0, 23.0, 27.0, 18.0, 24.0, 16.0, 23.0, 24.0, 24.0, 29.0, 31.0, 31.0, 33.0, 47.0, 71.0, 114.0, 141.0, 186.0, 420.0, 869.0, 1281.0, 547.0, 29.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.00069427490234375, 0.004748702049255371, 0.010191679000854492, 0.015634655952453613, 0.021077632904052734, 0.026520609855651855, 0.03196358680725098, 0.0374065637588501, 0.04284954071044922, 0.04829251766204834, 0.05373549461364746, 0.05917847156524658, 0.0646214485168457, 0.07006442546844482, 0.07550740242004395, 0.08095037937164307, 0.08639335632324219, 0.09183633327484131, 0.09727931022644043, 0.10272228717803955, 0.10816526412963867, 0.11360824108123779, 0.11905121803283691, 0.12449419498443604, 0.12993717193603516, 0.13538014888763428, 0.1408231258392334, 0.14626610279083252, 0.15170907974243164, 0.15715205669403076, 0.16259503364562988, 0.168038010597229, 0.17348098754882812, 0.17892396450042725, 0.18436694145202637, 0.1898099184036255, 0.1952528953552246, 0.20069587230682373, 0.20613884925842285, 0.21158182621002197, 0.2170248031616211, 0.22246778011322021, 0.22791075706481934, 0.23335373401641846, 0.23879671096801758, 0.2442396879196167, 0.24968266487121582, 0.25512564182281494, 0.26056861877441406, 0.2660115957260132, 0.2714545726776123, 0.2768975496292114, 0.28234052658081055, 0.28778350353240967, 0.2932264804840088, 0.2986694574356079, 0.30411243438720703, 0.30955541133880615, 0.3149983882904053, 0.3204413652420044, 0.3258843421936035, 0.33132731914520264, 0.33677029609680176, 0.3422132730484009, 0.34765625]}, "parameters/model.layers.7.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 1.0, 4.0, 5.0, 10.0, 17.0, 34.0, 70.0, 138.0, 264.0, 563.0, 1182.0, 2530.0, 5381.0, 11588.0, 25257.0, 54016.0, 110319.0, 217249.0, 429009.0, 804480.0, 1397605.0, 2273470.0, 3087735.0, 3061988.0, 2249959.0, 1400169.0, 797186.0, 417338.0, 218144.0, 113119.0, 52504.0, 24877.0, 10738.0, 5334.0, 2516.0, 1196.0, 637.0, 310.0, 133.0, 57.0, 31.0, 21.0, 6.0, 11.0, 4.0, 2.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.2333984375, -0.226531982421875, -0.21966552734375, -0.212799072265625, -0.2059326171875, -0.199066162109375, -0.19219970703125, -0.185333251953125, -0.178466796875, -0.171600341796875, -0.16473388671875, -0.157867431640625, -0.1510009765625, -0.144134521484375, -0.13726806640625, -0.130401611328125, -0.12353515625, -0.116668701171875, -0.10980224609375, -0.102935791015625, -0.0960693359375, -0.089202880859375, -0.08233642578125, -0.075469970703125, -0.068603515625, -0.061737060546875, -0.05487060546875, -0.048004150390625, -0.0411376953125, -0.034271240234375, -0.02740478515625, -0.020538330078125, -0.013671875, -0.006805419921875, 6.103515625e-05, 0.006927490234375, 0.0137939453125, 0.020660400390625, 0.02752685546875, 0.034393310546875, 0.041259765625, 0.048126220703125, 0.05499267578125, 0.061859130859375, 0.0687255859375, 0.075592041015625, 0.08245849609375, 0.089324951171875, 0.09619140625, 0.103057861328125, 0.10992431640625, 0.116790771484375, 0.1236572265625, 0.130523681640625, 0.13739013671875, 0.144256591796875, 0.151123046875, 0.157989501953125, 0.16485595703125, 0.171722412109375, 0.1785888671875, 0.185455322265625, 0.19232177734375, 0.199188232421875, 0.2060546875]}, "parameters/model.layers.7.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 5.0, 4.0, 3.0, 5.0, 18.0, 33.0, 42.0, 77.0, 130.0, 184.0, 330.0, 609.0, 1176.0, 1972.0, 4016.0, 6087.0, 11445.0, 17674.0, 30471.0, 49187.0, 76837.0, 113543.0, 165730.0, 241135.0, 319225.0, 416080.0, 483873.0, 501569.0, 459529.0, 377923.0, 290977.0, 209990.0, 144460.0, 99532.0, 66446.0, 41785.0, 25740.0, 14676.0, 9541.0, 5256.0, 3188.0, 1618.0, 983.0, 516.0, 300.0, 132.0, 95.0, 54.0, 35.0, 31.0, 10.0, 8.0, 9.0, 2.0, 0.0, 3.0], "bins": [-0.25, -0.242919921875, -0.23583984375, -0.228759765625, -0.2216796875, -0.214599609375, -0.20751953125, -0.200439453125, -0.193359375, -0.186279296875, -0.17919921875, -0.172119140625, -0.1650390625, -0.157958984375, -0.15087890625, -0.143798828125, -0.13671875, -0.129638671875, -0.12255859375, -0.115478515625, -0.1083984375, -0.101318359375, -0.09423828125, -0.087158203125, -0.080078125, -0.072998046875, -0.06591796875, -0.058837890625, -0.0517578125, -0.044677734375, -0.03759765625, -0.030517578125, -0.0234375, -0.016357421875, -0.00927734375, -0.002197265625, 0.0048828125, 0.011962890625, 0.01904296875, 0.026123046875, 0.033203125, 0.040283203125, 0.04736328125, 0.054443359375, 0.0615234375, 0.068603515625, 0.07568359375, 0.082763671875, 0.08984375, 0.096923828125, 0.10400390625, 0.111083984375, 0.1181640625, 0.125244140625, 0.13232421875, 0.139404296875, 0.146484375, 0.153564453125, 0.16064453125, 0.167724609375, 0.1748046875, 0.181884765625, 0.18896484375, 0.196044921875, 0.203125]}, "parameters/model.layers.7.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 0.0, 2.0, 3.0, 3.0, 15.0, 27.0, 39.0, 106.0, 223.0, 444.0, 898.0, 1817.0, 3643.0, 6370.0, 10830.0, 20043.0, 31249.0, 52083.0, 80543.0, 123597.0, 169173.0, 234837.0, 307708.0, 363931.0, 417933.0, 441241.0, 427640.0, 376082.0, 317902.0, 250529.0, 188896.0, 134001.0, 88475.0, 59331.0, 35026.0, 21380.0, 12544.0, 7617.0, 3849.0, 2098.0, 1152.0, 549.0, 231.0, 133.0, 56.0, 28.0, 11.0, 2.0, 3.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.051513671875, -0.049648284912109375, -0.04778289794921875, -0.045917510986328125, -0.0440521240234375, -0.042186737060546875, -0.04032135009765625, -0.038455963134765625, -0.036590576171875, -0.034725189208984375, -0.03285980224609375, -0.030994415283203125, -0.0291290283203125, -0.027263641357421875, -0.02539825439453125, -0.023532867431640625, -0.02166748046875, -0.019802093505859375, -0.01793670654296875, -0.016071319580078125, -0.0142059326171875, -0.012340545654296875, -0.01047515869140625, -0.008609771728515625, -0.006744384765625, -0.004878997802734375, -0.00301361083984375, -0.001148223876953125, 0.0007171630859375, 0.002582550048828125, 0.00444793701171875, 0.006313323974609375, 0.0081787109375, 0.010044097900390625, 0.01190948486328125, 0.013774871826171875, 0.0156402587890625, 0.017505645751953125, 0.01937103271484375, 0.021236419677734375, 0.023101806640625, 0.024967193603515625, 0.02683258056640625, 0.028697967529296875, 0.0305633544921875, 0.032428741455078125, 0.03429412841796875, 0.036159515380859375, 0.03802490234375, 0.039890289306640625, 0.04175567626953125, 0.043621063232421875, 0.0454864501953125, 0.047351837158203125, 0.04921722412109375, 0.051082611083984375, 0.052947998046875, 0.054813385009765625, 0.05667877197265625, 0.058544158935546875, 0.0604095458984375, 0.062274932861328125, 0.06414031982421875, 0.06600570678710938, 0.06787109375]}, "parameters/model.layers.7.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 2.0, 0.0, 3.0, 2.0, 2.0, 3.0, 4.0, 5.0, 10.0, 10.0, 18.0, 30.0, 56.0, 85.0, 177.0, 349.0, 1177.0, 7362.0, 73008.0, 563284.0, 2726350.0, 6125670.0, 5172344.0, 1755698.0, 312539.0, 34036.0, 3616.0, 719.0, 261.0, 149.0, 87.0, 48.0, 30.0, 21.0, 18.0, 4.0, 10.0, 6.0, 2.0, 6.0, 2.0, 3.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.255859375, -0.247283935546875, -0.23870849609375, -0.230133056640625, -0.2215576171875, -0.212982177734375, -0.20440673828125, -0.195831298828125, -0.187255859375, -0.178680419921875, -0.17010498046875, -0.161529541015625, -0.1529541015625, -0.144378662109375, -0.13580322265625, -0.127227783203125, -0.11865234375, -0.110076904296875, -0.10150146484375, -0.092926025390625, -0.0843505859375, -0.075775146484375, -0.06719970703125, -0.058624267578125, -0.050048828125, -0.041473388671875, -0.03289794921875, -0.024322509765625, -0.0157470703125, -0.007171630859375, 0.00140380859375, 0.009979248046875, 0.0185546875, 0.027130126953125, 0.03570556640625, 0.044281005859375, 0.0528564453125, 0.061431884765625, 0.07000732421875, 0.078582763671875, 0.087158203125, 0.095733642578125, 0.10430908203125, 0.112884521484375, 0.1214599609375, 0.130035400390625, 0.13861083984375, 0.147186279296875, 0.15576171875, 0.164337158203125, 0.17291259765625, 0.181488037109375, 0.1900634765625, 0.198638916015625, 0.20721435546875, 0.215789794921875, 0.224365234375, 0.232940673828125, 0.24151611328125, 0.250091552734375, 0.2586669921875, 0.267242431640625, 0.27581787109375, 0.284393310546875, 0.29296875]}, "parameters/model.layers.7.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 1.0, 4.0, 0.0, 4.0, 5.0, 5.0, 11.0, 20.0, 17.0, 28.0, 33.0, 81.0, 148.0, 265.0, 523.0, 1269.0, 4648.0, 23398.0, 136391.0, 843086.0, 4872271.0, 16777216.0, 16777216.0, 16777216.0, 14629602.0, 3398286.0, 563625.0, 94181.0, 16351.0, 3269.0, 1018.0, 431.0, 205.0, 126.0, 61.0, 47.0, 18.0, 19.0, 11.0, 2.0, 7.0, 3.0, 4.0, 2.0, 2.0, 1.0, 2.0, 1.0, 2.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.27734375, -0.26702880859375, -0.2567138671875, -0.24639892578125, -0.236083984375, -0.22576904296875, -0.2154541015625, -0.20513916015625, -0.19482421875, -0.18450927734375, -0.1741943359375, -0.16387939453125, -0.153564453125, -0.14324951171875, -0.1329345703125, -0.12261962890625, -0.1123046875, -0.10198974609375, -0.0916748046875, -0.08135986328125, -0.071044921875, -0.06072998046875, -0.0504150390625, -0.04010009765625, -0.02978515625, -0.01947021484375, -0.0091552734375, 0.00115966796875, 0.011474609375, 0.02178955078125, 0.0321044921875, 0.04241943359375, 0.052734375, 0.06304931640625, 0.0733642578125, 0.08367919921875, 0.093994140625, 0.10430908203125, 0.1146240234375, 0.12493896484375, 0.13525390625, 0.14556884765625, 0.1558837890625, 0.16619873046875, 0.176513671875, 0.18682861328125, 0.1971435546875, 0.20745849609375, 0.2177734375, 0.22808837890625, 0.2384033203125, 0.24871826171875, 0.259033203125, 0.26934814453125, 0.2796630859375, 0.28997802734375, 0.30029296875, 0.31060791015625, 0.3209228515625, 0.33123779296875, 0.341552734375, 0.35186767578125, 0.3621826171875, 0.37249755859375, 0.3828125]}, "parameters/model.layers.7.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 2.0, 4.0, 6.0, 6.0, 13.0, 17.0, 35.0, 45.0, 84.0, 158.0, 352.0, 806.0, 7041.0, 580188.0, 12006607.0, 16777216.0, 11108924.0, 500587.0, 6213.0, 828.0, 326.0, 169.0, 94.0, 52.0, 36.0, 15.0, 15.0, 9.0, 8.0, 5.0, 6.0, 2.0, 4.0, 3.0, 0.0, 3.0, 0.0, 3.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.5703125, -0.55377197265625, -0.5372314453125, -0.52069091796875, -0.504150390625, -0.48760986328125, -0.4710693359375, -0.45452880859375, -0.43798828125, -0.42144775390625, -0.4049072265625, -0.38836669921875, -0.371826171875, -0.35528564453125, -0.3387451171875, -0.32220458984375, -0.3056640625, -0.28912353515625, -0.2725830078125, -0.25604248046875, -0.239501953125, -0.22296142578125, -0.2064208984375, -0.18988037109375, -0.17333984375, -0.15679931640625, -0.1402587890625, -0.12371826171875, -0.107177734375, -0.09063720703125, -0.0740966796875, -0.05755615234375, -0.041015625, -0.02447509765625, -0.0079345703125, 0.00860595703125, 0.025146484375, 0.04168701171875, 0.0582275390625, 0.07476806640625, 0.09130859375, 0.10784912109375, 0.1243896484375, 0.14093017578125, 0.157470703125, 0.17401123046875, 0.1905517578125, 0.20709228515625, 0.2236328125, 0.24017333984375, 0.2567138671875, 0.27325439453125, 0.289794921875, 0.30633544921875, 0.3228759765625, 0.33941650390625, 0.35595703125, 0.37249755859375, 0.3890380859375, 0.40557861328125, 0.422119140625, 0.43865966796875, 0.4552001953125, 0.47174072265625, 0.48828125]}, "parameters/model.layers.7.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 3.0, 3.0, 3.0, 13.0, 36.0, 83.0, 204.0, 596.0, 649.0, 702.0, 561.0, 463.0, 253.0, 164.0, 106.0, 62.0, 52.0, 22.0, 19.0, 20.0, 8.0, 7.0, 6.0, 5.0, 11.0, 2.0, 4.0, 4.0, 0.0, 4.0, 2.0, 5.0, 3.0, 1.0, 1.0, 8.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0], "bins": [0.087890625, 0.100067138671875, 0.11224365234375, 0.124420166015625, 0.1365966796875, 0.148773193359375, 0.16094970703125, 0.173126220703125, 0.185302734375, 0.197479248046875, 0.20965576171875, 0.221832275390625, 0.2340087890625, 0.246185302734375, 0.25836181640625, 0.270538330078125, 0.28271484375, 0.294891357421875, 0.30706787109375, 0.319244384765625, 0.3314208984375, 0.343597412109375, 0.35577392578125, 0.367950439453125, 0.380126953125, 0.392303466796875, 0.40447998046875, 0.416656494140625, 0.4288330078125, 0.441009521484375, 0.45318603515625, 0.465362548828125, 0.4775390625, 0.489715576171875, 0.50189208984375, 0.514068603515625, 0.5262451171875, 0.538421630859375, 0.55059814453125, 0.562774658203125, 0.574951171875, 0.587127685546875, 0.59930419921875, 0.611480712890625, 0.6236572265625, 0.635833740234375, 0.64801025390625, 0.660186767578125, 0.67236328125, 0.684539794921875, 0.69671630859375, 0.708892822265625, 0.7210693359375, 0.733245849609375, 0.74542236328125, 0.757598876953125, 0.769775390625, 0.781951904296875, 0.79412841796875, 0.806304931640625, 0.8184814453125, 0.830657958984375, 0.84283447265625, 0.855010986328125, 0.8671875]}, "parameters/model.layers.7.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 2.0, 2.0, 7.0, 9.0, 8.0, 14.0, 33.0, 31.0, 19.0, 19.0, 23.0, 22.0, 32.0, 9.0, 32.0, 22.0, 46.0, 27.0, 26.0, 51.0, 56.0, 90.0, 80.0, 95.0, 265.0, 253.0, 381.0, 857.0, 704.0, 685.0, 143.0, 42.0, 5.0, 1.0, 1.0], "bins": [0.0458984375, 0.0505523681640625, 0.055206298828125, 0.0598602294921875, 0.06451416015625, 0.0691680908203125, 0.073822021484375, 0.0784759521484375, 0.0831298828125, 0.0877838134765625, 0.092437744140625, 0.0970916748046875, 0.10174560546875, 0.1063995361328125, 0.111053466796875, 0.1157073974609375, 0.120361328125, 0.1250152587890625, 0.129669189453125, 0.1343231201171875, 0.13897705078125, 0.1436309814453125, 0.148284912109375, 0.1529388427734375, 0.1575927734375, 0.1622467041015625, 0.166900634765625, 0.1715545654296875, 0.17620849609375, 0.1808624267578125, 0.185516357421875, 0.1901702880859375, 0.19482421875, 0.1994781494140625, 0.204132080078125, 0.2087860107421875, 0.21343994140625, 0.2180938720703125, 0.222747802734375, 0.2274017333984375, 0.2320556640625, 0.2367095947265625, 0.241363525390625, 0.2460174560546875, 0.25067138671875, 0.2553253173828125, 0.259979248046875, 0.2646331787109375, 0.269287109375, 0.2739410400390625, 0.278594970703125, 0.2832489013671875, 0.28790283203125, 0.2925567626953125, 0.297210693359375, 0.3018646240234375, 0.3065185546875, 0.3111724853515625, 0.315826416015625, 0.3204803466796875, 0.32513427734375, 0.3297882080078125, 0.334442138671875, 0.3390960693359375, 0.34375]}, "parameters/model.layers.8.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 4.0, 1.0, 0.0, 5.0, 2.0, 7.0, 5.0, 3.0, 4.0, 18.0, 13.0, 16.0, 28.0, 37.0, 40.0, 42.0, 66.0, 81.0, 121.0, 179.0, 304.0, 683.0, 1785.0, 5461.0, 16986.0, 65397.0, 226015.0, 791865.0, 2510289.0, 5319533.0, 4885195.0, 2062411.0, 636645.0, 183173.0, 49921.0, 13635.0, 4195.0, 1449.0, 683.0, 291.0, 148.0, 122.0, 86.0, 59.0, 45.0, 21.0, 38.0, 26.0, 19.0, 14.0, 10.0, 8.0, 6.0, 6.0, 8.0, 2.0, 4.0, 1.0, 0.0, 3.0], "bins": [-0.43359375, -0.420806884765625, -0.40802001953125, -0.395233154296875, -0.3824462890625, -0.369659423828125, -0.35687255859375, -0.344085693359375, -0.331298828125, -0.318511962890625, -0.30572509765625, -0.292938232421875, -0.2801513671875, -0.267364501953125, -0.25457763671875, -0.241790771484375, -0.22900390625, -0.216217041015625, -0.20343017578125, -0.190643310546875, -0.1778564453125, -0.165069580078125, -0.15228271484375, -0.139495849609375, -0.126708984375, -0.113922119140625, -0.10113525390625, -0.088348388671875, -0.0755615234375, -0.062774658203125, -0.04998779296875, -0.037200927734375, -0.0244140625, -0.011627197265625, 0.00115966796875, 0.013946533203125, 0.0267333984375, 0.039520263671875, 0.05230712890625, 0.065093994140625, 0.077880859375, 0.090667724609375, 0.10345458984375, 0.116241455078125, 0.1290283203125, 0.141815185546875, 0.15460205078125, 0.167388916015625, 0.18017578125, 0.192962646484375, 0.20574951171875, 0.218536376953125, 0.2313232421875, 0.244110107421875, 0.25689697265625, 0.269683837890625, 0.282470703125, 0.295257568359375, 0.30804443359375, 0.320831298828125, 0.3336181640625, 0.346405029296875, 0.35919189453125, 0.371978759765625, 0.384765625]}, "parameters/model.layers.8.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 3.0, 3.0, 5.0, 0.0, 3.0, 4.0, 5.0, 8.0, 13.0, 24.0, 26.0, 43.0, 49.0, 80.0, 138.0, 245.0, 403.0, 836.0, 2087.0, 4495.0, 10927.0, 26411.0, 62349.0, 148741.0, 319484.0, 625445.0, 958307.0, 928051.0, 585508.0, 289278.0, 133068.0, 57565.0, 22912.0, 10162.0, 4106.0, 1667.0, 859.0, 407.0, 199.0, 125.0, 87.0, 43.0, 45.0, 35.0, 8.0, 14.0, 10.0, 8.0, 0.0, 3.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.4453125, -0.431793212890625, -0.41827392578125, -0.404754638671875, -0.3912353515625, -0.377716064453125, -0.36419677734375, -0.350677490234375, -0.337158203125, -0.323638916015625, -0.31011962890625, -0.296600341796875, -0.2830810546875, -0.269561767578125, -0.25604248046875, -0.242523193359375, -0.22900390625, -0.215484619140625, -0.20196533203125, -0.188446044921875, -0.1749267578125, -0.161407470703125, -0.14788818359375, -0.134368896484375, -0.120849609375, -0.107330322265625, -0.09381103515625, -0.080291748046875, -0.0667724609375, -0.053253173828125, -0.03973388671875, -0.026214599609375, -0.0126953125, 0.000823974609375, 0.01434326171875, 0.027862548828125, 0.0413818359375, 0.054901123046875, 0.06842041015625, 0.081939697265625, 0.095458984375, 0.108978271484375, 0.12249755859375, 0.136016845703125, 0.1495361328125, 0.163055419921875, 0.17657470703125, 0.190093994140625, 0.20361328125, 0.217132568359375, 0.23065185546875, 0.244171142578125, 0.2576904296875, 0.271209716796875, 0.28472900390625, 0.298248291015625, 0.311767578125, 0.325286865234375, 0.33880615234375, 0.352325439453125, 0.3658447265625, 0.379364013671875, 0.39288330078125, 0.406402587890625, 0.419921875]}, "parameters/model.layers.8.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 3.0, 3.0, 5.0, 13.0, 16.0, 34.0, 59.0, 127.0, 199.0, 411.0, 828.0, 1398.0, 2756.0, 4647.0, 8775.0, 14455.0, 25423.0, 45856.0, 76615.0, 125862.0, 200714.0, 293398.0, 406263.0, 501703.0, 560445.0, 530122.0, 444951.0, 336744.0, 234401.0, 154441.0, 94363.0, 54810.0, 31684.0, 19157.0, 10399.0, 5788.0, 3436.0, 1740.0, 1060.0, 581.0, 285.0, 160.0, 97.0, 36.0, 25.0, 6.0, 4.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.06787109375, -0.0655059814453125, -0.063140869140625, -0.0607757568359375, -0.05841064453125, -0.0560455322265625, -0.053680419921875, -0.0513153076171875, -0.0489501953125, -0.0465850830078125, -0.044219970703125, -0.0418548583984375, -0.03948974609375, -0.0371246337890625, -0.034759521484375, -0.0323944091796875, -0.030029296875, -0.0276641845703125, -0.025299072265625, -0.0229339599609375, -0.02056884765625, -0.0182037353515625, -0.015838623046875, -0.0134735107421875, -0.0111083984375, -0.0087432861328125, -0.006378173828125, -0.0040130615234375, -0.00164794921875, 0.0007171630859375, 0.003082275390625, 0.0054473876953125, 0.0078125, 0.0101776123046875, 0.012542724609375, 0.0149078369140625, 0.01727294921875, 0.0196380615234375, 0.022003173828125, 0.0243682861328125, 0.0267333984375, 0.0290985107421875, 0.031463623046875, 0.0338287353515625, 0.03619384765625, 0.0385589599609375, 0.040924072265625, 0.0432891845703125, 0.045654296875, 0.0480194091796875, 0.050384521484375, 0.0527496337890625, 0.05511474609375, 0.0574798583984375, 0.059844970703125, 0.0622100830078125, 0.0645751953125, 0.0669403076171875, 0.069305419921875, 0.0716705322265625, 0.07403564453125, 0.0764007568359375, 0.078765869140625, 0.0811309814453125, 0.08349609375]}, "parameters/model.layers.8.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 2.0, 1.0, 0.0, 1.0, 1.0, 2.0, 1.0, 1.0, 2.0, 1.0, 2.0, 2.0, 3.0, 0.0, 8.0, 2.0, 7.0, 9.0, 16.0, 27.0, 48.0, 78.0, 95.0, 280.0, 877.0, 7699.0, 125453.0, 1338804.0, 6131594.0, 7070888.0, 1888052.0, 199190.0, 12338.0, 1108.0, 294.0, 124.0, 78.0, 43.0, 20.0, 13.0, 7.0, 5.0, 7.0, 2.0, 3.0, 2.0, 2.0, 4.0, 2.0, 1.0, 0.0, 2.0, 3.0, 2.0, 1.0, 2.0, 1.0, 0.0, 1.0, 0.0, 2.0], "bins": [-0.35546875, -0.344390869140625, -0.33331298828125, -0.322235107421875, -0.3111572265625, -0.300079345703125, -0.28900146484375, -0.277923583984375, -0.266845703125, -0.255767822265625, -0.24468994140625, -0.233612060546875, -0.2225341796875, -0.211456298828125, -0.20037841796875, -0.189300537109375, -0.17822265625, -0.167144775390625, -0.15606689453125, -0.144989013671875, -0.1339111328125, -0.122833251953125, -0.11175537109375, -0.100677490234375, -0.089599609375, -0.078521728515625, -0.06744384765625, -0.056365966796875, -0.0452880859375, -0.034210205078125, -0.02313232421875, -0.012054443359375, -0.0009765625, 0.010101318359375, 0.02117919921875, 0.032257080078125, 0.0433349609375, 0.054412841796875, 0.06549072265625, 0.076568603515625, 0.087646484375, 0.098724365234375, 0.10980224609375, 0.120880126953125, 0.1319580078125, 0.143035888671875, 0.15411376953125, 0.165191650390625, 0.17626953125, 0.187347412109375, 0.19842529296875, 0.209503173828125, 0.2205810546875, 0.231658935546875, 0.24273681640625, 0.253814697265625, 0.264892578125, 0.275970458984375, 0.28704833984375, 0.298126220703125, 0.3092041015625, 0.320281982421875, 0.33135986328125, 0.342437744140625, 0.353515625]}, "parameters/model.layers.8.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 3.0, 2.0, 1.0, 3.0, 2.0, 10.0, 13.0, 9.0, 20.0, 27.0, 53.0, 76.0, 108.0, 179.0, 295.0, 779.0, 1905.0, 6744.0, 30392.0, 150410.0, 728497.0, 3580109.0, 13417426.0, 16777216.0, 16777216.0, 16777216.0, 7241353.0, 1664404.0, 326018.0, 69088.0, 13784.0, 3472.0, 1031.0, 451.0, 206.0, 149.0, 84.0, 66.0, 35.0, 31.0, 14.0, 11.0, 12.0, 2.0, 7.0, 8.0, 1.0, 3.0, 1.0, 0.0, 1.0, 1.0, 2.0, 1.0, 1.0], "bins": [-0.3125, -0.3031005859375, -0.293701171875, -0.2843017578125, -0.27490234375, -0.2655029296875, -0.256103515625, -0.2467041015625, -0.2373046875, -0.2279052734375, -0.218505859375, -0.2091064453125, -0.19970703125, -0.1903076171875, -0.180908203125, -0.1715087890625, -0.162109375, -0.1527099609375, -0.143310546875, -0.1339111328125, -0.12451171875, -0.1151123046875, -0.105712890625, -0.0963134765625, -0.0869140625, -0.0775146484375, -0.068115234375, -0.0587158203125, -0.04931640625, -0.0399169921875, -0.030517578125, -0.0211181640625, -0.01171875, -0.0023193359375, 0.007080078125, 0.0164794921875, 0.02587890625, 0.0352783203125, 0.044677734375, 0.0540771484375, 0.0634765625, 0.0728759765625, 0.082275390625, 0.0916748046875, 0.10107421875, 0.1104736328125, 0.119873046875, 0.1292724609375, 0.138671875, 0.1480712890625, 0.157470703125, 0.1668701171875, 0.17626953125, 0.1856689453125, 0.195068359375, 0.2044677734375, 0.2138671875, 0.2232666015625, 0.232666015625, 0.2420654296875, 0.25146484375, 0.2608642578125, 0.270263671875, 0.2796630859375, 0.2890625]}, "parameters/model.layers.8.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 2.0, 0.0, 1.0, 1.0, 3.0, 3.0, 7.0, 3.0, 1.0, 4.0, 8.0, 28.0, 35.0, 72.0, 104.0, 233.0, 497.0, 2433.0, 331815.0, 12527768.0, 16777216.0, 7944014.0, 139273.0, 1557.0, 449.0, 207.0, 111.0, 62.0, 25.0, 19.0, 11.0, 11.0, 8.0, 3.0, 3.0, 3.0, 2.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.5546875, -0.53582763671875, -0.5169677734375, -0.49810791015625, -0.479248046875, -0.46038818359375, -0.4415283203125, -0.42266845703125, -0.40380859375, -0.38494873046875, -0.3660888671875, -0.34722900390625, -0.328369140625, -0.30950927734375, -0.2906494140625, -0.27178955078125, -0.2529296875, -0.23406982421875, -0.2152099609375, -0.19635009765625, -0.177490234375, -0.15863037109375, -0.1397705078125, -0.12091064453125, -0.10205078125, -0.08319091796875, -0.0643310546875, -0.04547119140625, -0.026611328125, -0.00775146484375, 0.0111083984375, 0.02996826171875, 0.048828125, 0.06768798828125, 0.0865478515625, 0.10540771484375, 0.124267578125, 0.14312744140625, 0.1619873046875, 0.18084716796875, 0.19970703125, 0.21856689453125, 0.2374267578125, 0.25628662109375, 0.275146484375, 0.29400634765625, 0.3128662109375, 0.33172607421875, 0.3505859375, 0.36944580078125, 0.3883056640625, 0.40716552734375, 0.426025390625, 0.44488525390625, 0.4637451171875, 0.48260498046875, 0.50146484375, 0.52032470703125, 0.5391845703125, 0.55804443359375, 0.576904296875, 0.59576416015625, 0.6146240234375, 0.63348388671875, 0.65234375]}, "parameters/model.layers.8.input_layernorm.weight": {"_type": "histogram", "values": [4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 3.0, 8.0, 11.0, 27.0, 53.0, 137.0, 297.0, 522.0, 713.0, 688.0, 537.0, 417.0, 254.0, 100.0, 92.0, 48.0, 32.0, 41.0, 18.0, 9.0, 10.0, 8.0, 10.0, 2.0, 4.0, 3.0, 5.0, 5.0, 4.0, 2.0, 2.0, 1.0, 1.0, 1.0, 1.0, 2.0, 5.0, 1.0, 3.0, 0.0, 2.0, 1.0, 0.0, 3.0, 0.0, 2.0, 1.0, 0.0, 2.0, 0.0, 0.0, 1.0], "bins": [0.12353515625, 0.13649749755859375, 0.1494598388671875, 0.16242218017578125, 0.175384521484375, 0.18834686279296875, 0.2013092041015625, 0.21427154541015625, 0.22723388671875, 0.24019622802734375, 0.2531585693359375, 0.26612091064453125, 0.279083251953125, 0.29204559326171875, 0.3050079345703125, 0.31797027587890625, 0.3309326171875, 0.34389495849609375, 0.3568572998046875, 0.36981964111328125, 0.382781982421875, 0.39574432373046875, 0.4087066650390625, 0.42166900634765625, 0.43463134765625, 0.44759368896484375, 0.4605560302734375, 0.47351837158203125, 0.486480712890625, 0.49944305419921875, 0.5124053955078125, 0.5253677368164062, 0.538330078125, 0.5512924194335938, 0.5642547607421875, 0.5772171020507812, 0.590179443359375, 0.6031417846679688, 0.6161041259765625, 0.6290664672851562, 0.64202880859375, 0.6549911499023438, 0.6679534912109375, 0.6809158325195312, 0.693878173828125, 0.7068405151367188, 0.7198028564453125, 0.7327651977539062, 0.7457275390625, 0.7586898803710938, 0.7716522216796875, 0.7846145629882812, 0.797576904296875, 0.8105392456054688, 0.8235015869140625, 0.8364639282226562, 0.84942626953125, 0.8623886108398438, 0.8753509521484375, 0.8883132934570312, 0.901275634765625, 0.9142379760742188, 0.9272003173828125, 0.9401626586914062, 0.953125]}, "parameters/model.layers.8.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 5.0, 3.0, 9.0, 9.0, 19.0, 23.0, 29.0, 27.0, 17.0, 29.0, 26.0, 37.0, 24.0, 35.0, 31.0, 40.0, 54.0, 69.0, 84.0, 158.0, 208.0, 358.0, 554.0, 861.0, 800.0, 428.0, 127.0, 20.0, 3.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0], "bins": [-0.00010824203491210938, 0.005874894559383392, 0.011858031153678894, 0.017841167747974396, 0.023824304342269897, 0.0298074409365654, 0.0357905775308609, 0.0417737141251564, 0.047756850719451904, 0.053739987313747406, 0.05972312390804291, 0.06570626050233841, 0.07168939709663391, 0.07767253369092941, 0.08365567028522491, 0.08963880687952042, 0.09562194347381592, 0.10160508006811142, 0.10758821666240692, 0.11357135325670242, 0.11955448985099792, 0.12553763389587402, 0.13152076303958893, 0.13750389218330383, 0.14348703622817993, 0.14947018027305603, 0.15545330941677094, 0.16143643856048584, 0.16741958260536194, 0.17340272665023804, 0.17938585579395294, 0.18536898493766785, 0.19135212898254395, 0.19733527302742004, 0.20331840217113495, 0.20930153131484985, 0.21528467535972595, 0.22126781940460205, 0.22725094854831696, 0.23323407769203186, 0.23921722173690796, 0.24520036578178406, 0.25118350982666016, 0.25716662406921387, 0.26314976811408997, 0.26913291215896606, 0.2751160264015198, 0.2810991704463959, 0.287082314491272, 0.29306545853614807, 0.29904860258102417, 0.3050317168235779, 0.311014860868454, 0.3169980049133301, 0.3229811191558838, 0.3289642632007599, 0.334947407245636, 0.3409305512905121, 0.3469136953353882, 0.3528968095779419, 0.358879953622818, 0.3648630976676941, 0.3708462119102478, 0.3768293559551239, 0.3828125]}, "parameters/model.layers.9.self_attn.q_proj.weight": {"_type": "histogram", "values": [4.0, 3.0, 1.0, 1.0, 2.0, 4.0, 5.0, 7.0, 22.0, 37.0, 47.0, 101.0, 198.0, 367.0, 554.0, 1100.0, 2056.0, 3652.0, 7379.0, 14153.0, 29958.0, 62227.0, 124384.0, 235151.0, 494906.0, 939050.0, 1668512.0, 2610124.0, 3270633.0, 2915198.0, 2012038.0, 1132780.0, 630157.0, 308952.0, 158185.0, 79501.0, 38811.0, 17831.0, 9096.0, 4536.0, 2399.0, 1428.0, 703.0, 414.0, 229.0, 144.0, 74.0, 33.0, 26.0, 11.0, 12.0, 8.0, 4.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 1.0], "bins": [-0.2021484375, -0.195098876953125, -0.18804931640625, -0.180999755859375, -0.1739501953125, -0.166900634765625, -0.15985107421875, -0.152801513671875, -0.145751953125, -0.138702392578125, -0.13165283203125, -0.124603271484375, -0.1175537109375, -0.110504150390625, -0.10345458984375, -0.096405029296875, -0.08935546875, -0.082305908203125, -0.07525634765625, -0.068206787109375, -0.0611572265625, -0.054107666015625, -0.04705810546875, -0.040008544921875, -0.032958984375, -0.025909423828125, -0.01885986328125, -0.011810302734375, -0.0047607421875, 0.002288818359375, 0.00933837890625, 0.016387939453125, 0.0234375, 0.030487060546875, 0.03753662109375, 0.044586181640625, 0.0516357421875, 0.058685302734375, 0.06573486328125, 0.072784423828125, 0.079833984375, 0.086883544921875, 0.09393310546875, 0.100982666015625, 0.1080322265625, 0.115081787109375, 0.12213134765625, 0.129180908203125, 0.13623046875, 0.143280029296875, 0.15032958984375, 0.157379150390625, 0.1644287109375, 0.171478271484375, 0.17852783203125, 0.185577392578125, 0.192626953125, 0.199676513671875, 0.20672607421875, 0.213775634765625, 0.2208251953125, 0.227874755859375, 0.23492431640625, 0.241973876953125, 0.2490234375]}, "parameters/model.layers.9.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 4.0, 6.0, 1.0, 9.0, 9.0, 10.0, 18.0, 36.0, 68.0, 107.0, 194.0, 409.0, 775.0, 1700.0, 3405.0, 6939.0, 13431.0, 24878.0, 48922.0, 89692.0, 153967.0, 267212.0, 424813.0, 624140.0, 729446.0, 651668.0, 465030.0, 296978.0, 172601.0, 101411.0, 56115.0, 28707.0, 15642.0, 7988.0, 4140.0, 1951.0, 937.0, 427.0, 221.0, 128.0, 57.0, 34.0, 21.0, 15.0, 7.0, 8.0, 9.0, 3.0, 2.0, 4.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.26953125, -0.259765625, -0.25, -0.240234375, -0.23046875, -0.220703125, -0.2109375, -0.201171875, -0.19140625, -0.181640625, -0.171875, -0.162109375, -0.15234375, -0.142578125, -0.1328125, -0.123046875, -0.11328125, -0.103515625, -0.09375, -0.083984375, -0.07421875, -0.064453125, -0.0546875, -0.044921875, -0.03515625, -0.025390625, -0.015625, -0.005859375, 0.00390625, 0.013671875, 0.0234375, 0.033203125, 0.04296875, 0.052734375, 0.0625, 0.072265625, 0.08203125, 0.091796875, 0.1015625, 0.111328125, 0.12109375, 0.130859375, 0.140625, 0.150390625, 0.16015625, 0.169921875, 0.1796875, 0.189453125, 0.19921875, 0.208984375, 0.21875, 0.228515625, 0.23828125, 0.248046875, 0.2578125, 0.267578125, 0.27734375, 0.287109375, 0.296875, 0.306640625, 0.31640625, 0.326171875, 0.3359375, 0.345703125, 0.35546875]}, "parameters/model.layers.9.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 3.0, 4.0, 4.0, 5.0, 22.0, 35.0, 84.0, 112.0, 200.0, 300.0, 499.0, 852.0, 1353.0, 2143.0, 2845.0, 5206.0, 7479.0, 12648.0, 21526.0, 33498.0, 58970.0, 94345.0, 154339.0, 232564.0, 326118.0, 436064.0, 521019.0, 550378.0, 494028.0, 402688.0, 297301.0, 202549.0, 129297.0, 80281.0, 49708.0, 28359.0, 18596.0, 10681.0, 6726.0, 4006.0, 2731.0, 1738.0, 1083.0, 791.0, 467.0, 282.0, 132.0, 113.0, 63.0, 30.0, 22.0, 8.0, 0.0, 3.0, 0.0, 0.0, 1.0], "bins": [-0.08984375, -0.08722686767578125, -0.0846099853515625, -0.08199310302734375, -0.079376220703125, -0.07675933837890625, -0.0741424560546875, -0.07152557373046875, -0.06890869140625, -0.06629180908203125, -0.0636749267578125, -0.06105804443359375, -0.058441162109375, -0.05582427978515625, -0.0532073974609375, -0.05059051513671875, -0.0479736328125, -0.04535675048828125, -0.0427398681640625, -0.04012298583984375, -0.037506103515625, -0.03488922119140625, -0.0322723388671875, -0.02965545654296875, -0.02703857421875, -0.02442169189453125, -0.0218048095703125, -0.01918792724609375, -0.016571044921875, -0.01395416259765625, -0.0113372802734375, -0.00872039794921875, -0.006103515625, -0.00348663330078125, -0.0008697509765625, 0.00174713134765625, 0.004364013671875, 0.00698089599609375, 0.0095977783203125, 0.01221466064453125, 0.01483154296875, 0.01744842529296875, 0.0200653076171875, 0.02268218994140625, 0.025299072265625, 0.02791595458984375, 0.0305328369140625, 0.03314971923828125, 0.0357666015625, 0.03838348388671875, 0.0410003662109375, 0.04361724853515625, 0.046234130859375, 0.04885101318359375, 0.0514678955078125, 0.05408477783203125, 0.05670166015625, 0.05931854248046875, 0.0619354248046875, 0.06455230712890625, 0.067169189453125, 0.06978607177734375, 0.0724029541015625, 0.07501983642578125, 0.07763671875]}, "parameters/model.layers.9.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 3.0, 4.0, 2.0, 1.0, 5.0, 5.0, 8.0, 9.0, 9.0, 10.0, 22.0, 29.0, 32.0, 58.0, 72.0, 229.0, 1052.0, 7946.0, 54576.0, 380472.0, 1989406.0, 5673344.0, 5919141.0, 2223529.0, 450314.0, 66084.0, 9019.0, 1322.0, 214.0, 98.0, 55.0, 40.0, 28.0, 16.0, 11.0, 8.0, 5.0, 8.0, 4.0, 3.0, 6.0, 1.0, 1.0, 2.0, 3.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.310546875, -0.3011474609375, -0.291748046875, -0.2823486328125, -0.27294921875, -0.2635498046875, -0.254150390625, -0.2447509765625, -0.2353515625, -0.2259521484375, -0.216552734375, -0.2071533203125, -0.19775390625, -0.1883544921875, -0.178955078125, -0.1695556640625, -0.16015625, -0.1507568359375, -0.141357421875, -0.1319580078125, -0.12255859375, -0.1131591796875, -0.103759765625, -0.0943603515625, -0.0849609375, -0.0755615234375, -0.066162109375, -0.0567626953125, -0.04736328125, -0.0379638671875, -0.028564453125, -0.0191650390625, -0.009765625, -0.0003662109375, 0.009033203125, 0.0184326171875, 0.02783203125, 0.0372314453125, 0.046630859375, 0.0560302734375, 0.0654296875, 0.0748291015625, 0.084228515625, 0.0936279296875, 0.10302734375, 0.1124267578125, 0.121826171875, 0.1312255859375, 0.140625, 0.1500244140625, 0.159423828125, 0.1688232421875, 0.17822265625, 0.1876220703125, 0.197021484375, 0.2064208984375, 0.2158203125, 0.2252197265625, 0.234619140625, 0.2440185546875, 0.25341796875, 0.2628173828125, 0.272216796875, 0.2816162109375, 0.291015625]}, "parameters/model.layers.9.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 3.0, 0.0, 3.0, 4.0, 7.0, 6.0, 5.0, 11.0, 20.0, 39.0, 55.0, 106.0, 159.0, 352.0, 667.0, 1667.0, 5906.0, 27847.0, 158908.0, 1039453.0, 6448588.0, 16777216.0, 16777216.0, 16777216.0, 8065477.0, 1360760.0, 200563.0, 34866.0, 6669.0, 2056.0, 742.0, 338.0, 174.0, 114.0, 36.0, 46.0, 32.0, 12.0, 11.0, 5.0, 7.0, 4.0, 3.0, 1.0, 1.0, 0.0, 0.0, 1.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.291015625, -0.279632568359375, -0.26824951171875, -0.256866455078125, -0.2454833984375, -0.234100341796875, -0.22271728515625, -0.211334228515625, -0.199951171875, -0.188568115234375, -0.17718505859375, -0.165802001953125, -0.1544189453125, -0.143035888671875, -0.13165283203125, -0.120269775390625, -0.10888671875, -0.097503662109375, -0.08612060546875, -0.074737548828125, -0.0633544921875, -0.051971435546875, -0.04058837890625, -0.029205322265625, -0.017822265625, -0.006439208984375, 0.00494384765625, 0.016326904296875, 0.0277099609375, 0.039093017578125, 0.05047607421875, 0.061859130859375, 0.0732421875, 0.084625244140625, 0.09600830078125, 0.107391357421875, 0.1187744140625, 0.130157470703125, 0.14154052734375, 0.152923583984375, 0.164306640625, 0.175689697265625, 0.18707275390625, 0.198455810546875, 0.2098388671875, 0.221221923828125, 0.23260498046875, 0.243988037109375, 0.25537109375, 0.266754150390625, 0.27813720703125, 0.289520263671875, 0.3009033203125, 0.312286376953125, 0.32366943359375, 0.335052490234375, 0.346435546875, 0.357818603515625, 0.36920166015625, 0.380584716796875, 0.3919677734375, 0.403350830078125, 0.41473388671875, 0.426116943359375, 0.4375]}, "parameters/model.layers.9.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 3.0, 1.0, 4.0, 3.0, 5.0, 9.0, 14.0, 24.0, 34.0, 54.0, 67.0, 162.0, 332.0, 1046.0, 57195.0, 4576188.0, 16777216.0, 16777216.0, 838945.0, 5796.0, 507.0, 238.0, 132.0, 60.0, 33.0, 16.0, 18.0, 5.0, 8.0, 1.0, 5.0, 3.0, 1.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.6328125, -0.61407470703125, -0.5953369140625, -0.57659912109375, -0.557861328125, -0.53912353515625, -0.5203857421875, -0.50164794921875, -0.48291015625, -0.46417236328125, -0.4454345703125, -0.42669677734375, -0.407958984375, -0.38922119140625, -0.3704833984375, -0.35174560546875, -0.3330078125, -0.31427001953125, -0.2955322265625, -0.27679443359375, -0.258056640625, -0.23931884765625, -0.2205810546875, -0.20184326171875, -0.18310546875, -0.16436767578125, -0.1456298828125, -0.12689208984375, -0.108154296875, -0.08941650390625, -0.0706787109375, -0.05194091796875, -0.033203125, -0.01446533203125, 0.0042724609375, 0.02301025390625, 0.041748046875, 0.06048583984375, 0.0792236328125, 0.09796142578125, 0.11669921875, 0.13543701171875, 0.1541748046875, 0.17291259765625, 0.191650390625, 0.21038818359375, 0.2291259765625, 0.24786376953125, 0.2666015625, 0.28533935546875, 0.3040771484375, 0.32281494140625, 0.341552734375, 0.36029052734375, 0.3790283203125, 0.39776611328125, 0.41650390625, 0.43524169921875, 0.4539794921875, 0.47271728515625, 0.491455078125, 0.51019287109375, 0.5289306640625, 0.54766845703125, 0.56640625]}, "parameters/model.layers.9.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 1.0, 0.0, 1.0, 1.0, 2.0, 3.0, 4.0, 10.0, 14.0, 14.0, 19.0, 31.0, 55.0, 70.0, 119.0, 188.0, 288.0, 454.0, 570.0, 508.0, 583.0, 410.0, 257.0, 177.0, 123.0, 44.0, 50.0, 12.0, 14.0, 5.0, 6.0, 6.0, 5.0, 5.0, 5.0, 2.0, 2.0, 3.0, 3.0, 0.0, 0.0, 1.0, 1.0, 2.0, 1.0, 1.0, 1.0, 0.0, 5.0, 3.0, 1.0, 1.0, 4.0, 1.0, 1.0, 1.0, 2.0, 0.0, 0.0, 2.0, 0.0, 1.0], "bins": [0.1298828125, 0.1428680419921875, 0.155853271484375, 0.1688385009765625, 0.18182373046875, 0.1948089599609375, 0.207794189453125, 0.2207794189453125, 0.2337646484375, 0.2467498779296875, 0.259735107421875, 0.2727203369140625, 0.28570556640625, 0.2986907958984375, 0.311676025390625, 0.3246612548828125, 0.337646484375, 0.3506317138671875, 0.363616943359375, 0.3766021728515625, 0.38958740234375, 0.4025726318359375, 0.415557861328125, 0.4285430908203125, 0.4415283203125, 0.4545135498046875, 0.467498779296875, 0.4804840087890625, 0.49346923828125, 0.5064544677734375, 0.519439697265625, 0.5324249267578125, 0.54541015625, 0.5583953857421875, 0.571380615234375, 0.5843658447265625, 0.59735107421875, 0.6103363037109375, 0.623321533203125, 0.6363067626953125, 0.6492919921875, 0.6622772216796875, 0.675262451171875, 0.6882476806640625, 0.70123291015625, 0.7142181396484375, 0.727203369140625, 0.7401885986328125, 0.753173828125, 0.7661590576171875, 0.779144287109375, 0.7921295166015625, 0.80511474609375, 0.8180999755859375, 0.831085205078125, 0.8440704345703125, 0.8570556640625, 0.8700408935546875, 0.883026123046875, 0.8960113525390625, 0.90899658203125, 0.9219818115234375, 0.934967041015625, 0.9479522705078125, 0.9609375]}, "parameters/model.layers.9.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 4.0, 7.0, 15.0, 15.0, 28.0, 33.0, 22.0, 36.0, 18.0, 36.0, 22.0, 31.0, 32.0, 36.0, 41.0, 24.0, 77.0, 73.0, 95.0, 154.0, 149.0, 320.0, 487.0, 758.0, 525.0, 638.0, 292.0, 95.0, 20.0, 0.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.072265625, 0.07769775390625, 0.0831298828125, 0.08856201171875, 0.093994140625, 0.09942626953125, 0.1048583984375, 0.11029052734375, 0.11572265625, 0.12115478515625, 0.1265869140625, 0.13201904296875, 0.137451171875, 0.14288330078125, 0.1483154296875, 0.15374755859375, 0.1591796875, 0.16461181640625, 0.1700439453125, 0.17547607421875, 0.180908203125, 0.18634033203125, 0.1917724609375, 0.19720458984375, 0.20263671875, 0.20806884765625, 0.2135009765625, 0.21893310546875, 0.224365234375, 0.22979736328125, 0.2352294921875, 0.24066162109375, 0.24609375, 0.25152587890625, 0.2569580078125, 0.26239013671875, 0.267822265625, 0.27325439453125, 0.2786865234375, 0.28411865234375, 0.28955078125, 0.29498291015625, 0.3004150390625, 0.30584716796875, 0.311279296875, 0.31671142578125, 0.3221435546875, 0.32757568359375, 0.3330078125, 0.33843994140625, 0.3438720703125, 0.34930419921875, 0.354736328125, 0.36016845703125, 0.3656005859375, 0.37103271484375, 0.37646484375, 0.38189697265625, 0.3873291015625, 0.39276123046875, 0.398193359375, 0.40362548828125, 0.4090576171875, 0.41448974609375, 0.419921875]}, "parameters/model.layers.10.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 1.0, 2.0, 5.0, 4.0, 13.0, 22.0, 23.0, 29.0, 28.0, 46.0, 70.0, 72.0, 134.0, 201.0, 324.0, 801.0, 1926.0, 5073.0, 14916.0, 42914.0, 124193.0, 342417.0, 947851.0, 2363474.0, 4703499.0, 4572118.0, 2270738.0, 886967.0, 316952.0, 120274.0, 39621.0, 14037.0, 5006.0, 1713.0, 772.0, 347.0, 185.0, 122.0, 85.0, 54.0, 51.0, 36.0, 20.0, 20.0, 15.0, 12.0, 7.0, 2.0, 4.0, 5.0, 2.0, 2.0, 1.0, 2.0, 1.0, 1.0], "bins": [-0.365234375, -0.354156494140625, -0.34307861328125, -0.332000732421875, -0.3209228515625, -0.309844970703125, -0.29876708984375, -0.287689208984375, -0.276611328125, -0.265533447265625, -0.25445556640625, -0.243377685546875, -0.2322998046875, -0.221221923828125, -0.21014404296875, -0.199066162109375, -0.18798828125, -0.176910400390625, -0.16583251953125, -0.154754638671875, -0.1436767578125, -0.132598876953125, -0.12152099609375, -0.110443115234375, -0.099365234375, -0.088287353515625, -0.07720947265625, -0.066131591796875, -0.0550537109375, -0.043975830078125, -0.03289794921875, -0.021820068359375, -0.0107421875, 0.000335693359375, 0.01141357421875, 0.022491455078125, 0.0335693359375, 0.044647216796875, 0.05572509765625, 0.066802978515625, 0.077880859375, 0.088958740234375, 0.10003662109375, 0.111114501953125, 0.1221923828125, 0.133270263671875, 0.14434814453125, 0.155426025390625, 0.16650390625, 0.177581787109375, 0.18865966796875, 0.199737548828125, 0.2108154296875, 0.221893310546875, 0.23297119140625, 0.244049072265625, 0.255126953125, 0.266204833984375, 0.27728271484375, 0.288360595703125, 0.2994384765625, 0.310516357421875, 0.32159423828125, 0.332672119140625, 0.34375]}, "parameters/model.layers.10.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 2.0, 2.0, 5.0, 12.0, 7.0, 11.0, 12.0, 29.0, 37.0, 38.0, 65.0, 78.0, 102.0, 184.0, 247.0, 436.0, 1032.0, 2091.0, 5114.0, 10375.0, 24551.0, 52822.0, 114902.0, 219272.0, 431201.0, 710215.0, 897472.0, 754907.0, 474315.0, 251605.0, 130380.0, 60623.0, 28365.0, 12930.0, 5765.0, 2445.0, 1173.0, 555.0, 279.0, 173.0, 134.0, 78.0, 65.0, 57.0, 36.0, 30.0, 24.0, 11.0, 11.0, 4.0, 9.0, 2.0, 4.0, 1.0, 1.0, 2.0], "bins": [-0.41796875, -0.4058837890625, -0.393798828125, -0.3817138671875, -0.36962890625, -0.3575439453125, -0.345458984375, -0.3333740234375, -0.3212890625, -0.3092041015625, -0.297119140625, -0.2850341796875, -0.27294921875, -0.2608642578125, -0.248779296875, -0.2366943359375, -0.224609375, -0.2125244140625, -0.200439453125, -0.1883544921875, -0.17626953125, -0.1641845703125, -0.152099609375, -0.1400146484375, -0.1279296875, -0.1158447265625, -0.103759765625, -0.0916748046875, -0.07958984375, -0.0675048828125, -0.055419921875, -0.0433349609375, -0.03125, -0.0191650390625, -0.007080078125, 0.0050048828125, 0.01708984375, 0.0291748046875, 0.041259765625, 0.0533447265625, 0.0654296875, 0.0775146484375, 0.089599609375, 0.1016845703125, 0.11376953125, 0.1258544921875, 0.137939453125, 0.1500244140625, 0.162109375, 0.1741943359375, 0.186279296875, 0.1983642578125, 0.21044921875, 0.2225341796875, 0.234619140625, 0.2467041015625, 0.2587890625, 0.2708740234375, 0.282958984375, 0.2950439453125, 0.30712890625, 0.3192138671875, 0.331298828125, 0.3433837890625, 0.35546875]}, "parameters/model.layers.10.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 2.0, 4.0, 4.0, 13.0, 28.0, 33.0, 59.0, 126.0, 180.0, 312.0, 498.0, 701.0, 1233.0, 1987.0, 3072.0, 4527.0, 7058.0, 10338.0, 15622.0, 23703.0, 34790.0, 53310.0, 76162.0, 107228.0, 145951.0, 197687.0, 254871.0, 315734.0, 369170.0, 413907.0, 418533.0, 385767.0, 327649.0, 276338.0, 209771.0, 158929.0, 117640.0, 86945.0, 57801.0, 38304.0, 26324.0, 17754.0, 11591.0, 8005.0, 5080.0, 3719.0, 2086.0, 1395.0, 904.0, 523.0, 357.0, 220.0, 153.0, 85.0, 49.0, 36.0, 11.0, 11.0, 9.0, 1.0, 1.0], "bins": [-0.056884765625, -0.05516815185546875, -0.0534515380859375, -0.05173492431640625, -0.050018310546875, -0.04830169677734375, -0.0465850830078125, -0.04486846923828125, -0.04315185546875, -0.04143524169921875, -0.0397186279296875, -0.03800201416015625, -0.036285400390625, -0.03456878662109375, -0.0328521728515625, -0.03113555908203125, -0.0294189453125, -0.02770233154296875, -0.0259857177734375, -0.02426910400390625, -0.022552490234375, -0.02083587646484375, -0.0191192626953125, -0.01740264892578125, -0.01568603515625, -0.01396942138671875, -0.0122528076171875, -0.01053619384765625, -0.008819580078125, -0.00710296630859375, -0.0053863525390625, -0.00366973876953125, -0.001953125, -0.00023651123046875, 0.0014801025390625, 0.00319671630859375, 0.004913330078125, 0.00662994384765625, 0.0083465576171875, 0.01006317138671875, 0.01177978515625, 0.01349639892578125, 0.0152130126953125, 0.01692962646484375, 0.018646240234375, 0.02036285400390625, 0.0220794677734375, 0.02379608154296875, 0.0255126953125, 0.02722930908203125, 0.0289459228515625, 0.03066253662109375, 0.032379150390625, 0.03409576416015625, 0.0358123779296875, 0.03752899169921875, 0.03924560546875, 0.04096221923828125, 0.0426788330078125, 0.04439544677734375, 0.046112060546875, 0.04782867431640625, 0.0495452880859375, 0.05126190185546875, 0.052978515625]}, "parameters/model.layers.10.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 1.0, 0.0, 1.0, 1.0, 3.0, 6.0, 9.0, 7.0, 2.0, 4.0, 6.0, 5.0, 17.0, 16.0, 23.0, 58.0, 127.0, 927.0, 17740.0, 376464.0, 3841503.0, 9097501.0, 3150831.0, 278537.0, 12469.0, 717.0, 103.0, 46.0, 25.0, 10.0, 10.0, 6.0, 8.0, 4.0, 4.0, 3.0, 3.0, 1.0, 2.0, 2.0, 3.0, 2.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.359375, -0.34674072265625, -0.3341064453125, -0.32147216796875, -0.308837890625, -0.29620361328125, -0.2835693359375, -0.27093505859375, -0.25830078125, -0.24566650390625, -0.2330322265625, -0.22039794921875, -0.207763671875, -0.19512939453125, -0.1824951171875, -0.16986083984375, -0.1572265625, -0.14459228515625, -0.1319580078125, -0.11932373046875, -0.106689453125, -0.09405517578125, -0.0814208984375, -0.06878662109375, -0.05615234375, -0.04351806640625, -0.0308837890625, -0.01824951171875, -0.005615234375, 0.00701904296875, 0.0196533203125, 0.03228759765625, 0.044921875, 0.05755615234375, 0.0701904296875, 0.08282470703125, 0.095458984375, 0.10809326171875, 0.1207275390625, 0.13336181640625, 0.14599609375, 0.15863037109375, 0.1712646484375, 0.18389892578125, 0.196533203125, 0.20916748046875, 0.2218017578125, 0.23443603515625, 0.2470703125, 0.25970458984375, 0.2723388671875, 0.28497314453125, 0.297607421875, 0.31024169921875, 0.3228759765625, 0.33551025390625, 0.34814453125, 0.36077880859375, 0.3734130859375, 0.38604736328125, 0.398681640625, 0.41131591796875, 0.4239501953125, 0.43658447265625, 0.44921875]}, "parameters/model.layers.10.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 2.0, 1.0, 1.0, 7.0, 1.0, 3.0, 7.0, 7.0, 6.0, 11.0, 12.0, 27.0, 31.0, 44.0, 64.0, 109.0, 149.0, 366.0, 778.0, 2085.0, 6499.0, 25986.0, 113812.0, 534168.0, 2652973.0, 10097983.0, 16777216.0, 16777216.0, 16777216.0, 11137519.0, 3037088.0, 622291.0, 132776.0, 29694.0, 7378.0, 2273.0, 791.0, 379.0, 175.0, 110.0, 70.0, 45.0, 27.0, 20.0, 17.0, 11.0, 12.0, 9.0, 2.0, 6.0, 2.0, 3.0, 2.0, 2.0, 1.0, 2.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.2578125, -0.248779296875, -0.23974609375, -0.230712890625, -0.2216796875, -0.212646484375, -0.20361328125, -0.194580078125, -0.185546875, -0.176513671875, -0.16748046875, -0.158447265625, -0.1494140625, -0.140380859375, -0.13134765625, -0.122314453125, -0.11328125, -0.104248046875, -0.09521484375, -0.086181640625, -0.0771484375, -0.068115234375, -0.05908203125, -0.050048828125, -0.041015625, -0.031982421875, -0.02294921875, -0.013916015625, -0.0048828125, 0.004150390625, 0.01318359375, 0.022216796875, 0.03125, 0.040283203125, 0.04931640625, 0.058349609375, 0.0673828125, 0.076416015625, 0.08544921875, 0.094482421875, 0.103515625, 0.112548828125, 0.12158203125, 0.130615234375, 0.1396484375, 0.148681640625, 0.15771484375, 0.166748046875, 0.17578125, 0.184814453125, 0.19384765625, 0.202880859375, 0.2119140625, 0.220947265625, 0.22998046875, 0.239013671875, 0.248046875, 0.257080078125, 0.26611328125, 0.275146484375, 0.2841796875, 0.293212890625, 0.30224609375, 0.311279296875, 0.3203125]}, "parameters/model.layers.10.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 3.0, 1.0, 1.0, 2.0, 3.0, 7.0, 1.0, 7.0, 10.0, 10.0, 8.0, 18.0, 22.0, 57.0, 108.0, 207.0, 476.0, 8474.0, 785209.0, 14031465.0, 16777216.0, 9298242.0, 353826.0, 3483.0, 443.0, 179.0, 100.0, 55.0, 31.0, 25.0, 15.0, 14.0, 7.0, 9.0, 2.0, 0.0, 3.0, 4.0, 1.0, 1.0, 2.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0], "bins": [-0.62890625, -0.612091064453125, -0.59527587890625, -0.578460693359375, -0.5616455078125, -0.544830322265625, -0.52801513671875, -0.511199951171875, -0.494384765625, -0.477569580078125, -0.46075439453125, -0.443939208984375, -0.4271240234375, -0.410308837890625, -0.39349365234375, -0.376678466796875, -0.35986328125, -0.343048095703125, -0.32623291015625, -0.309417724609375, -0.2926025390625, -0.275787353515625, -0.25897216796875, -0.242156982421875, -0.225341796875, -0.208526611328125, -0.19171142578125, -0.174896240234375, -0.1580810546875, -0.141265869140625, -0.12445068359375, -0.107635498046875, -0.0908203125, -0.074005126953125, -0.05718994140625, -0.040374755859375, -0.0235595703125, -0.006744384765625, 0.01007080078125, 0.026885986328125, 0.043701171875, 0.060516357421875, 0.07733154296875, 0.094146728515625, 0.1109619140625, 0.127777099609375, 0.14459228515625, 0.161407470703125, 0.17822265625, 0.195037841796875, 0.21185302734375, 0.228668212890625, 0.2454833984375, 0.262298583984375, 0.27911376953125, 0.295928955078125, 0.312744140625, 0.329559326171875, 0.34637451171875, 0.363189697265625, 0.3800048828125, 0.396820068359375, 0.41363525390625, 0.430450439453125, 0.447265625]}, "parameters/model.layers.10.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 4.0, 12.0, 22.0, 46.0, 119.0, 259.0, 465.0, 559.0, 652.0, 539.0, 509.0, 296.0, 166.0, 134.0, 97.0, 48.0, 30.0, 16.0, 18.0, 16.0, 12.0, 6.0, 9.0, 8.0, 5.0, 2.0, 1.0, 0.0, 4.0, 2.0, 6.0, 4.0, 0.0, 1.0, 2.0, 2.0, 4.0, 2.0, 3.0, 2.0, 1.0, 3.0, 1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.10986328125, 0.12401580810546875, 0.1381683349609375, 0.15232086181640625, 0.166473388671875, 0.18062591552734375, 0.1947784423828125, 0.20893096923828125, 0.22308349609375, 0.23723602294921875, 0.2513885498046875, 0.26554107666015625, 0.279693603515625, 0.29384613037109375, 0.3079986572265625, 0.32215118408203125, 0.3363037109375, 0.35045623779296875, 0.3646087646484375, 0.37876129150390625, 0.392913818359375, 0.40706634521484375, 0.4212188720703125, 0.43537139892578125, 0.44952392578125, 0.46367645263671875, 0.4778289794921875, 0.49198150634765625, 0.506134033203125, 0.5202865600585938, 0.5344390869140625, 0.5485916137695312, 0.562744140625, 0.5768966674804688, 0.5910491943359375, 0.6052017211914062, 0.619354248046875, 0.6335067749023438, 0.6476593017578125, 0.6618118286132812, 0.67596435546875, 0.6901168823242188, 0.7042694091796875, 0.7184219360351562, 0.732574462890625, 0.7467269897460938, 0.7608795166015625, 0.7750320434570312, 0.7891845703125, 0.8033370971679688, 0.8174896240234375, 0.8316421508789062, 0.845794677734375, 0.8599472045898438, 0.8740997314453125, 0.8882522583007812, 0.90240478515625, 0.9165573120117188, 0.9307098388671875, 0.9448623657226562, 0.959014892578125, 0.9731674194335938, 0.9873199462890625, 1.0014724731445312, 1.015625]}, "parameters/model.layers.10.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 5.0, 9.0, 10.0, 11.0, 27.0, 21.0, 15.0, 23.0, 21.0, 28.0, 31.0, 18.0, 21.0, 27.0, 35.0, 33.0, 24.0, 50.0, 61.0, 47.0, 77.0, 130.0, 99.0, 215.0, 298.0, 275.0, 572.0, 647.0, 422.0, 434.0, 257.0, 80.0, 48.0, 12.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.07373046875, 0.07895660400390625, 0.0841827392578125, 0.08940887451171875, 0.094635009765625, 0.09986114501953125, 0.1050872802734375, 0.11031341552734375, 0.11553955078125, 0.12076568603515625, 0.1259918212890625, 0.13121795654296875, 0.136444091796875, 0.14167022705078125, 0.1468963623046875, 0.15212249755859375, 0.1573486328125, 0.16257476806640625, 0.1678009033203125, 0.17302703857421875, 0.178253173828125, 0.18347930908203125, 0.1887054443359375, 0.19393157958984375, 0.19915771484375, 0.20438385009765625, 0.2096099853515625, 0.21483612060546875, 0.220062255859375, 0.22528839111328125, 0.2305145263671875, 0.23574066162109375, 0.240966796875, 0.24619293212890625, 0.2514190673828125, 0.25664520263671875, 0.261871337890625, 0.26709747314453125, 0.2723236083984375, 0.27754974365234375, 0.28277587890625, 0.28800201416015625, 0.2932281494140625, 0.29845428466796875, 0.303680419921875, 0.30890655517578125, 0.3141326904296875, 0.31935882568359375, 0.3245849609375, 0.32981109619140625, 0.3350372314453125, 0.34026336669921875, 0.345489501953125, 0.35071563720703125, 0.3559417724609375, 0.36116790771484375, 0.36639404296875, 0.37162017822265625, 0.3768463134765625, 0.38207244873046875, 0.387298583984375, 0.39252471923828125, 0.3977508544921875, 0.40297698974609375, 0.408203125]}, "parameters/model.layers.11.self_attn.q_proj.weight": {"_type": "histogram", "values": [5.0, 5.0, 6.0, 13.0, 18.0, 47.0, 78.0, 149.0, 256.0, 390.0, 747.0, 1424.0, 2771.0, 4130.0, 7668.0, 14003.0, 26343.0, 46475.0, 76208.0, 135751.0, 217504.0, 383022.0, 623543.0, 982047.0, 1530399.0, 2078425.0, 2521742.0, 2455370.0, 1980379.0, 1389845.0, 896878.0, 573114.0, 347920.0, 201847.0, 115643.0, 71344.0, 39536.0, 23435.0, 12782.0, 6918.0, 4063.0, 2045.0, 1341.0, 680.0, 421.0, 193.0, 135.0, 68.0, 27.0, 19.0, 12.0, 14.0, 9.0, 4.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.1416015625, -0.1363372802734375, -0.131072998046875, -0.1258087158203125, -0.12054443359375, -0.1152801513671875, -0.110015869140625, -0.1047515869140625, -0.0994873046875, -0.0942230224609375, -0.088958740234375, -0.0836944580078125, -0.07843017578125, -0.0731658935546875, -0.067901611328125, -0.0626373291015625, -0.057373046875, -0.0521087646484375, -0.046844482421875, -0.0415802001953125, -0.03631591796875, -0.0310516357421875, -0.025787353515625, -0.0205230712890625, -0.0152587890625, -0.0099945068359375, -0.004730224609375, 0.0005340576171875, 0.00579833984375, 0.0110626220703125, 0.016326904296875, 0.0215911865234375, 0.02685546875, 0.0321197509765625, 0.037384033203125, 0.0426483154296875, 0.04791259765625, 0.0531768798828125, 0.058441162109375, 0.0637054443359375, 0.0689697265625, 0.0742340087890625, 0.079498291015625, 0.0847625732421875, 0.09002685546875, 0.0952911376953125, 0.100555419921875, 0.1058197021484375, 0.111083984375, 0.1163482666015625, 0.121612548828125, 0.1268768310546875, 0.13214111328125, 0.1374053955078125, 0.142669677734375, 0.1479339599609375, 0.1531982421875, 0.1584625244140625, 0.163726806640625, 0.1689910888671875, 0.17425537109375, 0.1795196533203125, 0.184783935546875, 0.1900482177734375, 0.1953125]}, "parameters/model.layers.11.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 4.0, 5.0, 3.0, 3.0, 8.0, 18.0, 17.0, 32.0, 52.0, 66.0, 108.0, 169.0, 279.0, 544.0, 1077.0, 2309.0, 4750.0, 9294.0, 18374.0, 34045.0, 64541.0, 111266.0, 183875.0, 304097.0, 456245.0, 619911.0, 685401.0, 595690.0, 424790.0, 281574.0, 170560.0, 103354.0, 58162.0, 30353.0, 16389.0, 8418.0, 4230.0, 2063.0, 1034.0, 516.0, 241.0, 157.0, 107.0, 58.0, 33.0, 28.0, 13.0, 10.0, 8.0, 6.0, 4.0, 6.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.28125, -0.272003173828125, -0.26275634765625, -0.253509521484375, -0.2442626953125, -0.235015869140625, -0.22576904296875, -0.216522216796875, -0.207275390625, -0.198028564453125, -0.18878173828125, -0.179534912109375, -0.1702880859375, -0.161041259765625, -0.15179443359375, -0.142547607421875, -0.13330078125, -0.124053955078125, -0.11480712890625, -0.105560302734375, -0.0963134765625, -0.087066650390625, -0.07781982421875, -0.068572998046875, -0.059326171875, -0.050079345703125, -0.04083251953125, -0.031585693359375, -0.0223388671875, -0.013092041015625, -0.00384521484375, 0.005401611328125, 0.0146484375, 0.023895263671875, 0.03314208984375, 0.042388916015625, 0.0516357421875, 0.060882568359375, 0.07012939453125, 0.079376220703125, 0.088623046875, 0.097869873046875, 0.10711669921875, 0.116363525390625, 0.1256103515625, 0.134857177734375, 0.14410400390625, 0.153350830078125, 0.16259765625, 0.171844482421875, 0.18109130859375, 0.190338134765625, 0.1995849609375, 0.208831787109375, 0.21807861328125, 0.227325439453125, 0.236572265625, 0.245819091796875, 0.25506591796875, 0.264312744140625, 0.2735595703125, 0.282806396484375, 0.29205322265625, 0.301300048828125, 0.310546875]}, "parameters/model.layers.11.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 4.0, 4.0, 6.0, 11.0, 25.0, 44.0, 109.0, 236.0, 538.0, 1010.0, 2177.0, 4263.0, 8725.0, 15011.0, 26128.0, 43232.0, 71882.0, 114255.0, 160711.0, 219760.0, 296228.0, 358152.0, 413687.0, 447219.0, 439171.0, 398071.0, 331372.0, 263147.0, 202742.0, 139678.0, 94861.0, 58711.0, 36095.0, 22494.0, 11790.0, 6335.0, 3217.0, 1701.0, 795.0, 378.0, 174.0, 79.0, 32.0, 18.0, 5.0, 4.0, 4.0, 3.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.064453125, -0.062488555908203125, -0.06052398681640625, -0.058559417724609375, -0.0565948486328125, -0.054630279541015625, -0.05266571044921875, -0.050701141357421875, -0.048736572265625, -0.046772003173828125, -0.04480743408203125, -0.042842864990234375, -0.0408782958984375, -0.038913726806640625, -0.03694915771484375, -0.034984588623046875, -0.03302001953125, -0.031055450439453125, -0.02909088134765625, -0.027126312255859375, -0.0251617431640625, -0.023197174072265625, -0.02123260498046875, -0.019268035888671875, -0.017303466796875, -0.015338897705078125, -0.01337432861328125, -0.011409759521484375, -0.0094451904296875, -0.007480621337890625, -0.00551605224609375, -0.003551483154296875, -0.0015869140625, 0.000377655029296875, 0.00234222412109375, 0.004306793212890625, 0.0062713623046875, 0.008235931396484375, 0.01020050048828125, 0.012165069580078125, 0.014129638671875, 0.016094207763671875, 0.01805877685546875, 0.020023345947265625, 0.0219879150390625, 0.023952484130859375, 0.02591705322265625, 0.027881622314453125, 0.02984619140625, 0.031810760498046875, 0.03377532958984375, 0.035739898681640625, 0.0377044677734375, 0.039669036865234375, 0.04163360595703125, 0.043598175048828125, 0.045562744140625, 0.047527313232421875, 0.04949188232421875, 0.051456451416015625, 0.0534210205078125, 0.055385589599609375, 0.05735015869140625, 0.059314727783203125, 0.061279296875]}, "parameters/model.layers.11.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 3.0, 2.0, 0.0, 3.0, 10.0, 9.0, 7.0, 16.0, 21.0, 28.0, 31.0, 46.0, 66.0, 98.0, 152.0, 488.0, 2765.0, 18694.0, 128667.0, 678308.0, 2542344.0, 5400275.0, 5102726.0, 2219763.0, 563269.0, 101619.0, 14713.0, 2199.0, 441.0, 128.0, 92.0, 66.0, 31.0, 28.0, 25.0, 18.0, 12.0, 13.0, 11.0, 4.0, 4.0, 4.0, 3.0, 4.0, 2.0, 0.0, 2.0, 0.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.2490234375, -0.2412261962890625, -0.233428955078125, -0.2256317138671875, -0.21783447265625, -0.2100372314453125, -0.202239990234375, -0.1944427490234375, -0.1866455078125, -0.1788482666015625, -0.171051025390625, -0.1632537841796875, -0.15545654296875, -0.1476593017578125, -0.139862060546875, -0.1320648193359375, -0.124267578125, -0.1164703369140625, -0.108673095703125, -0.1008758544921875, -0.09307861328125, -0.0852813720703125, -0.077484130859375, -0.0696868896484375, -0.0618896484375, -0.0540924072265625, -0.046295166015625, -0.0384979248046875, -0.03070068359375, -0.0229034423828125, -0.015106201171875, -0.0073089599609375, 0.00048828125, 0.0082855224609375, 0.016082763671875, 0.0238800048828125, 0.03167724609375, 0.0394744873046875, 0.047271728515625, 0.0550689697265625, 0.0628662109375, 0.0706634521484375, 0.078460693359375, 0.0862579345703125, 0.09405517578125, 0.1018524169921875, 0.109649658203125, 0.1174468994140625, 0.125244140625, 0.1330413818359375, 0.140838623046875, 0.1486358642578125, 0.15643310546875, 0.1642303466796875, 0.172027587890625, 0.1798248291015625, 0.1876220703125, 0.1954193115234375, 0.203216552734375, 0.2110137939453125, 0.21881103515625, 0.2266082763671875, 0.234405517578125, 0.2422027587890625, 0.25]}, "parameters/model.layers.11.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 0.0, 2.0, 2.0, 4.0, 5.0, 10.0, 10.0, 17.0, 23.0, 46.0, 64.0, 106.0, 187.0, 352.0, 636.0, 1410.0, 4240.0, 17335.0, 99101.0, 654810.0, 4608241.0, 16777216.0, 16777216.0, 16777216.0, 11586437.0, 2034781.0, 284035.0, 44567.0, 8502.0, 2435.0, 984.0, 451.0, 228.0, 140.0, 80.0, 49.0, 36.0, 19.0, 13.0, 9.0, 7.0, 8.0, 1.0, 2.0, 4.0, 2.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.3671875, -0.355987548828125, -0.34478759765625, -0.333587646484375, -0.3223876953125, -0.311187744140625, -0.29998779296875, -0.288787841796875, -0.277587890625, -0.266387939453125, -0.25518798828125, -0.243988037109375, -0.2327880859375, -0.221588134765625, -0.21038818359375, -0.199188232421875, -0.18798828125, -0.176788330078125, -0.16558837890625, -0.154388427734375, -0.1431884765625, -0.131988525390625, -0.12078857421875, -0.109588623046875, -0.098388671875, -0.087188720703125, -0.07598876953125, -0.064788818359375, -0.0535888671875, -0.042388916015625, -0.03118896484375, -0.019989013671875, -0.0087890625, 0.002410888671875, 0.01361083984375, 0.024810791015625, 0.0360107421875, 0.047210693359375, 0.05841064453125, 0.069610595703125, 0.080810546875, 0.092010498046875, 0.10321044921875, 0.114410400390625, 0.1256103515625, 0.136810302734375, 0.14801025390625, 0.159210205078125, 0.17041015625, 0.181610107421875, 0.19281005859375, 0.204010009765625, 0.2152099609375, 0.226409912109375, 0.23760986328125, 0.248809814453125, 0.260009765625, 0.271209716796875, 0.28240966796875, 0.293609619140625, 0.3048095703125, 0.316009521484375, 0.32720947265625, 0.338409423828125, 0.349609375]}, "parameters/model.layers.11.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 4.0, 5.0, 3.0, 4.0, 7.0, 7.0, 14.0, 13.0, 17.0, 31.0, 49.0, 72.0, 119.0, 194.0, 383.0, 796.0, 6245.0, 360382.0, 7299603.0, 16777216.0, 16777216.0, 2168765.0, 58433.0, 1546.0, 569.0, 275.0, 162.0, 96.0, 53.0, 34.0, 20.0, 21.0, 8.0, 8.0, 7.0, 4.0, 5.0, 2.0, 2.0, 5.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.455078125, -0.439788818359375, -0.42449951171875, -0.409210205078125, -0.3939208984375, -0.378631591796875, -0.36334228515625, -0.348052978515625, -0.332763671875, -0.317474365234375, -0.30218505859375, -0.286895751953125, -0.2716064453125, -0.256317138671875, -0.24102783203125, -0.225738525390625, -0.21044921875, -0.195159912109375, -0.17987060546875, -0.164581298828125, -0.1492919921875, -0.134002685546875, -0.11871337890625, -0.103424072265625, -0.088134765625, -0.072845458984375, -0.05755615234375, -0.042266845703125, -0.0269775390625, -0.011688232421875, 0.00360107421875, 0.018890380859375, 0.0341796875, 0.049468994140625, 0.06475830078125, 0.080047607421875, 0.0953369140625, 0.110626220703125, 0.12591552734375, 0.141204833984375, 0.156494140625, 0.171783447265625, 0.18707275390625, 0.202362060546875, 0.2176513671875, 0.232940673828125, 0.24822998046875, 0.263519287109375, 0.27880859375, 0.294097900390625, 0.30938720703125, 0.324676513671875, 0.3399658203125, 0.355255126953125, 0.37054443359375, 0.385833740234375, 0.401123046875, 0.416412353515625, 0.43170166015625, 0.446990966796875, 0.4622802734375, 0.477569580078125, 0.49285888671875, 0.508148193359375, 0.5234375]}, "parameters/model.layers.11.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 4.0, 5.0, 9.0, 12.0, 27.0, 46.0, 73.0, 129.0, 177.0, 289.0, 361.0, 585.0, 496.0, 565.0, 422.0, 306.0, 191.0, 172.0, 68.0, 34.0, 34.0, 11.0, 14.0, 9.0, 4.0, 4.0, 4.0, 2.0, 1.0, 2.0, 1.0, 2.0, 1.0, 2.0, 5.0, 4.0, 0.0, 1.0, 0.0, 2.0, 3.0, 3.0, 2.0, 2.0, 0.0, 0.0, 0.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.1279296875, 0.1426544189453125, 0.157379150390625, 0.1721038818359375, 0.18682861328125, 0.2015533447265625, 0.216278076171875, 0.2310028076171875, 0.2457275390625, 0.2604522705078125, 0.275177001953125, 0.2899017333984375, 0.30462646484375, 0.3193511962890625, 0.334075927734375, 0.3488006591796875, 0.363525390625, 0.3782501220703125, 0.392974853515625, 0.4076995849609375, 0.42242431640625, 0.4371490478515625, 0.451873779296875, 0.4665985107421875, 0.4813232421875, 0.4960479736328125, 0.510772705078125, 0.5254974365234375, 0.54022216796875, 0.5549468994140625, 0.569671630859375, 0.5843963623046875, 0.59912109375, 0.6138458251953125, 0.628570556640625, 0.6432952880859375, 0.65802001953125, 0.6727447509765625, 0.687469482421875, 0.7021942138671875, 0.7169189453125, 0.7316436767578125, 0.746368408203125, 0.7610931396484375, 0.77581787109375, 0.7905426025390625, 0.805267333984375, 0.8199920654296875, 0.834716796875, 0.8494415283203125, 0.864166259765625, 0.8788909912109375, 0.89361572265625, 0.9083404541015625, 0.923065185546875, 0.9377899169921875, 0.9525146484375, 0.9672393798828125, 0.981964111328125, 0.9966888427734375, 1.01141357421875, 1.0261383056640625, 1.040863037109375, 1.0555877685546875, 1.0703125]}, "parameters/model.layers.11.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 6.0, 10.0, 14.0, 7.0, 16.0, 27.0, 26.0, 30.0, 24.0, 26.0, 37.0, 24.0, 38.0, 28.0, 48.0, 42.0, 74.0, 64.0, 105.0, 101.0, 207.0, 161.0, 344.0, 320.0, 583.0, 429.0, 396.0, 428.0, 200.0, 184.0, 53.0, 22.0, 6.0, 3.0, 1.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.095703125, 0.10052490234375, 0.1053466796875, 0.11016845703125, 0.114990234375, 0.11981201171875, 0.1246337890625, 0.12945556640625, 0.13427734375, 0.13909912109375, 0.1439208984375, 0.14874267578125, 0.153564453125, 0.15838623046875, 0.1632080078125, 0.16802978515625, 0.1728515625, 0.17767333984375, 0.1824951171875, 0.18731689453125, 0.192138671875, 0.19696044921875, 0.2017822265625, 0.20660400390625, 0.21142578125, 0.21624755859375, 0.2210693359375, 0.22589111328125, 0.230712890625, 0.23553466796875, 0.2403564453125, 0.24517822265625, 0.25, 0.25482177734375, 0.2596435546875, 0.26446533203125, 0.269287109375, 0.27410888671875, 0.2789306640625, 0.28375244140625, 0.28857421875, 0.29339599609375, 0.2982177734375, 0.30303955078125, 0.307861328125, 0.31268310546875, 0.3175048828125, 0.32232666015625, 0.3271484375, 0.33197021484375, 0.3367919921875, 0.34161376953125, 0.346435546875, 0.35125732421875, 0.3560791015625, 0.36090087890625, 0.36572265625, 0.37054443359375, 0.3753662109375, 0.38018798828125, 0.385009765625, 0.38983154296875, 0.3946533203125, 0.39947509765625, 0.404296875]}, "parameters/model.layers.12.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 4.0, 0.0, 6.0, 3.0, 8.0, 7.0, 24.0, 30.0, 65.0, 114.0, 271.0, 495.0, 931.0, 2265.0, 4797.0, 11277.0, 27393.0, 60136.0, 134678.0, 265501.0, 563920.0, 1019418.0, 1741362.0, 2497587.0, 3012398.0, 2756614.0, 2038011.0, 1237878.0, 718864.0, 361711.0, 176592.0, 79929.0, 37369.0, 15420.0, 6655.0, 2821.0, 1395.0, 650.0, 321.0, 136.0, 55.0, 36.0, 14.0, 17.0, 11.0, 3.0, 10.0, 5.0, 1.0, 0.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.2333984375, -0.2266693115234375, -0.219940185546875, -0.2132110595703125, -0.20648193359375, -0.1997528076171875, -0.193023681640625, -0.1862945556640625, -0.1795654296875, -0.1728363037109375, -0.166107177734375, -0.1593780517578125, -0.15264892578125, -0.1459197998046875, -0.139190673828125, -0.1324615478515625, -0.125732421875, -0.1190032958984375, -0.112274169921875, -0.1055450439453125, -0.09881591796875, -0.0920867919921875, -0.085357666015625, -0.0786285400390625, -0.0718994140625, -0.0651702880859375, -0.058441162109375, -0.0517120361328125, -0.04498291015625, -0.0382537841796875, -0.031524658203125, -0.0247955322265625, -0.01806640625, -0.0113372802734375, -0.004608154296875, 0.0021209716796875, 0.00885009765625, 0.0155792236328125, 0.022308349609375, 0.0290374755859375, 0.0357666015625, 0.0424957275390625, 0.049224853515625, 0.0559539794921875, 0.06268310546875, 0.0694122314453125, 0.076141357421875, 0.0828704833984375, 0.089599609375, 0.0963287353515625, 0.103057861328125, 0.1097869873046875, 0.11651611328125, 0.1232452392578125, 0.129974365234375, 0.1367034912109375, 0.1434326171875, 0.1501617431640625, 0.156890869140625, 0.1636199951171875, 0.17034912109375, 0.1770782470703125, 0.183807373046875, 0.1905364990234375, 0.197265625]}, "parameters/model.layers.12.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 3.0, 0.0, 2.0, 8.0, 10.0, 14.0, 15.0, 29.0, 62.0, 127.0, 202.0, 371.0, 725.0, 1310.0, 2545.0, 4361.0, 7925.0, 14418.0, 23087.0, 39877.0, 67667.0, 98780.0, 145928.0, 224093.0, 294495.0, 384903.0, 457210.0, 492347.0, 473857.0, 411837.0, 326074.0, 247938.0, 166494.0, 118153.0, 74607.0, 47957.0, 29618.0, 15678.0, 9587.0, 5396.0, 2986.0, 1657.0, 969.0, 455.0, 214.0, 121.0, 77.0, 27.0, 23.0, 23.0, 16.0, 12.0, 3.0, 4.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.205078125, -0.1981658935546875, -0.191253662109375, -0.1843414306640625, -0.17742919921875, -0.1705169677734375, -0.163604736328125, -0.1566925048828125, -0.1497802734375, -0.1428680419921875, -0.135955810546875, -0.1290435791015625, -0.12213134765625, -0.1152191162109375, -0.108306884765625, -0.1013946533203125, -0.094482421875, -0.0875701904296875, -0.080657958984375, -0.0737457275390625, -0.06683349609375, -0.0599212646484375, -0.053009033203125, -0.0460968017578125, -0.0391845703125, -0.0322723388671875, -0.025360107421875, -0.0184478759765625, -0.01153564453125, -0.0046234130859375, 0.002288818359375, 0.0092010498046875, 0.01611328125, 0.0230255126953125, 0.029937744140625, 0.0368499755859375, 0.04376220703125, 0.0506744384765625, 0.057586669921875, 0.0644989013671875, 0.0714111328125, 0.0783233642578125, 0.085235595703125, 0.0921478271484375, 0.09906005859375, 0.1059722900390625, 0.112884521484375, 0.1197967529296875, 0.126708984375, 0.1336212158203125, 0.140533447265625, 0.1474456787109375, 0.15435791015625, 0.1612701416015625, 0.168182373046875, 0.1750946044921875, 0.1820068359375, 0.1889190673828125, 0.195831298828125, 0.2027435302734375, 0.20965576171875, 0.2165679931640625, 0.223480224609375, 0.2303924560546875, 0.2373046875]}, "parameters/model.layers.12.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 2.0, 5.0, 7.0, 19.0, 37.0, 87.0, 197.0, 353.0, 710.0, 1189.0, 2487.0, 4613.0, 8422.0, 15114.0, 28134.0, 44282.0, 71143.0, 109577.0, 168327.0, 224996.0, 298326.0, 363923.0, 425146.0, 455068.0, 440743.0, 393345.0, 333297.0, 257240.0, 191212.0, 133601.0, 91387.0, 54356.0, 33274.0, 19668.0, 11111.0, 6115.0, 3255.0, 1751.0, 928.0, 397.0, 256.0, 97.0, 69.0, 20.0, 6.0, 6.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.06005859375, -0.0577239990234375, -0.055389404296875, -0.0530548095703125, -0.05072021484375, -0.0483856201171875, -0.046051025390625, -0.0437164306640625, -0.0413818359375, -0.0390472412109375, -0.036712646484375, -0.0343780517578125, -0.03204345703125, -0.0297088623046875, -0.027374267578125, -0.0250396728515625, -0.022705078125, -0.0203704833984375, -0.018035888671875, -0.0157012939453125, -0.01336669921875, -0.0110321044921875, -0.008697509765625, -0.0063629150390625, -0.0040283203125, -0.0016937255859375, 0.000640869140625, 0.0029754638671875, 0.00531005859375, 0.0076446533203125, 0.009979248046875, 0.0123138427734375, 0.0146484375, 0.0169830322265625, 0.019317626953125, 0.0216522216796875, 0.02398681640625, 0.0263214111328125, 0.028656005859375, 0.0309906005859375, 0.0333251953125, 0.0356597900390625, 0.037994384765625, 0.0403289794921875, 0.04266357421875, 0.0449981689453125, 0.047332763671875, 0.0496673583984375, 0.052001953125, 0.0543365478515625, 0.056671142578125, 0.0590057373046875, 0.06134033203125, 0.0636749267578125, 0.066009521484375, 0.0683441162109375, 0.0706787109375, 0.0730133056640625, 0.075347900390625, 0.0776824951171875, 0.08001708984375, 0.0823516845703125, 0.084686279296875, 0.0870208740234375, 0.08935546875]}, "parameters/model.layers.12.self_attn.o_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 3.0, 0.0, 0.0, 2.0, 1.0, 2.0, 3.0, 2.0, 5.0, 2.0, 4.0, 4.0, 4.0, 7.0, 11.0, 12.0, 19.0, 62.0, 114.0, 241.0, 927.0, 13630.0, 233277.0, 2042525.0, 6721619.0, 6044131.0, 1554428.0, 156200.0, 8816.0, 740.0, 214.0, 89.0, 43.0, 24.0, 13.0, 7.0, 5.0, 2.0, 3.0, 2.0, 3.0, 1.0, 0.0, 0.0, 3.0, 3.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.314453125, -0.303192138671875, -0.29193115234375, -0.280670166015625, -0.2694091796875, -0.258148193359375, -0.24688720703125, -0.235626220703125, -0.224365234375, -0.213104248046875, -0.20184326171875, -0.190582275390625, -0.1793212890625, -0.168060302734375, -0.15679931640625, -0.145538330078125, -0.13427734375, -0.123016357421875, -0.11175537109375, -0.100494384765625, -0.0892333984375, -0.077972412109375, -0.06671142578125, -0.055450439453125, -0.044189453125, -0.032928466796875, -0.02166748046875, -0.010406494140625, 0.0008544921875, 0.012115478515625, 0.02337646484375, 0.034637451171875, 0.0458984375, 0.057159423828125, 0.06842041015625, 0.079681396484375, 0.0909423828125, 0.102203369140625, 0.11346435546875, 0.124725341796875, 0.135986328125, 0.147247314453125, 0.15850830078125, 0.169769287109375, 0.1810302734375, 0.192291259765625, 0.20355224609375, 0.214813232421875, 0.22607421875, 0.237335205078125, 0.24859619140625, 0.259857177734375, 0.2711181640625, 0.282379150390625, 0.29364013671875, 0.304901123046875, 0.316162109375, 0.327423095703125, 0.33868408203125, 0.349945068359375, 0.3612060546875, 0.372467041015625, 0.38372802734375, 0.394989013671875, 0.40625]}, "parameters/model.layers.12.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 2.0, 1.0, 0.0, 1.0, 1.0, 3.0, 1.0, 6.0, 5.0, 13.0, 11.0, 9.0, 25.0, 18.0, 22.0, 56.0, 66.0, 104.0, 135.0, 229.0, 359.0, 676.0, 1396.0, 4134.0, 18228.0, 115767.0, 949850.0, 7250608.0, 16777216.0, 16777216.0, 16777216.0, 5026188.0, 605504.0, 77341.0, 12676.0, 3248.0, 1222.0, 602.0, 324.0, 182.0, 138.0, 82.0, 58.0, 47.0, 37.0, 27.0, 14.0, 14.0, 8.0, 5.0, 6.0, 2.0, 2.0, 3.0, 0.0, 2.0, 1.0, 1.0, 1.0, 1.0, 2.0], "bins": [-0.396484375, -0.384246826171875, -0.37200927734375, -0.359771728515625, -0.3475341796875, -0.335296630859375, -0.32305908203125, -0.310821533203125, -0.298583984375, -0.286346435546875, -0.27410888671875, -0.261871337890625, -0.2496337890625, -0.237396240234375, -0.22515869140625, -0.212921142578125, -0.20068359375, -0.188446044921875, -0.17620849609375, -0.163970947265625, -0.1517333984375, -0.139495849609375, -0.12725830078125, -0.115020751953125, -0.102783203125, -0.090545654296875, -0.07830810546875, -0.066070556640625, -0.0538330078125, -0.041595458984375, -0.02935791015625, -0.017120361328125, -0.0048828125, 0.007354736328125, 0.01959228515625, 0.031829833984375, 0.0440673828125, 0.056304931640625, 0.06854248046875, 0.080780029296875, 0.093017578125, 0.105255126953125, 0.11749267578125, 0.129730224609375, 0.1419677734375, 0.154205322265625, 0.16644287109375, 0.178680419921875, 0.19091796875, 0.203155517578125, 0.21539306640625, 0.227630615234375, 0.2398681640625, 0.252105712890625, 0.26434326171875, 0.276580810546875, 0.288818359375, 0.301055908203125, 0.31329345703125, 0.325531005859375, 0.3377685546875, 0.350006103515625, 0.36224365234375, 0.374481201171875, 0.38671875]}, "parameters/model.layers.12.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 5.0, 3.0, 4.0, 7.0, 10.0, 20.0, 36.0, 52.0, 116.0, 244.0, 693.0, 3351.0, 773830.0, 16777216.0, 16777216.0, 1214318.0, 4900.0, 693.0, 248.0, 122.0, 65.0, 32.0, 11.0, 16.0, 5.0, 6.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.8359375, -0.812744140625, -0.78955078125, -0.766357421875, -0.7431640625, -0.719970703125, -0.69677734375, -0.673583984375, -0.650390625, -0.627197265625, -0.60400390625, -0.580810546875, -0.5576171875, -0.534423828125, -0.51123046875, -0.488037109375, -0.46484375, -0.441650390625, -0.41845703125, -0.395263671875, -0.3720703125, -0.348876953125, -0.32568359375, -0.302490234375, -0.279296875, -0.256103515625, -0.23291015625, -0.209716796875, -0.1865234375, -0.163330078125, -0.14013671875, -0.116943359375, -0.09375, -0.070556640625, -0.04736328125, -0.024169921875, -0.0009765625, 0.022216796875, 0.04541015625, 0.068603515625, 0.091796875, 0.114990234375, 0.13818359375, 0.161376953125, 0.1845703125, 0.207763671875, 0.23095703125, 0.254150390625, 0.27734375, 0.300537109375, 0.32373046875, 0.346923828125, 0.3701171875, 0.393310546875, 0.41650390625, 0.439697265625, 0.462890625, 0.486083984375, 0.50927734375, 0.532470703125, 0.5556640625, 0.578857421875, 0.60205078125, 0.625244140625, 0.6484375]}, "parameters/model.layers.12.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 3.0, 4.0, 4.0, 11.0, 20.0, 21.0, 41.0, 42.0, 73.0, 85.0, 127.0, 156.0, 265.0, 306.0, 439.0, 436.0, 598.0, 416.0, 401.0, 233.0, 175.0, 76.0, 58.0, 34.0, 10.0, 9.0, 6.0, 5.0, 4.0, 1.0, 1.0, 1.0, 2.0, 4.0, 2.0, 3.0, 2.0, 3.0, 0.0, 2.0, 4.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.11474609375, 0.12747955322265625, 0.1402130126953125, 0.15294647216796875, 0.165679931640625, 0.17841339111328125, 0.1911468505859375, 0.20388031005859375, 0.21661376953125, 0.22934722900390625, 0.2420806884765625, 0.25481414794921875, 0.267547607421875, 0.28028106689453125, 0.2930145263671875, 0.30574798583984375, 0.3184814453125, 0.33121490478515625, 0.3439483642578125, 0.35668182373046875, 0.369415283203125, 0.38214874267578125, 0.3948822021484375, 0.40761566162109375, 0.42034912109375, 0.43308258056640625, 0.4458160400390625, 0.45854949951171875, 0.471282958984375, 0.48401641845703125, 0.4967498779296875, 0.5094833374023438, 0.522216796875, 0.5349502563476562, 0.5476837158203125, 0.5604171752929688, 0.573150634765625, 0.5858840942382812, 0.5986175537109375, 0.6113510131835938, 0.62408447265625, 0.6368179321289062, 0.6495513916015625, 0.6622848510742188, 0.675018310546875, 0.6877517700195312, 0.7004852294921875, 0.7132186889648438, 0.7259521484375, 0.7386856079101562, 0.7514190673828125, 0.7641525268554688, 0.776885986328125, 0.7896194458007812, 0.8023529052734375, 0.8150863647460938, 0.82781982421875, 0.8405532836914062, 0.8532867431640625, 0.8660202026367188, 0.878753662109375, 0.8914871215820312, 0.9042205810546875, 0.9169540405273438, 0.9296875]}, "parameters/model.layers.12.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 2.0, 6.0, 2.0, 7.0, 7.0, 16.0, 34.0, 36.0, 61.0, 73.0, 85.0, 61.0, 107.0, 131.0, 219.0, 208.0, 390.0, 600.0, 687.0, 629.0, 298.0, 258.0, 106.0, 39.0, 13.0, 6.0, 5.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.0751953125, 0.0806121826171875, 0.086029052734375, 0.0914459228515625, 0.09686279296875, 0.1022796630859375, 0.107696533203125, 0.1131134033203125, 0.1185302734375, 0.1239471435546875, 0.129364013671875, 0.1347808837890625, 0.14019775390625, 0.1456146240234375, 0.151031494140625, 0.1564483642578125, 0.161865234375, 0.1672821044921875, 0.172698974609375, 0.1781158447265625, 0.18353271484375, 0.1889495849609375, 0.194366455078125, 0.1997833251953125, 0.2052001953125, 0.2106170654296875, 0.216033935546875, 0.2214508056640625, 0.22686767578125, 0.2322845458984375, 0.237701416015625, 0.2431182861328125, 0.24853515625, 0.2539520263671875, 0.259368896484375, 0.2647857666015625, 0.27020263671875, 0.2756195068359375, 0.281036376953125, 0.2864532470703125, 0.2918701171875, 0.2972869873046875, 0.302703857421875, 0.3081207275390625, 0.31353759765625, 0.3189544677734375, 0.324371337890625, 0.3297882080078125, 0.335205078125, 0.3406219482421875, 0.346038818359375, 0.3514556884765625, 0.35687255859375, 0.3622894287109375, 0.367706298828125, 0.3731231689453125, 0.3785400390625, 0.3839569091796875, 0.389373779296875, 0.3947906494140625, 0.40020751953125, 0.4056243896484375, 0.411041259765625, 0.4164581298828125, 0.421875]}, "parameters/model.layers.13.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 4.0, 0.0, 7.0, 11.0, 18.0, 38.0, 77.0, 138.0, 257.0, 483.0, 829.0, 1568.0, 2917.0, 5077.0, 8748.0, 15958.0, 30778.0, 52253.0, 88152.0, 146576.0, 244013.0, 413392.0, 664617.0, 1049752.0, 1620694.0, 2209266.0, 2587402.0, 2420701.0, 1868408.0, 1275882.0, 807022.0, 511747.0, 312232.0, 179953.0, 107424.0, 65165.0, 38078.0, 21111.0, 11583.0, 6622.0, 3936.0, 1996.0, 1094.0, 574.0, 288.0, 173.0, 80.0, 50.0, 27.0, 21.0, 9.0, 4.0, 2.0, 0.0, 1.0, 1.0], "bins": [-0.1875, -0.182098388671875, -0.17669677734375, -0.171295166015625, -0.1658935546875, -0.160491943359375, -0.15509033203125, -0.149688720703125, -0.144287109375, -0.138885498046875, -0.13348388671875, -0.128082275390625, -0.1226806640625, -0.117279052734375, -0.11187744140625, -0.106475830078125, -0.10107421875, -0.095672607421875, -0.09027099609375, -0.084869384765625, -0.0794677734375, -0.074066162109375, -0.06866455078125, -0.063262939453125, -0.057861328125, -0.052459716796875, -0.04705810546875, -0.041656494140625, -0.0362548828125, -0.030853271484375, -0.02545166015625, -0.020050048828125, -0.0146484375, -0.009246826171875, -0.00384521484375, 0.001556396484375, 0.0069580078125, 0.012359619140625, 0.01776123046875, 0.023162841796875, 0.028564453125, 0.033966064453125, 0.03936767578125, 0.044769287109375, 0.0501708984375, 0.055572509765625, 0.06097412109375, 0.066375732421875, 0.07177734375, 0.077178955078125, 0.08258056640625, 0.087982177734375, 0.0933837890625, 0.098785400390625, 0.10418701171875, 0.109588623046875, 0.114990234375, 0.120391845703125, 0.12579345703125, 0.131195068359375, 0.1365966796875, 0.141998291015625, 0.14739990234375, 0.152801513671875, 0.158203125]}, "parameters/model.layers.13.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 3.0, 2.0, 1.0, 2.0, 4.0, 9.0, 8.0, 12.0, 16.0, 21.0, 32.0, 48.0, 75.0, 112.0, 194.0, 320.0, 762.0, 1703.0, 3512.0, 7646.0, 15177.0, 31135.0, 61694.0, 112235.0, 198834.0, 350808.0, 551879.0, 751310.0, 752363.0, 558405.0, 355979.0, 197711.0, 117016.0, 64005.0, 30306.0, 16360.0, 7645.0, 3692.0, 1602.0, 759.0, 356.0, 202.0, 112.0, 68.0, 43.0, 32.0, 26.0, 22.0, 9.0, 8.0, 8.0, 10.0, 2.0, 2.0, 0.0, 5.0], "bins": [-0.376953125, -0.366485595703125, -0.35601806640625, -0.345550537109375, -0.3350830078125, -0.324615478515625, -0.31414794921875, -0.303680419921875, -0.293212890625, -0.282745361328125, -0.27227783203125, -0.261810302734375, -0.2513427734375, -0.240875244140625, -0.23040771484375, -0.219940185546875, -0.20947265625, -0.199005126953125, -0.18853759765625, -0.178070068359375, -0.1676025390625, -0.157135009765625, -0.14666748046875, -0.136199951171875, -0.125732421875, -0.115264892578125, -0.10479736328125, -0.094329833984375, -0.0838623046875, -0.073394775390625, -0.06292724609375, -0.052459716796875, -0.0419921875, -0.031524658203125, -0.02105712890625, -0.010589599609375, -0.0001220703125, 0.010345458984375, 0.02081298828125, 0.031280517578125, 0.041748046875, 0.052215576171875, 0.06268310546875, 0.073150634765625, 0.0836181640625, 0.094085693359375, 0.10455322265625, 0.115020751953125, 0.12548828125, 0.135955810546875, 0.14642333984375, 0.156890869140625, 0.1673583984375, 0.177825927734375, 0.18829345703125, 0.198760986328125, 0.209228515625, 0.219696044921875, 0.23016357421875, 0.240631103515625, 0.2510986328125, 0.261566162109375, 0.27203369140625, 0.282501220703125, 0.29296875]}, "parameters/model.layers.13.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 2.0, 3.0, 0.0, 1.0, 7.0, 5.0, 12.0, 18.0, 63.0, 170.0, 395.0, 883.0, 2242.0, 5140.0, 11185.0, 21756.0, 42057.0, 76936.0, 129516.0, 203021.0, 295156.0, 393261.0, 481088.0, 537075.0, 518647.0, 449647.0, 355388.0, 258626.0, 175017.0, 107379.0, 62358.0, 34296.0, 17661.0, 8506.0, 3894.0, 1712.0, 704.0, 248.0, 123.0, 42.0, 25.0, 15.0, 6.0, 5.0, 3.0, 1.0, 2.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0712890625, -0.0687255859375, -0.066162109375, -0.0635986328125, -0.06103515625, -0.0584716796875, -0.055908203125, -0.0533447265625, -0.05078125, -0.0482177734375, -0.045654296875, -0.0430908203125, -0.04052734375, -0.0379638671875, -0.035400390625, -0.0328369140625, -0.0302734375, -0.0277099609375, -0.025146484375, -0.0225830078125, -0.02001953125, -0.0174560546875, -0.014892578125, -0.0123291015625, -0.009765625, -0.0072021484375, -0.004638671875, -0.0020751953125, 0.00048828125, 0.0030517578125, 0.005615234375, 0.0081787109375, 0.0107421875, 0.0133056640625, 0.015869140625, 0.0184326171875, 0.02099609375, 0.0235595703125, 0.026123046875, 0.0286865234375, 0.03125, 0.0338134765625, 0.036376953125, 0.0389404296875, 0.04150390625, 0.0440673828125, 0.046630859375, 0.0491943359375, 0.0517578125, 0.0543212890625, 0.056884765625, 0.0594482421875, 0.06201171875, 0.0645751953125, 0.067138671875, 0.0697021484375, 0.072265625, 0.0748291015625, 0.077392578125, 0.0799560546875, 0.08251953125, 0.0850830078125, 0.087646484375, 0.0902099609375, 0.0927734375]}, "parameters/model.layers.13.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 5.0, 2.0, 2.0, 0.0, 4.0, 3.0, 5.0, 4.0, 2.0, 3.0, 7.0, 15.0, 22.0, 337.0, 27458.0, 1491681.0, 10490438.0, 4591087.0, 174179.0, 1839.0, 56.0, 16.0, 6.0, 4.0, 2.0, 2.0, 4.0, 4.0, 2.0, 3.0, 1.0, 4.0, 1.0, 2.0, 1.0, 1.0, 2.0, 2.0, 1.0, 0.0, 3.0], "bins": [-0.69921875, -0.681610107421875, -0.66400146484375, -0.646392822265625, -0.6287841796875, -0.611175537109375, -0.59356689453125, -0.575958251953125, -0.558349609375, -0.540740966796875, -0.52313232421875, -0.505523681640625, -0.4879150390625, -0.470306396484375, -0.45269775390625, -0.435089111328125, -0.41748046875, -0.399871826171875, -0.38226318359375, -0.364654541015625, -0.3470458984375, -0.329437255859375, -0.31182861328125, -0.294219970703125, -0.276611328125, -0.259002685546875, -0.24139404296875, -0.223785400390625, -0.2061767578125, -0.188568115234375, -0.17095947265625, -0.153350830078125, -0.1357421875, -0.118133544921875, -0.10052490234375, -0.082916259765625, -0.0653076171875, -0.047698974609375, -0.03009033203125, -0.012481689453125, 0.005126953125, 0.022735595703125, 0.04034423828125, 0.057952880859375, 0.0755615234375, 0.093170166015625, 0.11077880859375, 0.128387451171875, 0.14599609375, 0.163604736328125, 0.18121337890625, 0.198822021484375, 0.2164306640625, 0.234039306640625, 0.25164794921875, 0.269256591796875, 0.286865234375, 0.304473876953125, 0.32208251953125, 0.339691162109375, 0.3572998046875, 0.374908447265625, 0.39251708984375, 0.410125732421875, 0.427734375]}, "parameters/model.layers.13.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 3.0, 2.0, 3.0, 6.0, 4.0, 9.0, 8.0, 9.0, 18.0, 14.0, 20.0, 29.0, 45.0, 63.0, 62.0, 115.0, 146.0, 260.0, 381.0, 717.0, 1545.0, 3928.0, 15497.0, 104302.0, 958739.0, 7967563.0, 16777216.0, 16777216.0, 16777216.0, 3810111.0, 407778.0, 48557.0, 8581.0, 2704.0, 1091.0, 606.0, 325.0, 181.0, 140.0, 115.0, 70.0, 49.0, 27.0, 30.0, 28.0, 17.0, 15.0, 8.0, 10.0, 12.0, 6.0, 6.0, 0.0, 2.0, 2.0, 1.0, 2.0, 2.0, 0.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.359375, -0.3466796875, -0.333984375, -0.3212890625, -0.30859375, -0.2958984375, -0.283203125, -0.2705078125, -0.2578125, -0.2451171875, -0.232421875, -0.2197265625, -0.20703125, -0.1943359375, -0.181640625, -0.1689453125, -0.15625, -0.1435546875, -0.130859375, -0.1181640625, -0.10546875, -0.0927734375, -0.080078125, -0.0673828125, -0.0546875, -0.0419921875, -0.029296875, -0.0166015625, -0.00390625, 0.0087890625, 0.021484375, 0.0341796875, 0.046875, 0.0595703125, 0.072265625, 0.0849609375, 0.09765625, 0.1103515625, 0.123046875, 0.1357421875, 0.1484375, 0.1611328125, 0.173828125, 0.1865234375, 0.19921875, 0.2119140625, 0.224609375, 0.2373046875, 0.25, 0.2626953125, 0.275390625, 0.2880859375, 0.30078125, 0.3134765625, 0.326171875, 0.3388671875, 0.3515625, 0.3642578125, 0.376953125, 0.3896484375, 0.40234375, 0.4150390625, 0.427734375, 0.4404296875, 0.453125]}, "parameters/model.layers.13.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 4.0, 0.0, 3.0, 4.0, 2.0, 10.0, 16.0, 50.0, 75.0, 176.0, 446.0, 6827.0, 1944925.0, 16777216.0, 16777216.0, 493805.0, 1756.0, 347.0, 141.0, 58.0, 28.0, 14.0, 8.0, 7.0, 2.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.6640625, -0.64105224609375, -0.6180419921875, -0.59503173828125, -0.572021484375, -0.54901123046875, -0.5260009765625, -0.50299072265625, -0.47998046875, -0.45697021484375, -0.4339599609375, -0.41094970703125, -0.387939453125, -0.36492919921875, -0.3419189453125, -0.31890869140625, -0.2958984375, -0.27288818359375, -0.2498779296875, -0.22686767578125, -0.203857421875, -0.18084716796875, -0.1578369140625, -0.13482666015625, -0.11181640625, -0.08880615234375, -0.0657958984375, -0.04278564453125, -0.019775390625, 0.00323486328125, 0.0262451171875, 0.04925537109375, 0.072265625, 0.09527587890625, 0.1182861328125, 0.14129638671875, 0.164306640625, 0.18731689453125, 0.2103271484375, 0.23333740234375, 0.25634765625, 0.27935791015625, 0.3023681640625, 0.32537841796875, 0.348388671875, 0.37139892578125, 0.3944091796875, 0.41741943359375, 0.4404296875, 0.46343994140625, 0.4864501953125, 0.50946044921875, 0.532470703125, 0.55548095703125, 0.5784912109375, 0.60150146484375, 0.62451171875, 0.64752197265625, 0.6705322265625, 0.69354248046875, 0.716552734375, 0.73956298828125, 0.7625732421875, 0.78558349609375, 0.80859375]}, "parameters/model.layers.13.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 2.0, 5.0, 5.0, 7.0, 12.0, 12.0, 20.0, 30.0, 41.0, 50.0, 87.0, 107.0, 114.0, 168.0, 252.0, 317.0, 444.0, 433.0, 490.0, 412.0, 375.0, 254.0, 115.0, 126.0, 53.0, 43.0, 31.0, 15.0, 13.0, 4.0, 6.0, 1.0, 2.0, 3.0, 0.0, 3.0, 1.0, 4.0, 1.0, 1.0, 1.0, 1.0, 2.0, 3.0, 2.0, 1.0, 0.0, 0.0, 3.0, 6.0, 2.0, 2.0, 2.0, 1.0, 1.0, 0.0, 3.0], "bins": [0.11669921875, 0.13062286376953125, 0.1445465087890625, 0.15847015380859375, 0.172393798828125, 0.18631744384765625, 0.2002410888671875, 0.21416473388671875, 0.22808837890625, 0.24201202392578125, 0.2559356689453125, 0.26985931396484375, 0.283782958984375, 0.29770660400390625, 0.3116302490234375, 0.32555389404296875, 0.3394775390625, 0.35340118408203125, 0.3673248291015625, 0.38124847412109375, 0.395172119140625, 0.40909576416015625, 0.4230194091796875, 0.43694305419921875, 0.45086669921875, 0.46479034423828125, 0.4787139892578125, 0.49263763427734375, 0.506561279296875, 0.5204849243164062, 0.5344085693359375, 0.5483322143554688, 0.562255859375, 0.5761795043945312, 0.5901031494140625, 0.6040267944335938, 0.617950439453125, 0.6318740844726562, 0.6457977294921875, 0.6597213745117188, 0.67364501953125, 0.6875686645507812, 0.7014923095703125, 0.7154159545898438, 0.729339599609375, 0.7432632446289062, 0.7571868896484375, 0.7711105346679688, 0.7850341796875, 0.7989578247070312, 0.8128814697265625, 0.8268051147460938, 0.840728759765625, 0.8546524047851562, 0.8685760498046875, 0.8824996948242188, 0.89642333984375, 0.9103469848632812, 0.9242706298828125, 0.9381942749023438, 0.952117919921875, 0.9660415649414062, 0.9799652099609375, 0.9938888549804688, 1.0078125]}, "parameters/model.layers.13.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 0.0, 3.0, 2.0, 7.0, 2.0, 2.0, 4.0, 2.0, 1.0, 3.0, 1.0, 8.0, 8.0, 6.0, 25.0, 38.0, 68.0, 99.0, 99.0, 151.0, 199.0, 400.0, 538.0, 805.0, 799.0, 482.0, 222.0, 70.0, 28.0, 5.0, 2.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.0927734375, 0.0986175537109375, 0.104461669921875, 0.1103057861328125, 0.11614990234375, 0.1219940185546875, 0.127838134765625, 0.1336822509765625, 0.1395263671875, 0.1453704833984375, 0.151214599609375, 0.1570587158203125, 0.16290283203125, 0.1687469482421875, 0.174591064453125, 0.1804351806640625, 0.186279296875, 0.1921234130859375, 0.197967529296875, 0.2038116455078125, 0.20965576171875, 0.2154998779296875, 0.221343994140625, 0.2271881103515625, 0.2330322265625, 0.2388763427734375, 0.244720458984375, 0.2505645751953125, 0.25640869140625, 0.2622528076171875, 0.268096923828125, 0.2739410400390625, 0.27978515625, 0.2856292724609375, 0.291473388671875, 0.2973175048828125, 0.30316162109375, 0.3090057373046875, 0.314849853515625, 0.3206939697265625, 0.3265380859375, 0.3323822021484375, 0.338226318359375, 0.3440704345703125, 0.34991455078125, 0.3557586669921875, 0.361602783203125, 0.3674468994140625, 0.373291015625, 0.3791351318359375, 0.384979248046875, 0.3908233642578125, 0.39666748046875, 0.4025115966796875, 0.408355712890625, 0.4141998291015625, 0.4200439453125, 0.4258880615234375, 0.431732177734375, 0.4375762939453125, 0.44342041015625, 0.4492645263671875, 0.455108642578125, 0.4609527587890625, 0.466796875]}, "parameters/model.layers.14.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 2.0, 0.0, 2.0, 5.0, 2.0, 4.0, 7.0, 11.0, 11.0, 20.0, 28.0, 34.0, 48.0, 56.0, 87.0, 83.0, 139.0, 169.0, 274.0, 442.0, 916.0, 2610.0, 8242.0, 25117.0, 77784.0, 217539.0, 613931.0, 1490950.0, 3180543.0, 4645130.0, 3560820.0, 1784418.0, 749554.0, 271117.0, 99225.0, 31831.0, 9912.0, 3461.0, 1200.0, 498.0, 240.0, 172.0, 132.0, 86.0, 95.0, 49.0, 64.0, 34.0, 23.0, 22.0, 20.0, 16.0, 12.0, 9.0, 12.0, 0.0, 0.0, 2.0, 1.0, 1.0, 1.0, 1.0], "bins": [-0.310546875, -0.30072021484375, -0.2908935546875, -0.28106689453125, -0.271240234375, -0.26141357421875, -0.2515869140625, -0.24176025390625, -0.23193359375, -0.22210693359375, -0.2122802734375, -0.20245361328125, -0.192626953125, -0.18280029296875, -0.1729736328125, -0.16314697265625, -0.1533203125, -0.14349365234375, -0.1336669921875, -0.12384033203125, -0.114013671875, -0.10418701171875, -0.0943603515625, -0.08453369140625, -0.07470703125, -0.06488037109375, -0.0550537109375, -0.04522705078125, -0.035400390625, -0.02557373046875, -0.0157470703125, -0.00592041015625, 0.00390625, 0.01373291015625, 0.0235595703125, 0.03338623046875, 0.043212890625, 0.05303955078125, 0.0628662109375, 0.07269287109375, 0.08251953125, 0.09234619140625, 0.1021728515625, 0.11199951171875, 0.121826171875, 0.13165283203125, 0.1414794921875, 0.15130615234375, 0.1611328125, 0.17095947265625, 0.1807861328125, 0.19061279296875, 0.200439453125, 0.21026611328125, 0.2200927734375, 0.22991943359375, 0.23974609375, 0.24957275390625, 0.2593994140625, 0.26922607421875, 0.279052734375, 0.28887939453125, 0.2987060546875, 0.30853271484375, 0.318359375]}, "parameters/model.layers.14.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 3.0, 5.0, 6.0, 12.0, 6.0, 12.0, 24.0, 22.0, 33.0, 39.0, 50.0, 72.0, 77.0, 120.0, 194.0, 255.0, 449.0, 784.0, 1495.0, 3256.0, 6787.0, 13063.0, 26604.0, 55908.0, 110066.0, 208873.0, 391218.0, 629056.0, 824521.0, 764522.0, 527416.0, 301973.0, 160474.0, 85774.0, 40262.0, 21019.0, 9941.0, 4570.0, 2449.0, 1178.0, 564.0, 334.0, 225.0, 145.0, 103.0, 78.0, 59.0, 48.0, 37.0, 28.0, 23.0, 20.0, 11.0, 7.0, 11.0, 9.0, 4.0, 3.0, 0.0, 4.0], "bins": [-0.36328125, -0.352203369140625, -0.34112548828125, -0.330047607421875, -0.3189697265625, -0.307891845703125, -0.29681396484375, -0.285736083984375, -0.274658203125, -0.263580322265625, -0.25250244140625, -0.241424560546875, -0.2303466796875, -0.219268798828125, -0.20819091796875, -0.197113037109375, -0.18603515625, -0.174957275390625, -0.16387939453125, -0.152801513671875, -0.1417236328125, -0.130645751953125, -0.11956787109375, -0.108489990234375, -0.097412109375, -0.086334228515625, -0.07525634765625, -0.064178466796875, -0.0531005859375, -0.042022705078125, -0.03094482421875, -0.019866943359375, -0.0087890625, 0.002288818359375, 0.01336669921875, 0.024444580078125, 0.0355224609375, 0.046600341796875, 0.05767822265625, 0.068756103515625, 0.079833984375, 0.090911865234375, 0.10198974609375, 0.113067626953125, 0.1241455078125, 0.135223388671875, 0.14630126953125, 0.157379150390625, 0.16845703125, 0.179534912109375, 0.19061279296875, 0.201690673828125, 0.2127685546875, 0.223846435546875, 0.23492431640625, 0.246002197265625, 0.257080078125, 0.268157958984375, 0.27923583984375, 0.290313720703125, 0.3013916015625, 0.312469482421875, 0.32354736328125, 0.334625244140625, 0.345703125]}, "parameters/model.layers.14.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 3.0, 4.0, 4.0, 3.0, 9.0, 19.0, 23.0, 38.0, 53.0, 119.0, 183.0, 323.0, 635.0, 1086.0, 2282.0, 3914.0, 7178.0, 13800.0, 22175.0, 37692.0, 64416.0, 98058.0, 153592.0, 219614.0, 296498.0, 379656.0, 452085.0, 488800.0, 472296.0, 407924.0, 335997.0, 249773.0, 174729.0, 121540.0, 76033.0, 46807.0, 29646.0, 16362.0, 9362.0, 5513.0, 2786.0, 1462.0, 863.0, 386.0, 241.0, 119.0, 68.0, 38.0, 33.0, 19.0, 13.0, 4.0, 5.0, 6.0, 3.0, 4.0, 2.0, 1.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.06640625, -0.06417083740234375, -0.0619354248046875, -0.05970001220703125, -0.057464599609375, -0.05522918701171875, -0.0529937744140625, -0.05075836181640625, -0.04852294921875, -0.04628753662109375, -0.0440521240234375, -0.04181671142578125, -0.039581298828125, -0.03734588623046875, -0.0351104736328125, -0.03287506103515625, -0.0306396484375, -0.02840423583984375, -0.0261688232421875, -0.02393341064453125, -0.021697998046875, -0.01946258544921875, -0.0172271728515625, -0.01499176025390625, -0.01275634765625, -0.01052093505859375, -0.0082855224609375, -0.00605010986328125, -0.003814697265625, -0.00157928466796875, 0.0006561279296875, 0.00289154052734375, 0.005126953125, 0.00736236572265625, 0.0095977783203125, 0.01183319091796875, 0.014068603515625, 0.01630401611328125, 0.0185394287109375, 0.02077484130859375, 0.02301025390625, 0.02524566650390625, 0.0274810791015625, 0.02971649169921875, 0.031951904296875, 0.03418731689453125, 0.0364227294921875, 0.03865814208984375, 0.0408935546875, 0.04312896728515625, 0.0453643798828125, 0.04759979248046875, 0.049835205078125, 0.05207061767578125, 0.0543060302734375, 0.05654144287109375, 0.05877685546875, 0.06101226806640625, 0.0632476806640625, 0.06548309326171875, 0.067718505859375, 0.06995391845703125, 0.0721893310546875, 0.07442474365234375, 0.07666015625]}, "parameters/model.layers.14.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 1.0, 3.0, 2.0, 0.0, 1.0, 1.0, 4.0, 8.0, 10.0, 13.0, 16.0, 23.0, 37.0, 44.0, 84.0, 108.0, 222.0, 317.0, 629.0, 1967.0, 6903.0, 30854.0, 129632.0, 469895.0, 1369866.0, 2977765.0, 4418564.0, 3921455.0, 2198619.0, 885179.0, 271895.0, 71703.0, 15121.0, 3882.0, 1171.0, 479.0, 217.0, 184.0, 110.0, 65.0, 40.0, 36.0, 21.0, 11.0, 13.0, 6.0, 10.0, 11.0, 2.0, 3.0, 1.0, 1.0, 1.0, 2.0, 1.0, 2.0, 1.0], "bins": [-0.2041015625, -0.19805908203125, -0.1920166015625, -0.18597412109375, -0.179931640625, -0.17388916015625, -0.1678466796875, -0.16180419921875, -0.15576171875, -0.14971923828125, -0.1436767578125, -0.13763427734375, -0.131591796875, -0.12554931640625, -0.1195068359375, -0.11346435546875, -0.107421875, -0.10137939453125, -0.0953369140625, -0.08929443359375, -0.083251953125, -0.07720947265625, -0.0711669921875, -0.06512451171875, -0.05908203125, -0.05303955078125, -0.0469970703125, -0.04095458984375, -0.034912109375, -0.02886962890625, -0.0228271484375, -0.01678466796875, -0.0107421875, -0.00469970703125, 0.0013427734375, 0.00738525390625, 0.013427734375, 0.01947021484375, 0.0255126953125, 0.03155517578125, 0.03759765625, 0.04364013671875, 0.0496826171875, 0.05572509765625, 0.061767578125, 0.06781005859375, 0.0738525390625, 0.07989501953125, 0.0859375, 0.09197998046875, 0.0980224609375, 0.10406494140625, 0.110107421875, 0.11614990234375, 0.1221923828125, 0.12823486328125, 0.13427734375, 0.14031982421875, 0.1463623046875, 0.15240478515625, 0.158447265625, 0.16448974609375, 0.1705322265625, 0.17657470703125, 0.1826171875]}, "parameters/model.layers.14.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 3.0, 6.0, 4.0, 12.0, 16.0, 15.0, 24.0, 30.0, 35.0, 54.0, 81.0, 93.0, 173.0, 254.0, 432.0, 827.0, 1592.0, 3784.0, 12964.0, 78903.0, 731665.0, 7210057.0, 16777216.0, 16777216.0, 16777216.0, 2904886.0, 278987.0, 34710.0, 7447.0, 2741.0, 1237.0, 586.0, 347.0, 220.0, 175.0, 99.0, 78.0, 47.0, 36.0, 34.0, 13.0, 16.0, 13.0, 9.0, 6.0, 3.0, 7.0, 1.0, 4.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.4140625, -0.400390625, -0.38671875, -0.373046875, -0.359375, -0.345703125, -0.33203125, -0.318359375, -0.3046875, -0.291015625, -0.27734375, -0.263671875, -0.25, -0.236328125, -0.22265625, -0.208984375, -0.1953125, -0.181640625, -0.16796875, -0.154296875, -0.140625, -0.126953125, -0.11328125, -0.099609375, -0.0859375, -0.072265625, -0.05859375, -0.044921875, -0.03125, -0.017578125, -0.00390625, 0.009765625, 0.0234375, 0.037109375, 0.05078125, 0.064453125, 0.078125, 0.091796875, 0.10546875, 0.119140625, 0.1328125, 0.146484375, 0.16015625, 0.173828125, 0.1875, 0.201171875, 0.21484375, 0.228515625, 0.2421875, 0.255859375, 0.26953125, 0.283203125, 0.296875, 0.310546875, 0.32421875, 0.337890625, 0.3515625, 0.365234375, 0.37890625, 0.392578125, 0.40625, 0.419921875, 0.43359375, 0.447265625, 0.4609375]}, "parameters/model.layers.14.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 4.0, 2.0, 2.0, 2.0, 6.0, 3.0, 12.0, 20.0, 25.0, 51.0, 119.0, 277.0, 898.0, 36591.0, 2400534.0, 16777216.0, 16777216.0, 3350388.0, 56274.0, 1127.0, 317.0, 127.0, 64.0, 37.0, 14.0, 8.0, 11.0, 4.0, 3.0, 2.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.7265625, -0.7088623046875, -0.691162109375, -0.6734619140625, -0.65576171875, -0.6380615234375, -0.620361328125, -0.6026611328125, -0.5849609375, -0.5672607421875, -0.549560546875, -0.5318603515625, -0.51416015625, -0.4964599609375, -0.478759765625, -0.4610595703125, -0.443359375, -0.4256591796875, -0.407958984375, -0.3902587890625, -0.37255859375, -0.3548583984375, -0.337158203125, -0.3194580078125, -0.3017578125, -0.2840576171875, -0.266357421875, -0.2486572265625, -0.23095703125, -0.2132568359375, -0.195556640625, -0.1778564453125, -0.16015625, -0.1424560546875, -0.124755859375, -0.1070556640625, -0.08935546875, -0.0716552734375, -0.053955078125, -0.0362548828125, -0.0185546875, -0.0008544921875, 0.016845703125, 0.0345458984375, 0.05224609375, 0.0699462890625, 0.087646484375, 0.1053466796875, 0.123046875, 0.1407470703125, 0.158447265625, 0.1761474609375, 0.19384765625, 0.2115478515625, 0.229248046875, 0.2469482421875, 0.2646484375, 0.2823486328125, 0.300048828125, 0.3177490234375, 0.33544921875, 0.3531494140625, 0.370849609375, 0.3885498046875, 0.40625]}, "parameters/model.layers.14.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 7.0, 23.0, 50.0, 97.0, 180.0, 258.0, 400.0, 522.0, 505.0, 545.0, 562.0, 324.0, 248.0, 133.0, 54.0, 47.0, 44.0, 14.0, 8.0, 4.0, 6.0, 8.0, 3.0, 2.0, 2.0, 1.0, 2.0, 1.0, 2.0, 1.0, 0.0, 1.0, 5.0, 1.0, 1.0, 1.0, 3.0, 3.0, 2.0, 4.0, 1.0, 1.0, 1.0, 3.0, 2.0, 1.0, 0.0, 2.0, 2.0], "bins": [0.11767578125, 0.13170623779296875, 0.1457366943359375, 0.15976715087890625, 0.173797607421875, 0.18782806396484375, 0.2018585205078125, 0.21588897705078125, 0.22991943359375, 0.24394989013671875, 0.2579803466796875, 0.27201080322265625, 0.286041259765625, 0.30007171630859375, 0.3141021728515625, 0.32813262939453125, 0.3421630859375, 0.35619354248046875, 0.3702239990234375, 0.38425445556640625, 0.398284912109375, 0.41231536865234375, 0.4263458251953125, 0.44037628173828125, 0.45440673828125, 0.46843719482421875, 0.4824676513671875, 0.49649810791015625, 0.510528564453125, 0.5245590209960938, 0.5385894775390625, 0.5526199340820312, 0.566650390625, 0.5806808471679688, 0.5947113037109375, 0.6087417602539062, 0.622772216796875, 0.6368026733398438, 0.6508331298828125, 0.6648635864257812, 0.67889404296875, 0.6929244995117188, 0.7069549560546875, 0.7209854125976562, 0.735015869140625, 0.7490463256835938, 0.7630767822265625, 0.7771072387695312, 0.7911376953125, 0.8051681518554688, 0.8191986083984375, 0.8332290649414062, 0.847259521484375, 0.8612899780273438, 0.8753204345703125, 0.8893508911132812, 0.90338134765625, 0.9174118041992188, 0.9314422607421875, 0.9454727172851562, 0.959503173828125, 0.9735336303710938, 0.9875640869140625, 1.0015945434570312, 1.015625]}, "parameters/model.layers.14.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 1.0, 2.0, 2.0, 4.0, 3.0, 3.0, 2.0, 4.0, 1.0, 2.0, 2.0, 2.0, 0.0, 2.0, 6.0, 2.0, 2.0, 6.0, 6.0, 11.0, 14.0, 47.0, 90.0, 149.0, 312.0, 548.0, 824.0, 944.0, 691.0, 314.0, 55.0, 19.0, 7.0, 3.0, 3.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.07861328125, 0.08467864990234375, 0.0907440185546875, 0.09680938720703125, 0.102874755859375, 0.10894012451171875, 0.1150054931640625, 0.12107086181640625, 0.12713623046875, 0.13320159912109375, 0.1392669677734375, 0.14533233642578125, 0.151397705078125, 0.15746307373046875, 0.1635284423828125, 0.16959381103515625, 0.1756591796875, 0.18172454833984375, 0.1877899169921875, 0.19385528564453125, 0.199920654296875, 0.20598602294921875, 0.2120513916015625, 0.21811676025390625, 0.22418212890625, 0.23024749755859375, 0.2363128662109375, 0.24237823486328125, 0.248443603515625, 0.25450897216796875, 0.2605743408203125, 0.26663970947265625, 0.272705078125, 0.27877044677734375, 0.2848358154296875, 0.29090118408203125, 0.296966552734375, 0.30303192138671875, 0.3090972900390625, 0.31516265869140625, 0.32122802734375, 0.32729339599609375, 0.3333587646484375, 0.33942413330078125, 0.345489501953125, 0.35155487060546875, 0.3576202392578125, 0.36368560791015625, 0.3697509765625, 0.37581634521484375, 0.3818817138671875, 0.38794708251953125, 0.394012451171875, 0.40007781982421875, 0.4061431884765625, 0.41220855712890625, 0.41827392578125, 0.42433929443359375, 0.4304046630859375, 0.43647003173828125, 0.442535400390625, 0.44860076904296875, 0.4546661376953125, 0.46073150634765625, 0.466796875]}, "parameters/model.layers.15.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 2.0, 6.0, 5.0, 5.0, 7.0, 7.0, 14.0, 23.0, 23.0, 43.0, 52.0, 110.0, 164.0, 335.0, 808.0, 1922.0, 5836.0, 15491.0, 45760.0, 140274.0, 396563.0, 1171169.0, 3081443.0, 5613935.0, 3900336.0, 1567153.0, 548126.0, 190858.0, 63083.0, 21773.0, 7409.0, 2565.0, 1005.0, 400.0, 191.0, 90.0, 62.0, 37.0, 30.0, 22.0, 14.0, 13.0, 12.0, 10.0, 5.0, 2.0, 6.0, 1.0, 0.0, 4.0, 2.0, 2.0, 0.0, 2.0], "bins": [-0.419921875, -0.407440185546875, -0.39495849609375, -0.382476806640625, -0.3699951171875, -0.357513427734375, -0.34503173828125, -0.332550048828125, -0.320068359375, -0.307586669921875, -0.29510498046875, -0.282623291015625, -0.2701416015625, -0.257659912109375, -0.24517822265625, -0.232696533203125, -0.22021484375, -0.207733154296875, -0.19525146484375, -0.182769775390625, -0.1702880859375, -0.157806396484375, -0.14532470703125, -0.132843017578125, -0.120361328125, -0.107879638671875, -0.09539794921875, -0.082916259765625, -0.0704345703125, -0.057952880859375, -0.04547119140625, -0.032989501953125, -0.0205078125, -0.008026123046875, 0.00445556640625, 0.016937255859375, 0.0294189453125, 0.041900634765625, 0.05438232421875, 0.066864013671875, 0.079345703125, 0.091827392578125, 0.10430908203125, 0.116790771484375, 0.1292724609375, 0.141754150390625, 0.15423583984375, 0.166717529296875, 0.17919921875, 0.191680908203125, 0.20416259765625, 0.216644287109375, 0.2291259765625, 0.241607666015625, 0.25408935546875, 0.266571044921875, 0.279052734375, 0.291534423828125, 0.30401611328125, 0.316497802734375, 0.3289794921875, 0.341461181640625, 0.35394287109375, 0.366424560546875, 0.37890625]}, "parameters/model.layers.15.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 8.0, 3.0, 3.0, 11.0, 7.0, 19.0, 19.0, 33.0, 60.0, 94.0, 178.0, 232.0, 496.0, 1118.0, 2699.0, 6354.0, 14162.0, 32964.0, 70984.0, 147529.0, 273402.0, 477354.0, 702010.0, 795887.0, 680511.0, 461850.0, 262964.0, 140736.0, 67934.0, 30535.0, 13634.0, 5895.0, 2438.0, 1086.0, 485.0, 242.0, 142.0, 78.0, 45.0, 32.0, 21.0, 17.0, 9.0, 8.0, 6.0, 0.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.3828125, -0.3717041015625, -0.360595703125, -0.3494873046875, -0.33837890625, -0.3272705078125, -0.316162109375, -0.3050537109375, -0.2939453125, -0.2828369140625, -0.271728515625, -0.2606201171875, -0.24951171875, -0.2384033203125, -0.227294921875, -0.2161865234375, -0.205078125, -0.1939697265625, -0.182861328125, -0.1717529296875, -0.16064453125, -0.1495361328125, -0.138427734375, -0.1273193359375, -0.1162109375, -0.1051025390625, -0.093994140625, -0.0828857421875, -0.07177734375, -0.0606689453125, -0.049560546875, -0.0384521484375, -0.02734375, -0.0162353515625, -0.005126953125, 0.0059814453125, 0.01708984375, 0.0281982421875, 0.039306640625, 0.0504150390625, 0.0615234375, 0.0726318359375, 0.083740234375, 0.0948486328125, 0.10595703125, 0.1170654296875, 0.128173828125, 0.1392822265625, 0.150390625, 0.1614990234375, 0.172607421875, 0.1837158203125, 0.19482421875, 0.2059326171875, 0.217041015625, 0.2281494140625, 0.2392578125, 0.2503662109375, 0.261474609375, 0.2725830078125, 0.28369140625, 0.2947998046875, 0.305908203125, 0.3170166015625, 0.328125]}, "parameters/model.layers.15.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 5.0, 5.0, 10.0, 15.0, 33.0, 76.0, 147.0, 338.0, 775.0, 1896.0, 4608.0, 10517.0, 23600.0, 50205.0, 101740.0, 198702.0, 347271.0, 543316.0, 712492.0, 735915.0, 591945.0, 396626.0, 236154.0, 122788.0, 61280.0, 29975.0, 13447.0, 5800.0, 2663.0, 1097.0, 477.0, 203.0, 95.0, 36.0, 15.0, 11.0, 7.0, 6.0, 0.0, 3.0, 2.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.125, -0.121337890625, -0.11767578125, -0.114013671875, -0.1103515625, -0.106689453125, -0.10302734375, -0.099365234375, -0.095703125, -0.092041015625, -0.08837890625, -0.084716796875, -0.0810546875, -0.077392578125, -0.07373046875, -0.070068359375, -0.06640625, -0.062744140625, -0.05908203125, -0.055419921875, -0.0517578125, -0.048095703125, -0.04443359375, -0.040771484375, -0.037109375, -0.033447265625, -0.02978515625, -0.026123046875, -0.0224609375, -0.018798828125, -0.01513671875, -0.011474609375, -0.0078125, -0.004150390625, -0.00048828125, 0.003173828125, 0.0068359375, 0.010498046875, 0.01416015625, 0.017822265625, 0.021484375, 0.025146484375, 0.02880859375, 0.032470703125, 0.0361328125, 0.039794921875, 0.04345703125, 0.047119140625, 0.05078125, 0.054443359375, 0.05810546875, 0.061767578125, 0.0654296875, 0.069091796875, 0.07275390625, 0.076416015625, 0.080078125, 0.083740234375, 0.08740234375, 0.091064453125, 0.0947265625, 0.098388671875, 0.10205078125, 0.105712890625, 0.109375]}, "parameters/model.layers.15.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 3.0, 2.0, 2.0, 3.0, 1.0, 5.0, 3.0, 5.0, 5.0, 7.0, 5.0, 5.0, 13.0, 17.0, 39.0, 91.0, 206.0, 859.0, 7926.0, 113657.0, 1395353.0, 6803236.0, 6869246.0, 1457149.0, 119881.0, 8303.0, 819.0, 186.0, 86.0, 31.0, 13.0, 12.0, 5.0, 4.0, 2.0, 4.0, 2.0, 2.0, 4.0, 1.0, 1.0, 2.0, 2.0, 3.0, 1.0, 1.0, 3.0, 1.0, 0.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.416015625, -0.403411865234375, -0.39080810546875, -0.378204345703125, -0.3656005859375, -0.352996826171875, -0.34039306640625, -0.327789306640625, -0.315185546875, -0.302581787109375, -0.28997802734375, -0.277374267578125, -0.2647705078125, -0.252166748046875, -0.23956298828125, -0.226959228515625, -0.21435546875, -0.201751708984375, -0.18914794921875, -0.176544189453125, -0.1639404296875, -0.151336669921875, -0.13873291015625, -0.126129150390625, -0.113525390625, -0.100921630859375, -0.08831787109375, -0.075714111328125, -0.0631103515625, -0.050506591796875, -0.03790283203125, -0.025299072265625, -0.0126953125, -9.1552734375e-05, 0.01251220703125, 0.025115966796875, 0.0377197265625, 0.050323486328125, 0.06292724609375, 0.075531005859375, 0.088134765625, 0.100738525390625, 0.11334228515625, 0.125946044921875, 0.1385498046875, 0.151153564453125, 0.16375732421875, 0.176361083984375, 0.18896484375, 0.201568603515625, 0.21417236328125, 0.226776123046875, 0.2393798828125, 0.251983642578125, 0.26458740234375, 0.277191162109375, 0.289794921875, 0.302398681640625, 0.31500244140625, 0.327606201171875, 0.3402099609375, 0.352813720703125, 0.36541748046875, 0.378021240234375, 0.390625]}, "parameters/model.layers.15.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 1.0, 1.0, 1.0, 1.0, 3.0, 3.0, 8.0, 7.0, 11.0, 11.0, 11.0, 16.0, 26.0, 34.0, 48.0, 57.0, 78.0, 129.0, 147.0, 271.0, 429.0, 738.0, 1340.0, 2692.0, 6128.0, 19530.0, 93586.0, 573504.0, 3879540.0, 16777216.0, 16777216.0, 16777216.0, 14069488.0, 2795984.0, 407556.0, 68379.0, 15251.0, 5089.0, 2271.0, 1209.0, 665.0, 371.0, 248.0, 165.0, 110.0, 63.0, 78.0, 53.0, 32.0, 35.0, 20.0, 11.0, 10.0, 14.0, 4.0, 5.0, 6.0, 3.0, 4.0, 2.0, 3.0, 0.0, 1.0], "bins": [-0.359375, -0.348114013671875, -0.33685302734375, -0.325592041015625, -0.3143310546875, -0.303070068359375, -0.29180908203125, -0.280548095703125, -0.269287109375, -0.258026123046875, -0.24676513671875, -0.235504150390625, -0.2242431640625, -0.212982177734375, -0.20172119140625, -0.190460205078125, -0.17919921875, -0.167938232421875, -0.15667724609375, -0.145416259765625, -0.1341552734375, -0.122894287109375, -0.11163330078125, -0.100372314453125, -0.089111328125, -0.077850341796875, -0.06658935546875, -0.055328369140625, -0.0440673828125, -0.032806396484375, -0.02154541015625, -0.010284423828125, 0.0009765625, 0.012237548828125, 0.02349853515625, 0.034759521484375, 0.0460205078125, 0.057281494140625, 0.06854248046875, 0.079803466796875, 0.091064453125, 0.102325439453125, 0.11358642578125, 0.124847412109375, 0.1361083984375, 0.147369384765625, 0.15863037109375, 0.169891357421875, 0.18115234375, 0.192413330078125, 0.20367431640625, 0.214935302734375, 0.2261962890625, 0.237457275390625, 0.24871826171875, 0.259979248046875, 0.271240234375, 0.282501220703125, 0.29376220703125, 0.305023193359375, 0.3162841796875, 0.327545166015625, 0.33880615234375, 0.350067138671875, 0.361328125]}, "parameters/model.layers.15.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 2.0, 3.0, 2.0, 7.0, 3.0, 9.0, 19.0, 37.0, 59.0, 139.0, 251.0, 543.0, 3932.0, 370761.0, 9865943.0, 16777216.0, 13158200.0, 648941.0, 6514.0, 670.0, 255.0, 136.0, 68.0, 33.0, 16.0, 10.0, 5.0, 6.0, 4.0, 3.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.5625, -0.54522705078125, -0.5279541015625, -0.51068115234375, -0.493408203125, -0.47613525390625, -0.4588623046875, -0.44158935546875, -0.42431640625, -0.40704345703125, -0.3897705078125, -0.37249755859375, -0.355224609375, -0.33795166015625, -0.3206787109375, -0.30340576171875, -0.2861328125, -0.26885986328125, -0.2515869140625, -0.23431396484375, -0.217041015625, -0.19976806640625, -0.1824951171875, -0.16522216796875, -0.14794921875, -0.13067626953125, -0.1134033203125, -0.09613037109375, -0.078857421875, -0.06158447265625, -0.0443115234375, -0.02703857421875, -0.009765625, 0.00750732421875, 0.0247802734375, 0.04205322265625, 0.059326171875, 0.07659912109375, 0.0938720703125, 0.11114501953125, 0.12841796875, 0.14569091796875, 0.1629638671875, 0.18023681640625, 0.197509765625, 0.21478271484375, 0.2320556640625, 0.24932861328125, 0.2666015625, 0.28387451171875, 0.3011474609375, 0.31842041015625, 0.335693359375, 0.35296630859375, 0.3702392578125, 0.38751220703125, 0.40478515625, 0.42205810546875, 0.4393310546875, 0.45660400390625, 0.473876953125, 0.49114990234375, 0.5084228515625, 0.52569580078125, 0.54296875]}, "parameters/model.layers.15.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 3.0, 4.0, 13.0, 15.0, 40.0, 68.0, 152.0, 372.0, 663.0, 844.0, 870.0, 462.0, 296.0, 143.0, 49.0, 24.0, 10.0, 5.0, 1.0, 5.0, 4.0, 3.0, 6.0, 1.0, 2.0, 0.0, 0.0, 0.0, 3.0, 3.0, 4.0, 1.0, 2.0, 1.0, 2.0, 0.0, 4.0, 6.0, 2.0, 1.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.1064453125, 0.1197967529296875, 0.133148193359375, 0.1464996337890625, 0.15985107421875, 0.1732025146484375, 0.186553955078125, 0.1999053955078125, 0.2132568359375, 0.2266082763671875, 0.239959716796875, 0.2533111572265625, 0.26666259765625, 0.2800140380859375, 0.293365478515625, 0.3067169189453125, 0.320068359375, 0.3334197998046875, 0.346771240234375, 0.3601226806640625, 0.37347412109375, 0.3868255615234375, 0.400177001953125, 0.4135284423828125, 0.4268798828125, 0.4402313232421875, 0.453582763671875, 0.4669342041015625, 0.48028564453125, 0.4936370849609375, 0.506988525390625, 0.5203399658203125, 0.53369140625, 0.5470428466796875, 0.560394287109375, 0.5737457275390625, 0.58709716796875, 0.6004486083984375, 0.613800048828125, 0.6271514892578125, 0.6405029296875, 0.6538543701171875, 0.667205810546875, 0.6805572509765625, 0.69390869140625, 0.7072601318359375, 0.720611572265625, 0.7339630126953125, 0.747314453125, 0.7606658935546875, 0.774017333984375, 0.7873687744140625, 0.80072021484375, 0.8140716552734375, 0.827423095703125, 0.8407745361328125, 0.8541259765625, 0.8674774169921875, 0.880828857421875, 0.8941802978515625, 0.90753173828125, 0.9208831787109375, 0.934234619140625, 0.9475860595703125, 0.9609375]}, "parameters/model.layers.15.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 4.0, 6.0, 1.0, 2.0, 3.0, 6.0, 2.0, 3.0, 1.0, 2.0, 6.0, 2.0, 2.0, 3.0, 2.0, 6.0, 8.0, 11.0, 16.0, 24.0, 72.0, 116.0, 131.0, 318.0, 515.0, 1183.0, 1186.0, 347.0, 88.0, 8.0, 2.0, 3.0, 1.0, 2.0, 0.0, 0.0, 1.0, 2.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [0.0830078125, 0.0900726318359375, 0.097137451171875, 0.1042022705078125, 0.11126708984375, 0.1183319091796875, 0.125396728515625, 0.1324615478515625, 0.1395263671875, 0.1465911865234375, 0.153656005859375, 0.1607208251953125, 0.16778564453125, 0.1748504638671875, 0.181915283203125, 0.1889801025390625, 0.196044921875, 0.2031097412109375, 0.210174560546875, 0.2172393798828125, 0.22430419921875, 0.2313690185546875, 0.238433837890625, 0.2454986572265625, 0.2525634765625, 0.2596282958984375, 0.266693115234375, 0.2737579345703125, 0.28082275390625, 0.2878875732421875, 0.294952392578125, 0.3020172119140625, 0.30908203125, 0.3161468505859375, 0.323211669921875, 0.3302764892578125, 0.33734130859375, 0.3444061279296875, 0.351470947265625, 0.3585357666015625, 0.3656005859375, 0.3726654052734375, 0.379730224609375, 0.3867950439453125, 0.39385986328125, 0.4009246826171875, 0.407989501953125, 0.4150543212890625, 0.422119140625, 0.4291839599609375, 0.436248779296875, 0.4433135986328125, 0.45037841796875, 0.4574432373046875, 0.464508056640625, 0.4715728759765625, 0.4786376953125, 0.4857025146484375, 0.492767333984375, 0.4998321533203125, 0.50689697265625, 0.5139617919921875, 0.521026611328125, 0.5280914306640625, 0.53515625]}, "parameters/model.layers.16.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 1.0, 2.0, 1.0, 6.0, 6.0, 7.0, 3.0, 15.0, 18.0, 12.0, 21.0, 23.0, 29.0, 49.0, 48.0, 73.0, 97.0, 122.0, 279.0, 606.0, 1845.0, 6227.0, 20381.0, 73151.0, 237980.0, 750329.0, 2028380.0, 4285492.0, 5017719.0, 2728812.0, 1101463.0, 364471.0, 112835.0, 32750.0, 9338.0, 2736.0, 899.0, 385.0, 167.0, 106.0, 86.0, 55.0, 54.0, 27.0, 28.0, 23.0, 10.0, 10.0, 6.0, 8.0, 8.0, 2.0, 2.0, 5.0, 0.0, 0.0, 1.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.369140625, -0.3572998046875, -0.345458984375, -0.3336181640625, -0.32177734375, -0.3099365234375, -0.298095703125, -0.2862548828125, -0.2744140625, -0.2625732421875, -0.250732421875, -0.2388916015625, -0.22705078125, -0.2152099609375, -0.203369140625, -0.1915283203125, -0.1796875, -0.1678466796875, -0.156005859375, -0.1441650390625, -0.13232421875, -0.1204833984375, -0.108642578125, -0.0968017578125, -0.0849609375, -0.0731201171875, -0.061279296875, -0.0494384765625, -0.03759765625, -0.0257568359375, -0.013916015625, -0.0020751953125, 0.009765625, 0.0216064453125, 0.033447265625, 0.0452880859375, 0.05712890625, 0.0689697265625, 0.080810546875, 0.0926513671875, 0.1044921875, 0.1163330078125, 0.128173828125, 0.1400146484375, 0.15185546875, 0.1636962890625, 0.175537109375, 0.1873779296875, 0.19921875, 0.2110595703125, 0.222900390625, 0.2347412109375, 0.24658203125, 0.2584228515625, 0.270263671875, 0.2821044921875, 0.2939453125, 0.3057861328125, 0.317626953125, 0.3294677734375, 0.34130859375, 0.3531494140625, 0.364990234375, 0.3768310546875, 0.388671875]}, "parameters/model.layers.16.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 1.0, 2.0, 4.0, 2.0, 2.0, 4.0, 9.0, 9.0, 9.0, 17.0, 24.0, 30.0, 39.0, 65.0, 118.0, 196.0, 359.0, 643.0, 1452.0, 2854.0, 6042.0, 11754.0, 24678.0, 52204.0, 103743.0, 191530.0, 357032.0, 553792.0, 731679.0, 743896.0, 587986.0, 381852.0, 213789.0, 114914.0, 60232.0, 26443.0, 13711.0, 6505.0, 3317.0, 1616.0, 742.0, 379.0, 218.0, 145.0, 76.0, 45.0, 41.0, 32.0, 21.0, 8.0, 9.0, 8.0, 4.0, 7.0, 3.0, 4.0, 0.0, 0.0, 2.0, 0.0, 1.0, 2.0], "bins": [-0.328125, -0.31756591796875, -0.3070068359375, -0.29644775390625, -0.285888671875, -0.27532958984375, -0.2647705078125, -0.25421142578125, -0.24365234375, -0.23309326171875, -0.2225341796875, -0.21197509765625, -0.201416015625, -0.19085693359375, -0.1802978515625, -0.16973876953125, -0.1591796875, -0.14862060546875, -0.1380615234375, -0.12750244140625, -0.116943359375, -0.10638427734375, -0.0958251953125, -0.08526611328125, -0.07470703125, -0.06414794921875, -0.0535888671875, -0.04302978515625, -0.032470703125, -0.02191162109375, -0.0113525390625, -0.00079345703125, 0.009765625, 0.02032470703125, 0.0308837890625, 0.04144287109375, 0.052001953125, 0.06256103515625, 0.0731201171875, 0.08367919921875, 0.09423828125, 0.10479736328125, 0.1153564453125, 0.12591552734375, 0.136474609375, 0.14703369140625, 0.1575927734375, 0.16815185546875, 0.1787109375, 0.18927001953125, 0.1998291015625, 0.21038818359375, 0.220947265625, 0.23150634765625, 0.2420654296875, 0.25262451171875, 0.26318359375, 0.27374267578125, 0.2843017578125, 0.29486083984375, 0.305419921875, 0.31597900390625, 0.3265380859375, 0.33709716796875, 0.34765625]}, "parameters/model.layers.16.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 3.0, 6.0, 3.0, 7.0, 16.0, 19.0, 30.0, 58.0, 70.0, 89.0, 149.0, 271.0, 505.0, 1244.0, 2860.0, 6227.0, 15616.0, 34850.0, 69078.0, 144601.0, 251941.0, 409000.0, 567032.0, 682525.0, 661581.0, 532895.0, 359256.0, 225063.0, 119148.0, 60364.0, 28003.0, 12116.0, 5344.0, 2276.0, 911.0, 509.0, 248.0, 134.0, 78.0, 58.0, 51.0, 16.0, 15.0, 10.0, 9.0, 6.0, 2.0, 1.0, 2.0, 1.0, 3.0], "bins": [-0.1201171875, -0.11685943603515625, -0.1136016845703125, -0.11034393310546875, -0.107086181640625, -0.10382843017578125, -0.1005706787109375, -0.09731292724609375, -0.09405517578125, -0.09079742431640625, -0.0875396728515625, -0.08428192138671875, -0.081024169921875, -0.07776641845703125, -0.0745086669921875, -0.07125091552734375, -0.0679931640625, -0.06473541259765625, -0.0614776611328125, -0.05821990966796875, -0.054962158203125, -0.05170440673828125, -0.0484466552734375, -0.04518890380859375, -0.04193115234375, -0.03867340087890625, -0.0354156494140625, -0.03215789794921875, -0.028900146484375, -0.02564239501953125, -0.0223846435546875, -0.01912689208984375, -0.015869140625, -0.01261138916015625, -0.0093536376953125, -0.00609588623046875, -0.002838134765625, 0.00041961669921875, 0.0036773681640625, 0.00693511962890625, 0.01019287109375, 0.01345062255859375, 0.0167083740234375, 0.01996612548828125, 0.023223876953125, 0.02648162841796875, 0.0297393798828125, 0.03299713134765625, 0.0362548828125, 0.03951263427734375, 0.0427703857421875, 0.04602813720703125, 0.049285888671875, 0.05254364013671875, 0.0558013916015625, 0.05905914306640625, 0.06231689453125, 0.06557464599609375, 0.0688323974609375, 0.07209014892578125, 0.075347900390625, 0.07860565185546875, 0.0818634033203125, 0.08512115478515625, 0.08837890625]}, "parameters/model.layers.16.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 5.0, 2.0, 3.0, 1.0, 2.0, 3.0, 1.0, 2.0, 1.0, 1.0, 2.0, 6.0, 4.0, 7.0, 5.0, 11.0, 8.0, 23.0, 46.0, 78.0, 158.0, 407.0, 3114.0, 60940.0, 1167901.0, 6817803.0, 7239861.0, 1403543.0, 78740.0, 3670.0, 504.0, 182.0, 75.0, 33.0, 14.0, 15.0, 3.0, 5.0, 2.0, 2.0, 3.0, 3.0, 2.0, 4.0, 4.0, 2.0, 2.0, 2.0, 2.0, 1.0, 1.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.408203125, -0.395050048828125, -0.38189697265625, -0.368743896484375, -0.3555908203125, -0.342437744140625, -0.32928466796875, -0.316131591796875, -0.302978515625, -0.289825439453125, -0.27667236328125, -0.263519287109375, -0.2503662109375, -0.237213134765625, -0.22406005859375, -0.210906982421875, -0.19775390625, -0.184600830078125, -0.17144775390625, -0.158294677734375, -0.1451416015625, -0.131988525390625, -0.11883544921875, -0.105682373046875, -0.092529296875, -0.079376220703125, -0.06622314453125, -0.053070068359375, -0.0399169921875, -0.026763916015625, -0.01361083984375, -0.000457763671875, 0.0126953125, 0.025848388671875, 0.03900146484375, 0.052154541015625, 0.0653076171875, 0.078460693359375, 0.09161376953125, 0.104766845703125, 0.117919921875, 0.131072998046875, 0.14422607421875, 0.157379150390625, 0.1705322265625, 0.183685302734375, 0.19683837890625, 0.209991455078125, 0.22314453125, 0.236297607421875, 0.24945068359375, 0.262603759765625, 0.2757568359375, 0.288909912109375, 0.30206298828125, 0.315216064453125, 0.328369140625, 0.341522216796875, 0.35467529296875, 0.367828369140625, 0.3809814453125, 0.394134521484375, 0.40728759765625, 0.420440673828125, 0.43359375]}, "parameters/model.layers.16.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 2.0, 1.0, 3.0, 2.0, 5.0, 5.0, 7.0, 10.0, 12.0, 18.0, 36.0, 56.0, 87.0, 140.0, 223.0, 477.0, 1007.0, 2490.0, 6242.0, 22532.0, 119371.0, 769644.0, 5415075.0, 16777216.0, 16777216.0, 16777216.0, 8580611.0, 1329388.0, 195247.0, 35157.0, 8266.0, 3169.0, 1228.0, 634.0, 297.0, 180.0, 90.0, 48.0, 36.0, 20.0, 15.0, 17.0, 7.0, 9.0, 6.0, 4.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.306640625, -0.294677734375, -0.28271484375, -0.270751953125, -0.2587890625, -0.246826171875, -0.23486328125, -0.222900390625, -0.2109375, -0.198974609375, -0.18701171875, -0.175048828125, -0.1630859375, -0.151123046875, -0.13916015625, -0.127197265625, -0.115234375, -0.103271484375, -0.09130859375, -0.079345703125, -0.0673828125, -0.055419921875, -0.04345703125, -0.031494140625, -0.01953125, -0.007568359375, 0.00439453125, 0.016357421875, 0.0283203125, 0.040283203125, 0.05224609375, 0.064208984375, 0.076171875, 0.088134765625, 0.10009765625, 0.112060546875, 0.1240234375, 0.135986328125, 0.14794921875, 0.159912109375, 0.171875, 0.183837890625, 0.19580078125, 0.207763671875, 0.2197265625, 0.231689453125, 0.24365234375, 0.255615234375, 0.267578125, 0.279541015625, 0.29150390625, 0.303466796875, 0.3154296875, 0.327392578125, 0.33935546875, 0.351318359375, 0.36328125, 0.375244140625, 0.38720703125, 0.399169921875, 0.4111328125, 0.423095703125, 0.43505859375, 0.447021484375, 0.458984375]}, "parameters/model.layers.16.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 3.0, 0.0, 2.0, 7.0, 8.0, 17.0, 33.0, 67.0, 154.0, 344.0, 2091.0, 403332.0, 15312251.0, 16777216.0, 5651002.0, 60604.0, 883.0, 278.0, 110.0, 47.0, 28.0, 15.0, 12.0, 10.0, 3.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.72265625, -0.7027587890625, -0.682861328125, -0.6629638671875, -0.64306640625, -0.6231689453125, -0.603271484375, -0.5833740234375, -0.5634765625, -0.5435791015625, -0.523681640625, -0.5037841796875, -0.48388671875, -0.4639892578125, -0.444091796875, -0.4241943359375, -0.404296875, -0.3843994140625, -0.364501953125, -0.3446044921875, -0.32470703125, -0.3048095703125, -0.284912109375, -0.2650146484375, -0.2451171875, -0.2252197265625, -0.205322265625, -0.1854248046875, -0.16552734375, -0.1456298828125, -0.125732421875, -0.1058349609375, -0.0859375, -0.0660400390625, -0.046142578125, -0.0262451171875, -0.00634765625, 0.0135498046875, 0.033447265625, 0.0533447265625, 0.0732421875, 0.0931396484375, 0.113037109375, 0.1329345703125, 0.15283203125, 0.1727294921875, 0.192626953125, 0.2125244140625, 0.232421875, 0.2523193359375, 0.272216796875, 0.2921142578125, 0.31201171875, 0.3319091796875, 0.351806640625, 0.3717041015625, 0.3916015625, 0.4114990234375, 0.431396484375, 0.4512939453125, 0.47119140625, 0.4910888671875, 0.510986328125, 0.5308837890625, 0.55078125]}, "parameters/model.layers.16.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 5.0, 19.0, 96.0, 501.0, 1064.0, 1089.0, 722.0, 252.0, 129.0, 56.0, 43.0, 10.0, 14.0, 11.0, 8.0, 5.0, 5.0, 2.0, 2.0, 4.0, 2.0, 4.0, 2.0, 1.0, 4.0, 2.0, 2.0, 3.0, 3.0, 1.0, 2.0, 2.0, 1.0, 4.0, 5.0, 3.0, 1.0, 5.0, 1.0, 2.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [0.0830078125, 0.0973358154296875, 0.111663818359375, 0.1259918212890625, 0.14031982421875, 0.1546478271484375, 0.168975830078125, 0.1833038330078125, 0.1976318359375, 0.2119598388671875, 0.226287841796875, 0.2406158447265625, 0.25494384765625, 0.2692718505859375, 0.283599853515625, 0.2979278564453125, 0.312255859375, 0.3265838623046875, 0.340911865234375, 0.3552398681640625, 0.36956787109375, 0.3838958740234375, 0.398223876953125, 0.4125518798828125, 0.4268798828125, 0.4412078857421875, 0.455535888671875, 0.4698638916015625, 0.48419189453125, 0.4985198974609375, 0.512847900390625, 0.5271759033203125, 0.54150390625, 0.5558319091796875, 0.570159912109375, 0.5844879150390625, 0.59881591796875, 0.6131439208984375, 0.627471923828125, 0.6417999267578125, 0.6561279296875, 0.6704559326171875, 0.684783935546875, 0.6991119384765625, 0.71343994140625, 0.7277679443359375, 0.742095947265625, 0.7564239501953125, 0.770751953125, 0.7850799560546875, 0.799407958984375, 0.8137359619140625, 0.82806396484375, 0.8423919677734375, 0.856719970703125, 0.8710479736328125, 0.8853759765625, 0.8997039794921875, 0.914031982421875, 0.9283599853515625, 0.94268798828125, 0.9570159912109375, 0.971343994140625, 0.9856719970703125, 1.0]}, "parameters/model.layers.16.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 2.0, 1.0, 1.0, 1.0, 4.0, 4.0, 2.0, 2.0, 4.0, 6.0, 1.0, 2.0, 0.0, 1.0, 4.0, 3.0, 3.0, 0.0, 4.0, 4.0, 7.0, 3.0, 7.0, 8.0, 15.0, 25.0, 24.0, 32.0, 51.0, 52.0, 70.0, 65.0, 133.0, 192.0, 283.0, 804.0, 894.0, 967.0, 300.0, 78.0, 21.0, 2.0, 3.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [0.0888671875, 0.0955352783203125, 0.102203369140625, 0.1088714599609375, 0.11553955078125, 0.1222076416015625, 0.128875732421875, 0.1355438232421875, 0.1422119140625, 0.1488800048828125, 0.155548095703125, 0.1622161865234375, 0.16888427734375, 0.1755523681640625, 0.182220458984375, 0.1888885498046875, 0.195556640625, 0.2022247314453125, 0.208892822265625, 0.2155609130859375, 0.22222900390625, 0.2288970947265625, 0.235565185546875, 0.2422332763671875, 0.2489013671875, 0.2555694580078125, 0.262237548828125, 0.2689056396484375, 0.27557373046875, 0.2822418212890625, 0.288909912109375, 0.2955780029296875, 0.30224609375, 0.3089141845703125, 0.315582275390625, 0.3222503662109375, 0.32891845703125, 0.3355865478515625, 0.342254638671875, 0.3489227294921875, 0.3555908203125, 0.3622589111328125, 0.368927001953125, 0.3755950927734375, 0.38226318359375, 0.3889312744140625, 0.395599365234375, 0.4022674560546875, 0.408935546875, 0.4156036376953125, 0.422271728515625, 0.4289398193359375, 0.43560791015625, 0.4422760009765625, 0.448944091796875, 0.4556121826171875, 0.4622802734375, 0.4689483642578125, 0.475616455078125, 0.4822845458984375, 0.48895263671875, 0.4956207275390625, 0.502288818359375, 0.5089569091796875, 0.515625]}, "parameters/model.layers.17.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 3.0, 1.0, 8.0, 3.0, 19.0, 21.0, 37.0, 47.0, 73.0, 124.0, 203.0, 412.0, 665.0, 1277.0, 2483.0, 4530.0, 7648.0, 15519.0, 28749.0, 56712.0, 100435.0, 184881.0, 312492.0, 559902.0, 914949.0, 1471955.0, 2186619.0, 2826137.0, 2727902.0, 2036589.0, 1338025.0, 841339.0, 506691.0, 290590.0, 160878.0, 93696.0, 49887.0, 26534.0, 13380.0, 7349.0, 3755.0, 2145.0, 1110.0, 576.0, 341.0, 193.0, 138.0, 75.0, 34.0, 21.0, 22.0, 11.0, 6.0, 11.0, 2.0, 2.0, 0.0, 1.0, 3.0, 1.0, 1.0], "bins": [-0.1923828125, -0.1861572265625, -0.179931640625, -0.1737060546875, -0.16748046875, -0.1612548828125, -0.155029296875, -0.1488037109375, -0.142578125, -0.1363525390625, -0.130126953125, -0.1239013671875, -0.11767578125, -0.1114501953125, -0.105224609375, -0.0989990234375, -0.0927734375, -0.0865478515625, -0.080322265625, -0.0740966796875, -0.06787109375, -0.0616455078125, -0.055419921875, -0.0491943359375, -0.04296875, -0.0367431640625, -0.030517578125, -0.0242919921875, -0.01806640625, -0.0118408203125, -0.005615234375, 0.0006103515625, 0.0068359375, 0.0130615234375, 0.019287109375, 0.0255126953125, 0.03173828125, 0.0379638671875, 0.044189453125, 0.0504150390625, 0.056640625, 0.0628662109375, 0.069091796875, 0.0753173828125, 0.08154296875, 0.0877685546875, 0.093994140625, 0.1002197265625, 0.1064453125, 0.1126708984375, 0.118896484375, 0.1251220703125, 0.13134765625, 0.1375732421875, 0.143798828125, 0.1500244140625, 0.15625, 0.1624755859375, 0.168701171875, 0.1749267578125, 0.18115234375, 0.1873779296875, 0.193603515625, 0.1998291015625, 0.2060546875]}, "parameters/model.layers.17.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 0.0, 3.0, 1.0, 6.0, 10.0, 3.0, 11.0, 16.0, 13.0, 15.0, 20.0, 39.0, 67.0, 67.0, 98.0, 180.0, 287.0, 510.0, 1083.0, 2315.0, 5029.0, 11810.0, 23733.0, 52690.0, 109214.0, 197842.0, 366166.0, 575649.0, 760414.0, 754512.0, 570815.0, 362968.0, 197892.0, 104342.0, 52568.0, 23042.0, 11155.0, 5215.0, 2292.0, 977.0, 529.0, 236.0, 154.0, 81.0, 60.0, 41.0, 41.0, 24.0, 13.0, 9.0, 11.0, 14.0, 4.0, 6.0, 2.0, 3.0, 0.0, 2.0, 0.0, 2.0], "bins": [-0.359375, -0.348480224609375, -0.33758544921875, -0.326690673828125, -0.3157958984375, -0.304901123046875, -0.29400634765625, -0.283111572265625, -0.272216796875, -0.261322021484375, -0.25042724609375, -0.239532470703125, -0.2286376953125, -0.217742919921875, -0.20684814453125, -0.195953369140625, -0.18505859375, -0.174163818359375, -0.16326904296875, -0.152374267578125, -0.1414794921875, -0.130584716796875, -0.11968994140625, -0.108795166015625, -0.097900390625, -0.087005615234375, -0.07611083984375, -0.065216064453125, -0.0543212890625, -0.043426513671875, -0.03253173828125, -0.021636962890625, -0.0107421875, 0.000152587890625, 0.01104736328125, 0.021942138671875, 0.0328369140625, 0.043731689453125, 0.05462646484375, 0.065521240234375, 0.076416015625, 0.087310791015625, 0.09820556640625, 0.109100341796875, 0.1199951171875, 0.130889892578125, 0.14178466796875, 0.152679443359375, 0.16357421875, 0.174468994140625, 0.18536376953125, 0.196258544921875, 0.2071533203125, 0.218048095703125, 0.22894287109375, 0.239837646484375, 0.250732421875, 0.261627197265625, 0.27252197265625, 0.283416748046875, 0.2943115234375, 0.305206298828125, 0.31610107421875, 0.326995849609375, 0.337890625]}, "parameters/model.layers.17.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 2.0, 2.0, 4.0, 1.0, 12.0, 7.0, 24.0, 45.0, 60.0, 112.0, 185.0, 295.0, 609.0, 881.0, 1696.0, 3088.0, 5152.0, 7925.0, 13923.0, 22391.0, 31797.0, 51224.0, 74203.0, 108355.0, 144390.0, 195717.0, 241649.0, 291897.0, 343359.0, 379250.0, 395560.0, 374223.0, 337609.0, 292541.0, 239744.0, 184490.0, 141216.0, 105515.0, 69789.0, 49263.0, 32545.0, 21242.0, 12494.0, 8270.0, 4930.0, 2657.0, 1738.0, 931.0, 604.0, 292.0, 152.0, 113.0, 52.0, 34.0, 17.0, 10.0, 3.0, 5.0, 2.0, 5.0], "bins": [-0.0703125, -0.06827163696289062, -0.06623077392578125, -0.06418991088867188, -0.0621490478515625, -0.060108184814453125, -0.05806732177734375, -0.056026458740234375, -0.053985595703125, -0.051944732666015625, -0.04990386962890625, -0.047863006591796875, -0.0458221435546875, -0.043781280517578125, -0.04174041748046875, -0.039699554443359375, -0.03765869140625, -0.035617828369140625, -0.03357696533203125, -0.031536102294921875, -0.0294952392578125, -0.027454376220703125, -0.02541351318359375, -0.023372650146484375, -0.021331787109375, -0.019290924072265625, -0.01725006103515625, -0.015209197998046875, -0.0131683349609375, -0.011127471923828125, -0.00908660888671875, -0.007045745849609375, -0.0050048828125, -0.002964019775390625, -0.00092315673828125, 0.001117706298828125, 0.0031585693359375, 0.005199432373046875, 0.00724029541015625, 0.009281158447265625, 0.011322021484375, 0.013362884521484375, 0.01540374755859375, 0.017444610595703125, 0.0194854736328125, 0.021526336669921875, 0.02356719970703125, 0.025608062744140625, 0.02764892578125, 0.029689788818359375, 0.03173065185546875, 0.033771514892578125, 0.0358123779296875, 0.037853240966796875, 0.03989410400390625, 0.041934967041015625, 0.043975830078125, 0.046016693115234375, 0.04805755615234375, 0.050098419189453125, 0.0521392822265625, 0.054180145263671875, 0.05622100830078125, 0.058261871337890625, 0.060302734375]}, "parameters/model.layers.17.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 3.0, 2.0, 2.0, 3.0, 1.0, 2.0, 1.0, 1.0, 3.0, 0.0, 3.0, 3.0, 3.0, 11.0, 13.0, 18.0, 16.0, 38.0, 79.0, 238.0, 2484.0, 30024.0, 457364.0, 3961354.0, 8531421.0, 3401586.0, 365192.0, 24734.0, 2083.0, 270.0, 87.0, 43.0, 35.0, 20.0, 17.0, 10.0, 12.0, 8.0, 4.0, 3.0, 4.0, 3.0, 1.0, 3.0, 2.0, 0.0, 0.0, 3.0, 1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.419921875, -0.406982421875, -0.39404296875, -0.381103515625, -0.3681640625, -0.355224609375, -0.34228515625, -0.329345703125, -0.31640625, -0.303466796875, -0.29052734375, -0.277587890625, -0.2646484375, -0.251708984375, -0.23876953125, -0.225830078125, -0.212890625, -0.199951171875, -0.18701171875, -0.174072265625, -0.1611328125, -0.148193359375, -0.13525390625, -0.122314453125, -0.109375, -0.096435546875, -0.08349609375, -0.070556640625, -0.0576171875, -0.044677734375, -0.03173828125, -0.018798828125, -0.005859375, 0.007080078125, 0.02001953125, 0.032958984375, 0.0458984375, 0.058837890625, 0.07177734375, 0.084716796875, 0.09765625, 0.110595703125, 0.12353515625, 0.136474609375, 0.1494140625, 0.162353515625, 0.17529296875, 0.188232421875, 0.201171875, 0.214111328125, 0.22705078125, 0.239990234375, 0.2529296875, 0.265869140625, 0.27880859375, 0.291748046875, 0.3046875, 0.317626953125, 0.33056640625, 0.343505859375, 0.3564453125, 0.369384765625, 0.38232421875, 0.395263671875, 0.408203125]}, "parameters/model.layers.17.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 2.0, 4.0, 2.0, 2.0, 4.0, 7.0, 7.0, 12.0, 22.0, 42.0, 32.0, 40.0, 75.0, 100.0, 136.0, 197.0, 385.0, 611.0, 1317.0, 2630.0, 6795.0, 20917.0, 81818.0, 377012.0, 1842203.0, 8312732.0, 16777216.0, 16777216.0, 16777216.0, 10723886.0, 2510529.0, 502967.0, 105985.0, 26116.0, 8336.0, 2978.0, 1520.0, 719.0, 453.0, 226.0, 144.0, 119.0, 82.0, 52.0, 29.0, 38.0, 21.0, 16.0, 23.0, 6.0, 5.0, 3.0, 3.0, 3.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.287109375, -0.277069091796875, -0.26702880859375, -0.256988525390625, -0.2469482421875, -0.236907958984375, -0.22686767578125, -0.216827392578125, -0.206787109375, -0.196746826171875, -0.18670654296875, -0.176666259765625, -0.1666259765625, -0.156585693359375, -0.14654541015625, -0.136505126953125, -0.12646484375, -0.116424560546875, -0.10638427734375, -0.096343994140625, -0.0863037109375, -0.076263427734375, -0.06622314453125, -0.056182861328125, -0.046142578125, -0.036102294921875, -0.02606201171875, -0.016021728515625, -0.0059814453125, 0.004058837890625, 0.01409912109375, 0.024139404296875, 0.0341796875, 0.044219970703125, 0.05426025390625, 0.064300537109375, 0.0743408203125, 0.084381103515625, 0.09442138671875, 0.104461669921875, 0.114501953125, 0.124542236328125, 0.13458251953125, 0.144622802734375, 0.1546630859375, 0.164703369140625, 0.17474365234375, 0.184783935546875, 0.19482421875, 0.204864501953125, 0.21490478515625, 0.224945068359375, 0.2349853515625, 0.245025634765625, 0.25506591796875, 0.265106201171875, 0.275146484375, 0.285186767578125, 0.29522705078125, 0.305267333984375, 0.3153076171875, 0.325347900390625, 0.33538818359375, 0.345428466796875, 0.35546875]}, "parameters/model.layers.17.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 2.0, 5.0, 2.0, 8.0, 11.0, 14.0, 26.0, 33.0, 48.0, 113.0, 145.0, 372.0, 1092.0, 41325.0, 2680984.0, 16777216.0, 16777216.0, 3519732.0, 62346.0, 1284.0, 391.0, 185.0, 112.0, 55.0, 31.0, 24.0, 21.0, 11.0, 7.0, 3.0, 3.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.55859375, -0.54168701171875, -0.5247802734375, -0.50787353515625, -0.490966796875, -0.47406005859375, -0.4571533203125, -0.44024658203125, -0.42333984375, -0.40643310546875, -0.3895263671875, -0.37261962890625, -0.355712890625, -0.33880615234375, -0.3218994140625, -0.30499267578125, -0.2880859375, -0.27117919921875, -0.2542724609375, -0.23736572265625, -0.220458984375, -0.20355224609375, -0.1866455078125, -0.16973876953125, -0.15283203125, -0.13592529296875, -0.1190185546875, -0.10211181640625, -0.085205078125, -0.06829833984375, -0.0513916015625, -0.03448486328125, -0.017578125, -0.00067138671875, 0.0162353515625, 0.03314208984375, 0.050048828125, 0.06695556640625, 0.0838623046875, 0.10076904296875, 0.11767578125, 0.13458251953125, 0.1514892578125, 0.16839599609375, 0.185302734375, 0.20220947265625, 0.2191162109375, 0.23602294921875, 0.2529296875, 0.26983642578125, 0.2867431640625, 0.30364990234375, 0.320556640625, 0.33746337890625, 0.3543701171875, 0.37127685546875, 0.38818359375, 0.40509033203125, 0.4219970703125, 0.43890380859375, 0.455810546875, 0.47271728515625, 0.4896240234375, 0.50653076171875, 0.5234375]}, "parameters/model.layers.17.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 2.0, 4.0, 24.0, 76.0, 468.0, 1225.0, 1252.0, 627.0, 184.0, 80.0, 40.0, 13.0, 13.0, 8.0, 8.0, 3.0, 5.0, 2.0, 2.0, 8.0, 2.0, 3.0, 2.0, 1.0, 2.0, 1.0, 1.0, 3.0, 4.0, 3.0, 5.0, 2.0, 3.0, 2.0, 5.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.08740234375, 0.10123443603515625, 0.1150665283203125, 0.12889862060546875, 0.142730712890625, 0.15656280517578125, 0.1703948974609375, 0.18422698974609375, 0.19805908203125, 0.21189117431640625, 0.2257232666015625, 0.23955535888671875, 0.253387451171875, 0.26721954345703125, 0.2810516357421875, 0.29488372802734375, 0.3087158203125, 0.32254791259765625, 0.3363800048828125, 0.35021209716796875, 0.364044189453125, 0.37787628173828125, 0.3917083740234375, 0.40554046630859375, 0.41937255859375, 0.43320465087890625, 0.4470367431640625, 0.46086883544921875, 0.474700927734375, 0.48853302001953125, 0.5023651123046875, 0.5161972045898438, 0.530029296875, 0.5438613891601562, 0.5576934814453125, 0.5715255737304688, 0.585357666015625, 0.5991897583007812, 0.6130218505859375, 0.6268539428710938, 0.64068603515625, 0.6545181274414062, 0.6683502197265625, 0.6821823120117188, 0.696014404296875, 0.7098464965820312, 0.7236785888671875, 0.7375106811523438, 0.7513427734375, 0.7651748657226562, 0.7790069580078125, 0.7928390502929688, 0.806671142578125, 0.8205032348632812, 0.8343353271484375, 0.8481674194335938, 0.86199951171875, 0.8758316040039062, 0.8896636962890625, 0.9034957885742188, 0.917327880859375, 0.9311599731445312, 0.9449920654296875, 0.9588241577148438, 0.97265625]}, "parameters/model.layers.17.post_attention_layernorm.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 3.0, 5.0, 2.0, 1.0, 5.0, 3.0, 1.0, 2.0, 5.0, 3.0, 0.0, 1.0, 5.0, 4.0, 0.0, 1.0, 4.0, 1.0, 2.0, 6.0, 5.0, 5.0, 6.0, 8.0, 8.0, 20.0, 26.0, 40.0, 29.0, 36.0, 61.0, 43.0, 100.0, 103.0, 306.0, 379.0, 610.0, 1247.0, 648.0, 292.0, 50.0, 5.0, 2.0, 1.0, 2.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0], "bins": [0.07666015625, 0.08333587646484375, 0.0900115966796875, 0.09668731689453125, 0.103363037109375, 0.11003875732421875, 0.1167144775390625, 0.12339019775390625, 0.13006591796875, 0.13674163818359375, 0.1434173583984375, 0.15009307861328125, 0.156768798828125, 0.16344451904296875, 0.1701202392578125, 0.17679595947265625, 0.1834716796875, 0.19014739990234375, 0.1968231201171875, 0.20349884033203125, 0.210174560546875, 0.21685028076171875, 0.2235260009765625, 0.23020172119140625, 0.23687744140625, 0.24355316162109375, 0.2502288818359375, 0.25690460205078125, 0.263580322265625, 0.27025604248046875, 0.2769317626953125, 0.28360748291015625, 0.290283203125, 0.29695892333984375, 0.3036346435546875, 0.31031036376953125, 0.316986083984375, 0.32366180419921875, 0.3303375244140625, 0.33701324462890625, 0.34368896484375, 0.35036468505859375, 0.3570404052734375, 0.36371612548828125, 0.370391845703125, 0.37706756591796875, 0.3837432861328125, 0.39041900634765625, 0.3970947265625, 0.40377044677734375, 0.4104461669921875, 0.41712188720703125, 0.423797607421875, 0.43047332763671875, 0.4371490478515625, 0.44382476806640625, 0.45050048828125, 0.45717620849609375, 0.4638519287109375, 0.47052764892578125, 0.477203369140625, 0.48387908935546875, 0.4905548095703125, 0.49723052978515625, 0.50390625]}, "parameters/model.layers.18.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 1.0, 1.0, 2.0, 1.0, 4.0, 8.0, 9.0, 12.0, 15.0, 7.0, 28.0, 44.0, 46.0, 80.0, 146.0, 227.0, 408.0, 863.0, 1820.0, 3324.0, 7312.0, 15283.0, 30689.0, 64725.0, 118263.0, 218421.0, 409856.0, 729735.0, 1137976.0, 1809369.0, 2412372.0, 2728061.0, 2435565.0, 1833483.0, 1182592.0, 742626.0, 409046.0, 235422.0, 121206.0, 66314.0, 31554.0, 15111.0, 7684.0, 3837.0, 1688.0, 953.0, 457.0, 225.0, 123.0, 69.0, 45.0, 31.0, 24.0, 12.0, 12.0, 5.0, 6.0, 7.0, 5.0, 1.0, 3.0], "bins": [-0.2138671875, -0.207672119140625, -0.20147705078125, -0.195281982421875, -0.1890869140625, -0.182891845703125, -0.17669677734375, -0.170501708984375, -0.164306640625, -0.158111572265625, -0.15191650390625, -0.145721435546875, -0.1395263671875, -0.133331298828125, -0.12713623046875, -0.120941162109375, -0.11474609375, -0.108551025390625, -0.10235595703125, -0.096160888671875, -0.0899658203125, -0.083770751953125, -0.07757568359375, -0.071380615234375, -0.065185546875, -0.058990478515625, -0.05279541015625, -0.046600341796875, -0.0404052734375, -0.034210205078125, -0.02801513671875, -0.021820068359375, -0.015625, -0.009429931640625, -0.00323486328125, 0.002960205078125, 0.0091552734375, 0.015350341796875, 0.02154541015625, 0.027740478515625, 0.033935546875, 0.040130615234375, 0.04632568359375, 0.052520751953125, 0.0587158203125, 0.064910888671875, 0.07110595703125, 0.077301025390625, 0.08349609375, 0.089691162109375, 0.09588623046875, 0.102081298828125, 0.1082763671875, 0.114471435546875, 0.12066650390625, 0.126861572265625, 0.133056640625, 0.139251708984375, 0.14544677734375, 0.151641845703125, 0.1578369140625, 0.164031982421875, 0.17022705078125, 0.176422119140625, 0.1826171875]}, "parameters/model.layers.18.self_attn.k_proj.weight": {"_type": "histogram", "values": [3.0, 5.0, 2.0, 3.0, 4.0, 7.0, 8.0, 9.0, 16.0, 13.0, 30.0, 41.0, 46.0, 48.0, 96.0, 151.0, 218.0, 358.0, 620.0, 1132.0, 2108.0, 4209.0, 7771.0, 15114.0, 26949.0, 52461.0, 90677.0, 146782.0, 235155.0, 354935.0, 491248.0, 598668.0, 613340.0, 520997.0, 380156.0, 260593.0, 162103.0, 100864.0, 58765.0, 32162.0, 17232.0, 9092.0, 4782.0, 2304.0, 1288.0, 617.0, 390.0, 233.0, 154.0, 108.0, 76.0, 41.0, 30.0, 26.0, 26.0, 14.0, 7.0, 5.0, 5.0, 2.0, 2.0, 1.0, 1.0, 1.0], "bins": [-0.283203125, -0.274383544921875, -0.26556396484375, -0.256744384765625, -0.2479248046875, -0.239105224609375, -0.23028564453125, -0.221466064453125, -0.212646484375, -0.203826904296875, -0.19500732421875, -0.186187744140625, -0.1773681640625, -0.168548583984375, -0.15972900390625, -0.150909423828125, -0.14208984375, -0.133270263671875, -0.12445068359375, -0.115631103515625, -0.1068115234375, -0.097991943359375, -0.08917236328125, -0.080352783203125, -0.071533203125, -0.062713623046875, -0.05389404296875, -0.045074462890625, -0.0362548828125, -0.027435302734375, -0.01861572265625, -0.009796142578125, -0.0009765625, 0.007843017578125, 0.01666259765625, 0.025482177734375, 0.0343017578125, 0.043121337890625, 0.05194091796875, 0.060760498046875, 0.069580078125, 0.078399658203125, 0.08721923828125, 0.096038818359375, 0.1048583984375, 0.113677978515625, 0.12249755859375, 0.131317138671875, 0.14013671875, 0.148956298828125, 0.15777587890625, 0.166595458984375, 0.1754150390625, 0.184234619140625, 0.19305419921875, 0.201873779296875, 0.210693359375, 0.219512939453125, 0.22833251953125, 0.237152099609375, 0.2459716796875, 0.254791259765625, 0.26361083984375, 0.272430419921875, 0.28125]}, "parameters/model.layers.18.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 6.0, 5.0, 5.0, 14.0, 17.0, 18.0, 30.0, 37.0, 48.0, 83.0, 114.0, 211.0, 286.0, 517.0, 785.0, 1532.0, 2570.0, 4939.0, 8767.0, 15684.0, 27324.0, 46398.0, 78637.0, 125267.0, 187431.0, 266747.0, 353786.0, 433175.0, 489286.0, 492878.0, 445044.0, 372075.0, 283330.0, 204019.0, 139174.0, 87268.0, 53646.0, 31817.0, 18144.0, 10119.0, 5682.0, 3079.0, 1724.0, 1027.0, 491.0, 402.0, 222.0, 138.0, 96.0, 63.0, 51.0, 28.0, 24.0, 12.0, 15.0, 5.0, 5.0, 1.0, 0.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.06982421875, -0.0675048828125, -0.065185546875, -0.0628662109375, -0.060546875, -0.0582275390625, -0.055908203125, -0.0535888671875, -0.05126953125, -0.0489501953125, -0.046630859375, -0.0443115234375, -0.0419921875, -0.0396728515625, -0.037353515625, -0.0350341796875, -0.03271484375, -0.0303955078125, -0.028076171875, -0.0257568359375, -0.0234375, -0.0211181640625, -0.018798828125, -0.0164794921875, -0.01416015625, -0.0118408203125, -0.009521484375, -0.0072021484375, -0.0048828125, -0.0025634765625, -0.000244140625, 0.0020751953125, 0.00439453125, 0.0067138671875, 0.009033203125, 0.0113525390625, 0.013671875, 0.0159912109375, 0.018310546875, 0.0206298828125, 0.02294921875, 0.0252685546875, 0.027587890625, 0.0299072265625, 0.0322265625, 0.0345458984375, 0.036865234375, 0.0391845703125, 0.04150390625, 0.0438232421875, 0.046142578125, 0.0484619140625, 0.05078125, 0.0531005859375, 0.055419921875, 0.0577392578125, 0.06005859375, 0.0623779296875, 0.064697265625, 0.0670166015625, 0.0693359375, 0.0716552734375, 0.073974609375, 0.0762939453125, 0.07861328125]}, "parameters/model.layers.18.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 5.0, 10.0, 18.0, 38.0, 121.0, 315.0, 1634.0, 34518.0, 1755806.0, 10536580.0, 4295022.0, 148712.0, 3647.0, 488.0, 171.0, 64.0, 35.0, 10.0, 5.0, 1.0, 3.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.29296875, -0.275390625, -0.2578125, -0.240234375, -0.22265625, -0.205078125, -0.1875, -0.169921875, -0.15234375, -0.134765625, -0.1171875, -0.099609375, -0.08203125, -0.064453125, -0.046875, -0.029296875, -0.01171875, 0.005859375, 0.0234375, 0.041015625, 0.05859375, 0.076171875, 0.09375, 0.111328125, 0.12890625, 0.146484375, 0.1640625, 0.181640625, 0.19921875, 0.216796875, 0.234375, 0.251953125, 0.26953125, 0.287109375, 0.3046875, 0.322265625, 0.33984375, 0.357421875, 0.375, 0.392578125, 0.41015625, 0.427734375, 0.4453125, 0.462890625, 0.48046875, 0.498046875, 0.515625, 0.533203125, 0.55078125, 0.568359375, 0.5859375, 0.603515625, 0.62109375, 0.638671875, 0.65625, 0.673828125, 0.69140625, 0.708984375, 0.7265625, 0.744140625, 0.76171875, 0.779296875, 0.796875, 0.814453125, 0.83203125]}, "parameters/model.layers.18.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 2.0, 2.0, 3.0, 8.0, 5.0, 17.0, 11.0, 22.0, 35.0, 75.0, 86.0, 168.0, 373.0, 759.0, 1852.0, 6449.0, 31582.0, 206456.0, 1717980.0, 12179028.0, 16777216.0, 16777216.0, 15521156.0, 2373601.0, 278782.0, 40970.0, 7736.0, 2135.0, 928.0, 419.0, 197.0, 117.0, 60.0, 36.0, 26.0, 15.0, 13.0, 7.0, 5.0, 2.0, 4.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.45703125, -0.444000244140625, -0.43096923828125, -0.417938232421875, -0.4049072265625, -0.391876220703125, -0.37884521484375, -0.365814208984375, -0.352783203125, -0.339752197265625, -0.32672119140625, -0.313690185546875, -0.3006591796875, -0.287628173828125, -0.27459716796875, -0.261566162109375, -0.24853515625, -0.235504150390625, -0.22247314453125, -0.209442138671875, -0.1964111328125, -0.183380126953125, -0.17034912109375, -0.157318115234375, -0.144287109375, -0.131256103515625, -0.11822509765625, -0.105194091796875, -0.0921630859375, -0.079132080078125, -0.06610107421875, -0.053070068359375, -0.0400390625, -0.027008056640625, -0.01397705078125, -0.000946044921875, 0.0120849609375, 0.025115966796875, 0.03814697265625, 0.051177978515625, 0.064208984375, 0.077239990234375, 0.09027099609375, 0.103302001953125, 0.1163330078125, 0.129364013671875, 0.14239501953125, 0.155426025390625, 0.16845703125, 0.181488037109375, 0.19451904296875, 0.207550048828125, 0.2205810546875, 0.233612060546875, 0.24664306640625, 0.259674072265625, 0.272705078125, 0.285736083984375, 0.29876708984375, 0.311798095703125, 0.3248291015625, 0.337860107421875, 0.35089111328125, 0.363922119140625, 0.376953125]}, "parameters/model.layers.18.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 2.0, 2.0, 1.0, 3.0, 6.0, 9.0, 13.0, 27.0, 44.0, 113.0, 253.0, 1057.0, 131436.0, 6340071.0, 16777216.0, 16777216.0, 1038861.0, 7750.0, 402.0, 148.0, 56.0, 33.0, 21.0, 13.0, 6.0, 3.0, 4.0, 1.0, 0.0, 2.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.7109375, -0.6934814453125, -0.676025390625, -0.6585693359375, -0.64111328125, -0.6236572265625, -0.606201171875, -0.5887451171875, -0.5712890625, -0.5538330078125, -0.536376953125, -0.5189208984375, -0.50146484375, -0.4840087890625, -0.466552734375, -0.4490966796875, -0.431640625, -0.4141845703125, -0.396728515625, -0.3792724609375, -0.36181640625, -0.3443603515625, -0.326904296875, -0.3094482421875, -0.2919921875, -0.2745361328125, -0.257080078125, -0.2396240234375, -0.22216796875, -0.2047119140625, -0.187255859375, -0.1697998046875, -0.15234375, -0.1348876953125, -0.117431640625, -0.0999755859375, -0.08251953125, -0.0650634765625, -0.047607421875, -0.0301513671875, -0.0126953125, 0.0047607421875, 0.022216796875, 0.0396728515625, 0.05712890625, 0.0745849609375, 0.092041015625, 0.1094970703125, 0.126953125, 0.1444091796875, 0.161865234375, 0.1793212890625, 0.19677734375, 0.2142333984375, 0.231689453125, 0.2491455078125, 0.2666015625, 0.2840576171875, 0.301513671875, 0.3189697265625, 0.33642578125, 0.3538818359375, 0.371337890625, 0.3887939453125, 0.40625]}, "parameters/model.layers.18.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 0.0, 1.0, 7.0, 11.0, 45.0, 337.0, 1359.0, 1266.0, 651.0, 210.0, 57.0, 35.0, 16.0, 16.0, 7.0, 9.0, 4.0, 5.0, 1.0, 5.0, 4.0, 1.0, 4.0, 2.0, 2.0, 1.0, 3.0, 3.0, 4.0, 5.0, 2.0, 2.0, 5.0, 1.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.06298828125, 0.07775115966796875, 0.0925140380859375, 0.10727691650390625, 0.122039794921875, 0.13680267333984375, 0.1515655517578125, 0.16632843017578125, 0.18109130859375, 0.19585418701171875, 0.2106170654296875, 0.22537994384765625, 0.240142822265625, 0.25490570068359375, 0.2696685791015625, 0.28443145751953125, 0.2991943359375, 0.31395721435546875, 0.3287200927734375, 0.34348297119140625, 0.358245849609375, 0.37300872802734375, 0.3877716064453125, 0.40253448486328125, 0.41729736328125, 0.43206024169921875, 0.4468231201171875, 0.46158599853515625, 0.476348876953125, 0.49111175537109375, 0.5058746337890625, 0.5206375122070312, 0.535400390625, 0.5501632690429688, 0.5649261474609375, 0.5796890258789062, 0.594451904296875, 0.6092147827148438, 0.6239776611328125, 0.6387405395507812, 0.65350341796875, 0.6682662963867188, 0.6830291748046875, 0.6977920532226562, 0.712554931640625, 0.7273178100585938, 0.7420806884765625, 0.7568435668945312, 0.7716064453125, 0.7863693237304688, 0.8011322021484375, 0.8158950805664062, 0.830657958984375, 0.8454208374023438, 0.8601837158203125, 0.8749465942382812, 0.88970947265625, 0.9044723510742188, 0.9192352294921875, 0.9339981079101562, 0.948760986328125, 0.9635238647460938, 0.9782867431640625, 0.9930496215820312, 1.0078125]}, "parameters/model.layers.18.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 2.0, 1.0, 3.0, 2.0, 3.0, 1.0, 3.0, 1.0, 3.0, 0.0, 5.0, 2.0, 2.0, 3.0, 0.0, 6.0, 2.0, 1.0, 0.0, 2.0, 2.0, 1.0, 2.0, 5.0, 2.0, 5.0, 4.0, 5.0, 8.0, 7.0, 10.0, 11.0, 25.0, 18.0, 13.0, 23.0, 22.0, 15.0, 19.0, 28.0, 36.0, 38.0, 46.0, 86.0, 92.0, 163.0, 281.0, 459.0, 700.0, 1168.0, 535.0, 182.0, 29.0, 5.0, 1.0, 0.0, 1.0, 3.0], "bins": [0.0751953125, 0.0812835693359375, 0.087371826171875, 0.0934600830078125, 0.09954833984375, 0.1056365966796875, 0.111724853515625, 0.1178131103515625, 0.1239013671875, 0.1299896240234375, 0.136077880859375, 0.1421661376953125, 0.14825439453125, 0.1543426513671875, 0.160430908203125, 0.1665191650390625, 0.172607421875, 0.1786956787109375, 0.184783935546875, 0.1908721923828125, 0.19696044921875, 0.2030487060546875, 0.209136962890625, 0.2152252197265625, 0.2213134765625, 0.2274017333984375, 0.233489990234375, 0.2395782470703125, 0.24566650390625, 0.2517547607421875, 0.257843017578125, 0.2639312744140625, 0.27001953125, 0.2761077880859375, 0.282196044921875, 0.2882843017578125, 0.29437255859375, 0.3004608154296875, 0.306549072265625, 0.3126373291015625, 0.3187255859375, 0.3248138427734375, 0.330902099609375, 0.3369903564453125, 0.34307861328125, 0.3491668701171875, 0.355255126953125, 0.3613433837890625, 0.367431640625, 0.3735198974609375, 0.379608154296875, 0.3856964111328125, 0.39178466796875, 0.3978729248046875, 0.403961181640625, 0.4100494384765625, 0.4161376953125, 0.4222259521484375, 0.428314208984375, 0.4344024658203125, 0.44049072265625, 0.4465789794921875, 0.452667236328125, 0.4587554931640625, 0.46484375]}, "parameters/model.layers.19.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 3.0, 2.0, 3.0, 6.0, 7.0, 11.0, 13.0, 18.0, 27.0, 53.0, 70.0, 138.0, 292.0, 645.0, 1120.0, 2227.0, 5120.0, 9559.0, 22799.0, 50397.0, 106513.0, 218579.0, 421170.0, 831041.0, 1419089.0, 2255722.0, 2993079.0, 2994986.0, 2277345.0, 1458342.0, 837551.0, 442012.0, 223937.0, 110042.0, 52118.0, 22843.0, 10490.0, 4904.0, 2354.0, 1289.0, 606.0, 325.0, 158.0, 84.0, 49.0, 35.0, 9.0, 12.0, 6.0, 1.0, 3.0, 4.0, 1.0, 2.0, 2.0], "bins": [-0.25390625, -0.246856689453125, -0.23980712890625, -0.232757568359375, -0.2257080078125, -0.218658447265625, -0.21160888671875, -0.204559326171875, -0.197509765625, -0.190460205078125, -0.18341064453125, -0.176361083984375, -0.1693115234375, -0.162261962890625, -0.15521240234375, -0.148162841796875, -0.14111328125, -0.134063720703125, -0.12701416015625, -0.119964599609375, -0.1129150390625, -0.105865478515625, -0.09881591796875, -0.091766357421875, -0.084716796875, -0.077667236328125, -0.07061767578125, -0.063568115234375, -0.0565185546875, -0.049468994140625, -0.04241943359375, -0.035369873046875, -0.0283203125, -0.021270751953125, -0.01422119140625, -0.007171630859375, -0.0001220703125, 0.006927490234375, 0.01397705078125, 0.021026611328125, 0.028076171875, 0.035125732421875, 0.04217529296875, 0.049224853515625, 0.0562744140625, 0.063323974609375, 0.07037353515625, 0.077423095703125, 0.08447265625, 0.091522216796875, 0.09857177734375, 0.105621337890625, 0.1126708984375, 0.119720458984375, 0.12677001953125, 0.133819580078125, 0.140869140625, 0.147918701171875, 0.15496826171875, 0.162017822265625, 0.1690673828125, 0.176116943359375, 0.18316650390625, 0.190216064453125, 0.197265625]}, "parameters/model.layers.19.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 4.0, 0.0, 4.0, 6.0, 4.0, 14.0, 14.0, 17.0, 23.0, 26.0, 56.0, 66.0, 108.0, 200.0, 341.0, 600.0, 1235.0, 2711.0, 5183.0, 10749.0, 22264.0, 44304.0, 86333.0, 147677.0, 260377.0, 406072.0, 574945.0, 674936.0, 642002.0, 506577.0, 348897.0, 205494.0, 120842.0, 66149.0, 32962.0, 16601.0, 8412.0, 4038.0, 1975.0, 920.0, 451.0, 253.0, 154.0, 105.0, 68.0, 33.0, 36.0, 18.0, 9.0, 12.0, 8.0, 6.0, 1.0, 2.0, 0.0, 3.0, 0.0, 0.0, 1.0, 2.0], "bins": [-0.30078125, -0.29132080078125, -0.2818603515625, -0.27239990234375, -0.262939453125, -0.25347900390625, -0.2440185546875, -0.23455810546875, -0.22509765625, -0.21563720703125, -0.2061767578125, -0.19671630859375, -0.187255859375, -0.17779541015625, -0.1683349609375, -0.15887451171875, -0.1494140625, -0.13995361328125, -0.1304931640625, -0.12103271484375, -0.111572265625, -0.10211181640625, -0.0926513671875, -0.08319091796875, -0.07373046875, -0.06427001953125, -0.0548095703125, -0.04534912109375, -0.035888671875, -0.02642822265625, -0.0169677734375, -0.00750732421875, 0.001953125, 0.01141357421875, 0.0208740234375, 0.03033447265625, 0.039794921875, 0.04925537109375, 0.0587158203125, 0.06817626953125, 0.07763671875, 0.08709716796875, 0.0965576171875, 0.10601806640625, 0.115478515625, 0.12493896484375, 0.1343994140625, 0.14385986328125, 0.1533203125, 0.16278076171875, 0.1722412109375, 0.18170166015625, 0.191162109375, 0.20062255859375, 0.2100830078125, 0.21954345703125, 0.22900390625, 0.23846435546875, 0.2479248046875, 0.25738525390625, 0.266845703125, 0.27630615234375, 0.2857666015625, 0.29522705078125, 0.3046875]}, "parameters/model.layers.19.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 4.0, 5.0, 10.0, 14.0, 27.0, 42.0, 86.0, 176.0, 359.0, 714.0, 1471.0, 2942.0, 5667.0, 11038.0, 20364.0, 35795.0, 62791.0, 108204.0, 173930.0, 259888.0, 365926.0, 462856.0, 533811.0, 539354.0, 477176.0, 383741.0, 277013.0, 190093.0, 118352.0, 71372.0, 41597.0, 23438.0, 12597.0, 6745.0, 3271.0, 1725.0, 835.0, 443.0, 208.0, 96.0, 55.0, 30.0, 16.0, 12.0, 3.0, 7.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.078125, -0.075439453125, -0.07275390625, -0.070068359375, -0.0673828125, -0.064697265625, -0.06201171875, -0.059326171875, -0.056640625, -0.053955078125, -0.05126953125, -0.048583984375, -0.0458984375, -0.043212890625, -0.04052734375, -0.037841796875, -0.03515625, -0.032470703125, -0.02978515625, -0.027099609375, -0.0244140625, -0.021728515625, -0.01904296875, -0.016357421875, -0.013671875, -0.010986328125, -0.00830078125, -0.005615234375, -0.0029296875, -0.000244140625, 0.00244140625, 0.005126953125, 0.0078125, 0.010498046875, 0.01318359375, 0.015869140625, 0.0185546875, 0.021240234375, 0.02392578125, 0.026611328125, 0.029296875, 0.031982421875, 0.03466796875, 0.037353515625, 0.0400390625, 0.042724609375, 0.04541015625, 0.048095703125, 0.05078125, 0.053466796875, 0.05615234375, 0.058837890625, 0.0615234375, 0.064208984375, 0.06689453125, 0.069580078125, 0.072265625, 0.074951171875, 0.07763671875, 0.080322265625, 0.0830078125, 0.085693359375, 0.08837890625, 0.091064453125, 0.09375]}, "parameters/model.layers.19.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 1.0, 8.0, 1.0, 10.0, 17.0, 32.0, 66.0, 141.0, 406.0, 3753.0, 164218.0, 2893028.0, 9327566.0, 4079866.0, 300309.0, 6987.0, 492.0, 162.0, 68.0, 36.0, 11.0, 13.0, 1.0, 4.0, 0.0, 0.0, 4.0, 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.56640625, -0.551727294921875, -0.53704833984375, -0.522369384765625, -0.5076904296875, -0.493011474609375, -0.47833251953125, -0.463653564453125, -0.448974609375, -0.434295654296875, -0.41961669921875, -0.404937744140625, -0.3902587890625, -0.375579833984375, -0.36090087890625, -0.346221923828125, -0.33154296875, -0.316864013671875, -0.30218505859375, -0.287506103515625, -0.2728271484375, -0.258148193359375, -0.24346923828125, -0.228790283203125, -0.214111328125, -0.199432373046875, -0.18475341796875, -0.170074462890625, -0.1553955078125, -0.140716552734375, -0.12603759765625, -0.111358642578125, -0.0966796875, -0.082000732421875, -0.06732177734375, -0.052642822265625, -0.0379638671875, -0.023284912109375, -0.00860595703125, 0.006072998046875, 0.020751953125, 0.035430908203125, 0.05010986328125, 0.064788818359375, 0.0794677734375, 0.094146728515625, 0.10882568359375, 0.123504638671875, 0.13818359375, 0.152862548828125, 0.16754150390625, 0.182220458984375, 0.1968994140625, 0.211578369140625, 0.22625732421875, 0.240936279296875, 0.255615234375, 0.270294189453125, 0.28497314453125, 0.299652099609375, 0.3143310546875, 0.329010009765625, 0.34368896484375, 0.358367919921875, 0.373046875]}, "parameters/model.layers.19.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 1.0, 1.0, 3.0, 2.0, 3.0, 1.0, 4.0, 6.0, 6.0, 14.0, 17.0, 37.0, 42.0, 70.0, 92.0, 195.0, 383.0, 780.0, 2038.0, 6610.0, 26832.0, 162155.0, 1108962.0, 7479580.0, 16777216.0, 16777216.0, 16777216.0, 6376924.0, 908679.0, 136620.0, 22944.0, 5826.0, 1798.0, 755.0, 353.0, 178.0, 83.0, 69.0, 41.0, 29.0, 21.0, 12.0, 14.0, 6.0, 5.0, 9.0, 2.0, 5.0, 0.0, 1.0, 3.0, 0.0, 1.0], "bins": [-0.439453125, -0.427398681640625, -0.41534423828125, -0.403289794921875, -0.3912353515625, -0.379180908203125, -0.36712646484375, -0.355072021484375, -0.343017578125, -0.330963134765625, -0.31890869140625, -0.306854248046875, -0.2947998046875, -0.282745361328125, -0.27069091796875, -0.258636474609375, -0.24658203125, -0.234527587890625, -0.22247314453125, -0.210418701171875, -0.1983642578125, -0.186309814453125, -0.17425537109375, -0.162200927734375, -0.150146484375, -0.138092041015625, -0.12603759765625, -0.113983154296875, -0.1019287109375, -0.089874267578125, -0.07781982421875, -0.065765380859375, -0.0537109375, -0.041656494140625, -0.02960205078125, -0.017547607421875, -0.0054931640625, 0.006561279296875, 0.01861572265625, 0.030670166015625, 0.042724609375, 0.054779052734375, 0.06683349609375, 0.078887939453125, 0.0909423828125, 0.102996826171875, 0.11505126953125, 0.127105712890625, 0.13916015625, 0.151214599609375, 0.16326904296875, 0.175323486328125, 0.1873779296875, 0.199432373046875, 0.21148681640625, 0.223541259765625, 0.235595703125, 0.247650146484375, 0.25970458984375, 0.271759033203125, 0.2838134765625, 0.295867919921875, 0.30792236328125, 0.319976806640625, 0.33203125]}, "parameters/model.layers.19.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 5.0, 4.0, 5.0, 6.0, 18.0, 26.0, 66.0, 161.0, 1503.0, 591336.0, 16777216.0, 16777216.0, 2735563.0, 10723.0, 306.0, 96.0, 45.0, 21.0, 11.0, 1.0, 4.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.5703125, -0.54931640625, -0.5283203125, -0.50732421875, -0.486328125, -0.46533203125, -0.4443359375, -0.42333984375, -0.40234375, -0.38134765625, -0.3603515625, -0.33935546875, -0.318359375, -0.29736328125, -0.2763671875, -0.25537109375, -0.234375, -0.21337890625, -0.1923828125, -0.17138671875, -0.150390625, -0.12939453125, -0.1083984375, -0.08740234375, -0.06640625, -0.04541015625, -0.0244140625, -0.00341796875, 0.017578125, 0.03857421875, 0.0595703125, 0.08056640625, 0.1015625, 0.12255859375, 0.1435546875, 0.16455078125, 0.185546875, 0.20654296875, 0.2275390625, 0.24853515625, 0.26953125, 0.29052734375, 0.3115234375, 0.33251953125, 0.353515625, 0.37451171875, 0.3955078125, 0.41650390625, 0.4375, 0.45849609375, 0.4794921875, 0.50048828125, 0.521484375, 0.54248046875, 0.5634765625, 0.58447265625, 0.60546875, 0.62646484375, 0.6474609375, 0.66845703125, 0.689453125, 0.71044921875, 0.7314453125, 0.75244140625, 0.7734375]}, "parameters/model.layers.19.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 4.0, 3.0, 12.0, 7.0, 30.0, 372.0, 1432.0, 1424.0, 397.0, 157.0, 67.0, 33.0, 29.0, 15.0, 14.0, 6.0, 19.0, 6.0, 8.0, 4.0, 3.0, 3.0, 5.0, 5.0, 3.0, 7.0, 2.0, 3.0, 6.0, 3.0, 2.0, 5.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.09130859375, 0.10514068603515625, 0.1189727783203125, 0.13280487060546875, 0.146636962890625, 0.16046905517578125, 0.1743011474609375, 0.18813323974609375, 0.20196533203125, 0.21579742431640625, 0.2296295166015625, 0.24346160888671875, 0.257293701171875, 0.27112579345703125, 0.2849578857421875, 0.29878997802734375, 0.3126220703125, 0.32645416259765625, 0.3402862548828125, 0.35411834716796875, 0.367950439453125, 0.38178253173828125, 0.3956146240234375, 0.40944671630859375, 0.42327880859375, 0.43711090087890625, 0.4509429931640625, 0.46477508544921875, 0.478607177734375, 0.49243927001953125, 0.5062713623046875, 0.5201034545898438, 0.533935546875, 0.5477676391601562, 0.5615997314453125, 0.5754318237304688, 0.589263916015625, 0.6030960083007812, 0.6169281005859375, 0.6307601928710938, 0.64459228515625, 0.6584243774414062, 0.6722564697265625, 0.6860885620117188, 0.699920654296875, 0.7137527465820312, 0.7275848388671875, 0.7414169311523438, 0.7552490234375, 0.7690811157226562, 0.7829132080078125, 0.7967453002929688, 0.810577392578125, 0.8244094848632812, 0.8382415771484375, 0.8520736694335938, 0.86590576171875, 0.8797378540039062, 0.8935699462890625, 0.9074020385742188, 0.921234130859375, 0.9350662231445312, 0.9488983154296875, 0.9627304077148438, 0.9765625]}, "parameters/model.layers.19.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 2.0, 1.0, 0.0, 3.0, 2.0, 1.0, 2.0, 3.0, 2.0, 1.0, 4.0, 1.0, 3.0, 1.0, 3.0, 2.0, 2.0, 3.0, 1.0, 3.0, 0.0, 2.0, 3.0, 1.0, 2.0, 0.0, 2.0, 0.0, 2.0, 1.0, 4.0, 7.0, 3.0, 7.0, 2.0, 5.0, 14.0, 13.0, 12.0, 17.0, 16.0, 15.0, 16.0, 23.0, 13.0, 11.0, 25.0, 36.0, 38.0, 47.0, 40.0, 103.0, 121.0, 214.0, 358.0, 669.0, 868.0, 831.0, 411.0, 78.0, 19.0, 3.0, 3.0], "bins": [0.080078125, 0.0860595703125, 0.092041015625, 0.0980224609375, 0.10400390625, 0.1099853515625, 0.115966796875, 0.1219482421875, 0.1279296875, 0.1339111328125, 0.139892578125, 0.1458740234375, 0.15185546875, 0.1578369140625, 0.163818359375, 0.1697998046875, 0.17578125, 0.1817626953125, 0.187744140625, 0.1937255859375, 0.19970703125, 0.2056884765625, 0.211669921875, 0.2176513671875, 0.2236328125, 0.2296142578125, 0.235595703125, 0.2415771484375, 0.24755859375, 0.2535400390625, 0.259521484375, 0.2655029296875, 0.271484375, 0.2774658203125, 0.283447265625, 0.2894287109375, 0.29541015625, 0.3013916015625, 0.307373046875, 0.3133544921875, 0.3193359375, 0.3253173828125, 0.331298828125, 0.3372802734375, 0.34326171875, 0.3492431640625, 0.355224609375, 0.3612060546875, 0.3671875, 0.3731689453125, 0.379150390625, 0.3851318359375, 0.39111328125, 0.3970947265625, 0.403076171875, 0.4090576171875, 0.4150390625, 0.4210205078125, 0.427001953125, 0.4329833984375, 0.43896484375, 0.4449462890625, 0.450927734375, 0.4569091796875, 0.462890625]}, "parameters/model.layers.20.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 0.0, 0.0, 2.0, 2.0, 2.0, 3.0, 1.0, 8.0, 15.0, 16.0, 16.0, 33.0, 48.0, 71.0, 146.0, 298.0, 596.0, 1317.0, 2729.0, 5802.0, 12291.0, 26682.0, 57482.0, 116830.0, 229301.0, 447600.0, 831994.0, 1419399.0, 2245472.0, 2990594.0, 2981672.0, 2238209.0, 1417386.0, 840108.0, 447962.0, 233613.0, 118991.0, 58785.0, 27615.0, 12662.0, 5916.0, 2955.0, 1320.0, 632.0, 290.0, 168.0, 82.0, 38.0, 23.0, 15.0, 6.0, 5.0, 3.0, 2.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.21875, -0.2119140625, -0.205078125, -0.1982421875, -0.19140625, -0.1845703125, -0.177734375, -0.1708984375, -0.1640625, -0.1572265625, -0.150390625, -0.1435546875, -0.13671875, -0.1298828125, -0.123046875, -0.1162109375, -0.109375, -0.1025390625, -0.095703125, -0.0888671875, -0.08203125, -0.0751953125, -0.068359375, -0.0615234375, -0.0546875, -0.0478515625, -0.041015625, -0.0341796875, -0.02734375, -0.0205078125, -0.013671875, -0.0068359375, 0.0, 0.0068359375, 0.013671875, 0.0205078125, 0.02734375, 0.0341796875, 0.041015625, 0.0478515625, 0.0546875, 0.0615234375, 0.068359375, 0.0751953125, 0.08203125, 0.0888671875, 0.095703125, 0.1025390625, 0.109375, 0.1162109375, 0.123046875, 0.1298828125, 0.13671875, 0.1435546875, 0.150390625, 0.1572265625, 0.1640625, 0.1708984375, 0.177734375, 0.1845703125, 0.19140625, 0.1982421875, 0.205078125, 0.2119140625, 0.21875]}, "parameters/model.layers.20.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 4.0, 1.0, 0.0, 2.0, 6.0, 4.0, 11.0, 9.0, 14.0, 28.0, 33.0, 53.0, 74.0, 94.0, 118.0, 224.0, 315.0, 564.0, 1023.0, 2134.0, 4084.0, 8102.0, 16862.0, 32059.0, 65924.0, 122178.0, 207521.0, 352061.0, 520573.0, 664654.0, 684360.0, 566765.0, 397346.0, 248350.0, 138816.0, 79735.0, 40092.0, 19342.0, 10294.0, 4771.0, 2536.0, 1377.0, 689.0, 361.0, 228.0, 164.0, 94.0, 69.0, 56.0, 41.0, 22.0, 21.0, 14.0, 8.0, 6.0, 5.0, 1.0, 2.0, 2.0, 3.0, 2.0], "bins": [-0.3125, -0.303070068359375, -0.29364013671875, -0.284210205078125, -0.2747802734375, -0.265350341796875, -0.25592041015625, -0.246490478515625, -0.237060546875, -0.227630615234375, -0.21820068359375, -0.208770751953125, -0.1993408203125, -0.189910888671875, -0.18048095703125, -0.171051025390625, -0.16162109375, -0.152191162109375, -0.14276123046875, -0.133331298828125, -0.1239013671875, -0.114471435546875, -0.10504150390625, -0.095611572265625, -0.086181640625, -0.076751708984375, -0.06732177734375, -0.057891845703125, -0.0484619140625, -0.039031982421875, -0.02960205078125, -0.020172119140625, -0.0107421875, -0.001312255859375, 0.00811767578125, 0.017547607421875, 0.0269775390625, 0.036407470703125, 0.04583740234375, 0.055267333984375, 0.064697265625, 0.074127197265625, 0.08355712890625, 0.092987060546875, 0.1024169921875, 0.111846923828125, 0.12127685546875, 0.130706787109375, 0.14013671875, 0.149566650390625, 0.15899658203125, 0.168426513671875, 0.1778564453125, 0.187286376953125, 0.19671630859375, 0.206146240234375, 0.215576171875, 0.225006103515625, 0.23443603515625, 0.243865966796875, 0.2532958984375, 0.262725830078125, 0.27215576171875, 0.281585693359375, 0.291015625]}, "parameters/model.layers.20.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 4.0, 10.0, 3.0, 9.0, 22.0, 27.0, 58.0, 102.0, 184.0, 270.0, 522.0, 970.0, 1530.0, 2922.0, 4963.0, 8908.0, 14769.0, 24489.0, 40138.0, 62732.0, 94753.0, 142495.0, 199297.0, 262547.0, 334617.0, 390302.0, 437583.0, 443546.0, 409387.0, 349577.0, 290225.0, 218987.0, 160673.0, 108590.0, 71824.0, 46373.0, 28920.0, 17718.0, 10622.0, 5877.0, 3417.0, 1843.0, 1044.0, 670.0, 323.0, 197.0, 104.0, 65.0, 39.0, 20.0, 10.0, 7.0, 5.0, 3.0, 2.0, 3.0, 0.0, 1.0], "bins": [-0.07666015625, -0.07434844970703125, -0.0720367431640625, -0.06972503662109375, -0.067413330078125, -0.06510162353515625, -0.0627899169921875, -0.06047821044921875, -0.05816650390625, -0.05585479736328125, -0.0535430908203125, -0.05123138427734375, -0.048919677734375, -0.04660797119140625, -0.0442962646484375, -0.04198455810546875, -0.0396728515625, -0.03736114501953125, -0.0350494384765625, -0.03273773193359375, -0.030426025390625, -0.02811431884765625, -0.0258026123046875, -0.02349090576171875, -0.02117919921875, -0.01886749267578125, -0.0165557861328125, -0.01424407958984375, -0.011932373046875, -0.00962066650390625, -0.0073089599609375, -0.00499725341796875, -0.002685546875, -0.00037384033203125, 0.0019378662109375, 0.00424957275390625, 0.006561279296875, 0.00887298583984375, 0.0111846923828125, 0.01349639892578125, 0.01580810546875, 0.01811981201171875, 0.0204315185546875, 0.02274322509765625, 0.025054931640625, 0.02736663818359375, 0.0296783447265625, 0.03199005126953125, 0.0343017578125, 0.03661346435546875, 0.0389251708984375, 0.04123687744140625, 0.043548583984375, 0.04586029052734375, 0.0481719970703125, 0.05048370361328125, 0.05279541015625, 0.05510711669921875, 0.0574188232421875, 0.05973052978515625, 0.062042236328125, 0.06435394287109375, 0.0666656494140625, 0.06897735595703125, 0.0712890625]}, "parameters/model.layers.20.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 0.0, 1.0, 1.0, 1.0, 2.0, 2.0, 4.0, 4.0, 4.0, 2.0, 8.0, 10.0, 15.0, 27.0, 37.0, 77.0, 248.0, 694.0, 3342.0, 98291.0, 2243715.0, 9163265.0, 4866512.0, 389218.0, 10011.0, 1166.0, 298.0, 120.0, 39.0, 27.0, 23.0, 10.0, 12.0, 6.0, 5.0, 3.0, 1.0, 2.0, 1.0, 0.0, 3.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.5703125, -0.555572509765625, -0.54083251953125, -0.526092529296875, -0.5113525390625, -0.496612548828125, -0.48187255859375, -0.467132568359375, -0.452392578125, -0.437652587890625, -0.42291259765625, -0.408172607421875, -0.3934326171875, -0.378692626953125, -0.36395263671875, -0.349212646484375, -0.33447265625, -0.319732666015625, -0.30499267578125, -0.290252685546875, -0.2755126953125, -0.260772705078125, -0.24603271484375, -0.231292724609375, -0.216552734375, -0.201812744140625, -0.18707275390625, -0.172332763671875, -0.1575927734375, -0.142852783203125, -0.12811279296875, -0.113372802734375, -0.0986328125, -0.083892822265625, -0.06915283203125, -0.054412841796875, -0.0396728515625, -0.024932861328125, -0.01019287109375, 0.004547119140625, 0.019287109375, 0.034027099609375, 0.04876708984375, 0.063507080078125, 0.0782470703125, 0.092987060546875, 0.10772705078125, 0.122467041015625, 0.13720703125, 0.151947021484375, 0.16668701171875, 0.181427001953125, 0.1961669921875, 0.210906982421875, 0.22564697265625, 0.240386962890625, 0.255126953125, 0.269866943359375, 0.28460693359375, 0.299346923828125, 0.3140869140625, 0.328826904296875, 0.34356689453125, 0.358306884765625, 0.373046875]}, "parameters/model.layers.20.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 2.0, 0.0, 3.0, 0.0, 6.0, 4.0, 3.0, 6.0, 7.0, 12.0, 15.0, 25.0, 25.0, 35.0, 63.0, 79.0, 167.0, 385.0, 751.0, 2195.0, 6964.0, 29168.0, 165778.0, 1045444.0, 6303727.0, 16777216.0, 16777216.0, 16777216.0, 9744028.0, 1763018.0, 270290.0, 44960.0, 9966.0, 2897.0, 1102.0, 463.0, 226.0, 122.0, 68.0, 38.0, 30.0, 18.0, 11.0, 20.0, 8.0, 8.0, 6.0, 3.0, 5.0, 1.0, 3.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.380859375, -0.369537353515625, -0.35821533203125, -0.346893310546875, -0.3355712890625, -0.324249267578125, -0.31292724609375, -0.301605224609375, -0.290283203125, -0.278961181640625, -0.26763916015625, -0.256317138671875, -0.2449951171875, -0.233673095703125, -0.22235107421875, -0.211029052734375, -0.19970703125, -0.188385009765625, -0.17706298828125, -0.165740966796875, -0.1544189453125, -0.143096923828125, -0.13177490234375, -0.120452880859375, -0.109130859375, -0.097808837890625, -0.08648681640625, -0.075164794921875, -0.0638427734375, -0.052520751953125, -0.04119873046875, -0.029876708984375, -0.0185546875, -0.007232666015625, 0.00408935546875, 0.015411376953125, 0.0267333984375, 0.038055419921875, 0.04937744140625, 0.060699462890625, 0.072021484375, 0.083343505859375, 0.09466552734375, 0.105987548828125, 0.1173095703125, 0.128631591796875, 0.13995361328125, 0.151275634765625, 0.16259765625, 0.173919677734375, 0.18524169921875, 0.196563720703125, 0.2078857421875, 0.219207763671875, 0.23052978515625, 0.241851806640625, 0.253173828125, 0.264495849609375, 0.27581787109375, 0.287139892578125, 0.2984619140625, 0.309783935546875, 0.32110595703125, 0.332427978515625, 0.34375]}, "parameters/model.layers.20.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 1.0, 0.0, 2.0, 1.0, 4.0, 4.0, 11.0, 14.0, 20.0, 33.0, 69.0, 139.0, 306.0, 3390.0, 163142.0, 3104593.0, 16777216.0, 16777216.0, 10437429.0, 1191837.0, 41180.0, 1013.0, 212.0, 104.0, 47.0, 25.0, 12.0, 11.0, 12.0, 9.0, 1.0, 3.0, 3.0, 6.0, 4.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.41015625, -0.3974609375, -0.384765625, -0.3720703125, -0.359375, -0.3466796875, -0.333984375, -0.3212890625, -0.30859375, -0.2958984375, -0.283203125, -0.2705078125, -0.2578125, -0.2451171875, -0.232421875, -0.2197265625, -0.20703125, -0.1943359375, -0.181640625, -0.1689453125, -0.15625, -0.1435546875, -0.130859375, -0.1181640625, -0.10546875, -0.0927734375, -0.080078125, -0.0673828125, -0.0546875, -0.0419921875, -0.029296875, -0.0166015625, -0.00390625, 0.0087890625, 0.021484375, 0.0341796875, 0.046875, 0.0595703125, 0.072265625, 0.0849609375, 0.09765625, 0.1103515625, 0.123046875, 0.1357421875, 0.1484375, 0.1611328125, 0.173828125, 0.1865234375, 0.19921875, 0.2119140625, 0.224609375, 0.2373046875, 0.25, 0.2626953125, 0.275390625, 0.2880859375, 0.30078125, 0.3134765625, 0.326171875, 0.3388671875, 0.3515625, 0.3642578125, 0.376953125, 0.3896484375, 0.40234375]}, "parameters/model.layers.20.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 3.0, 2.0, 2.0, 12.0, 23.0, 228.0, 1468.0, 1375.0, 468.0, 182.0, 53.0, 45.0, 43.0, 28.0, 23.0, 20.0, 19.0, 16.0, 6.0, 7.0, 7.0, 8.0, 3.0, 3.0, 7.0, 6.0, 1.0, 3.0, 2.0, 0.0, 3.0, 5.0, 6.0, 3.0, 1.0, 3.0, 0.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.0693359375, 0.0850982666015625, 0.100860595703125, 0.1166229248046875, 0.13238525390625, 0.1481475830078125, 0.163909912109375, 0.1796722412109375, 0.1954345703125, 0.2111968994140625, 0.226959228515625, 0.2427215576171875, 0.25848388671875, 0.2742462158203125, 0.290008544921875, 0.3057708740234375, 0.321533203125, 0.3372955322265625, 0.353057861328125, 0.3688201904296875, 0.38458251953125, 0.4003448486328125, 0.416107177734375, 0.4318695068359375, 0.4476318359375, 0.4633941650390625, 0.479156494140625, 0.4949188232421875, 0.51068115234375, 0.5264434814453125, 0.542205810546875, 0.5579681396484375, 0.57373046875, 0.5894927978515625, 0.605255126953125, 0.6210174560546875, 0.63677978515625, 0.6525421142578125, 0.668304443359375, 0.6840667724609375, 0.6998291015625, 0.7155914306640625, 0.731353759765625, 0.7471160888671875, 0.76287841796875, 0.7786407470703125, 0.794403076171875, 0.8101654052734375, 0.825927734375, 0.8416900634765625, 0.857452392578125, 0.8732147216796875, 0.88897705078125, 0.9047393798828125, 0.920501708984375, 0.9362640380859375, 0.9520263671875, 0.9677886962890625, 0.983551025390625, 0.9993133544921875, 1.01507568359375, 1.0308380126953125, 1.046600341796875, 1.0623626708984375, 1.078125]}, "parameters/model.layers.20.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 2.0, 4.0, 2.0, 2.0, 3.0, 0.0, 4.0, 0.0, 3.0, 3.0, 3.0, 1.0, 2.0, 5.0, 3.0, 1.0, 3.0, 0.0, 4.0, 0.0, 2.0, 2.0, 0.0, 2.0, 2.0, 4.0, 6.0, 4.0, 6.0, 6.0, 11.0, 11.0, 13.0, 15.0, 22.0, 10.0, 19.0, 15.0, 19.0, 12.0, 21.0, 37.0, 27.0, 39.0, 44.0, 107.0, 145.0, 239.0, 387.0, 966.0, 1028.0, 593.0, 195.0, 33.0, 1.0, 2.0, 1.0], "bins": [0.072265625, 0.078643798828125, 0.08502197265625, 0.091400146484375, 0.0977783203125, 0.104156494140625, 0.11053466796875, 0.116912841796875, 0.123291015625, 0.129669189453125, 0.13604736328125, 0.142425537109375, 0.1488037109375, 0.155181884765625, 0.16156005859375, 0.167938232421875, 0.17431640625, 0.180694580078125, 0.18707275390625, 0.193450927734375, 0.1998291015625, 0.206207275390625, 0.21258544921875, 0.218963623046875, 0.225341796875, 0.231719970703125, 0.23809814453125, 0.244476318359375, 0.2508544921875, 0.257232666015625, 0.26361083984375, 0.269989013671875, 0.2763671875, 0.282745361328125, 0.28912353515625, 0.295501708984375, 0.3018798828125, 0.308258056640625, 0.31463623046875, 0.321014404296875, 0.327392578125, 0.333770751953125, 0.34014892578125, 0.346527099609375, 0.3529052734375, 0.359283447265625, 0.36566162109375, 0.372039794921875, 0.37841796875, 0.384796142578125, 0.39117431640625, 0.397552490234375, 0.4039306640625, 0.410308837890625, 0.41668701171875, 0.423065185546875, 0.429443359375, 0.435821533203125, 0.44219970703125, 0.448577880859375, 0.4549560546875, 0.461334228515625, 0.46771240234375, 0.474090576171875, 0.48046875]}, "parameters/model.layers.21.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 2.0, 2.0, 7.0, 2.0, 7.0, 1.0, 10.0, 10.0, 23.0, 31.0, 41.0, 65.0, 79.0, 155.0, 261.0, 433.0, 943.0, 1910.0, 4392.0, 9434.0, 22806.0, 50855.0, 114749.0, 243386.0, 512476.0, 953357.0, 1673838.0, 2496004.0, 3078333.0, 2856324.0, 2097828.0, 1279913.0, 717874.0, 350031.0, 171011.0, 78832.0, 35142.0, 14487.0, 6173.0, 3114.0, 1304.0, 689.0, 330.0, 177.0, 91.0, 88.0, 53.0, 37.0, 28.0, 18.0, 13.0, 12.0, 11.0, 2.0, 8.0, 0.0, 3.0, 1.0, 2.0, 1.0, 1.0, 3.0], "bins": [-0.216796875, -0.209747314453125, -0.20269775390625, -0.195648193359375, -0.1885986328125, -0.181549072265625, -0.17449951171875, -0.167449951171875, -0.160400390625, -0.153350830078125, -0.14630126953125, -0.139251708984375, -0.1322021484375, -0.125152587890625, -0.11810302734375, -0.111053466796875, -0.10400390625, -0.096954345703125, -0.08990478515625, -0.082855224609375, -0.0758056640625, -0.068756103515625, -0.06170654296875, -0.054656982421875, -0.047607421875, -0.040557861328125, -0.03350830078125, -0.026458740234375, -0.0194091796875, -0.012359619140625, -0.00531005859375, 0.001739501953125, 0.0087890625, 0.015838623046875, 0.02288818359375, 0.029937744140625, 0.0369873046875, 0.044036865234375, 0.05108642578125, 0.058135986328125, 0.065185546875, 0.072235107421875, 0.07928466796875, 0.086334228515625, 0.0933837890625, 0.100433349609375, 0.10748291015625, 0.114532470703125, 0.12158203125, 0.128631591796875, 0.13568115234375, 0.142730712890625, 0.1497802734375, 0.156829833984375, 0.16387939453125, 0.170928955078125, 0.177978515625, 0.185028076171875, 0.19207763671875, 0.199127197265625, 0.2061767578125, 0.213226318359375, 0.22027587890625, 0.227325439453125, 0.234375]}, "parameters/model.layers.21.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 0.0, 5.0, 4.0, 6.0, 15.0, 17.0, 34.0, 40.0, 62.0, 97.0, 129.0, 204.0, 379.0, 595.0, 1349.0, 3104.0, 6779.0, 14066.0, 32449.0, 67061.0, 134440.0, 241394.0, 408633.0, 588353.0, 712669.0, 682724.0, 530747.0, 355898.0, 203289.0, 108261.0, 55410.0, 24472.0, 11658.0, 5120.0, 2403.0, 1099.0, 535.0, 297.0, 165.0, 130.0, 57.0, 47.0, 40.0, 19.0, 14.0, 8.0, 5.0, 4.0, 4.0, 1.0, 3.0, 1.0, 0.0, 1.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.3125, -0.3023681640625, -0.292236328125, -0.2821044921875, -0.27197265625, -0.2618408203125, -0.251708984375, -0.2415771484375, -0.2314453125, -0.2213134765625, -0.211181640625, -0.2010498046875, -0.19091796875, -0.1807861328125, -0.170654296875, -0.1605224609375, -0.150390625, -0.1402587890625, -0.130126953125, -0.1199951171875, -0.10986328125, -0.0997314453125, -0.089599609375, -0.0794677734375, -0.0693359375, -0.0592041015625, -0.049072265625, -0.0389404296875, -0.02880859375, -0.0186767578125, -0.008544921875, 0.0015869140625, 0.01171875, 0.0218505859375, 0.031982421875, 0.0421142578125, 0.05224609375, 0.0623779296875, 0.072509765625, 0.0826416015625, 0.0927734375, 0.1029052734375, 0.113037109375, 0.1231689453125, 0.13330078125, 0.1434326171875, 0.153564453125, 0.1636962890625, 0.173828125, 0.1839599609375, 0.194091796875, 0.2042236328125, 0.21435546875, 0.2244873046875, 0.234619140625, 0.2447509765625, 0.2548828125, 0.2650146484375, 0.275146484375, 0.2852783203125, 0.29541015625, 0.3055419921875, 0.315673828125, 0.3258056640625, 0.3359375]}, "parameters/model.layers.21.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 3.0, 1.0, 6.0, 8.0, 3.0, 13.0, 14.0, 20.0, 32.0, 42.0, 61.0, 90.0, 163.0, 219.0, 474.0, 813.0, 1551.0, 3338.0, 7029.0, 14324.0, 28759.0, 56885.0, 98243.0, 169254.0, 265571.0, 369107.0, 477988.0, 558499.0, 563298.0, 492211.0, 387076.0, 281341.0, 184806.0, 106164.0, 63325.0, 32033.0, 16084.0, 8048.0, 3682.0, 1627.0, 908.0, 453.0, 265.0, 157.0, 88.0, 66.0, 48.0, 35.0, 20.0, 20.0, 13.0, 6.0, 5.0, 5.0, 0.0, 2.0, 1.0, 0.0, 1.0, 0.0, 2.0, 0.0, 2.0], "bins": [-0.08984375, -0.08675384521484375, -0.0836639404296875, -0.08057403564453125, -0.077484130859375, -0.07439422607421875, -0.0713043212890625, -0.06821441650390625, -0.06512451171875, -0.06203460693359375, -0.0589447021484375, -0.05585479736328125, -0.052764892578125, -0.04967498779296875, -0.0465850830078125, -0.04349517822265625, -0.0404052734375, -0.03731536865234375, -0.0342254638671875, -0.03113555908203125, -0.028045654296875, -0.02495574951171875, -0.0218658447265625, -0.01877593994140625, -0.01568603515625, -0.01259613037109375, -0.0095062255859375, -0.00641632080078125, -0.003326416015625, -0.00023651123046875, 0.0028533935546875, 0.00594329833984375, 0.009033203125, 0.01212310791015625, 0.0152130126953125, 0.01830291748046875, 0.021392822265625, 0.02448272705078125, 0.0275726318359375, 0.03066253662109375, 0.03375244140625, 0.03684234619140625, 0.0399322509765625, 0.04302215576171875, 0.046112060546875, 0.04920196533203125, 0.0522918701171875, 0.05538177490234375, 0.0584716796875, 0.06156158447265625, 0.0646514892578125, 0.06774139404296875, 0.070831298828125, 0.07392120361328125, 0.0770111083984375, 0.08010101318359375, 0.08319091796875, 0.08628082275390625, 0.0893707275390625, 0.09246063232421875, 0.095550537109375, 0.09864044189453125, 0.1017303466796875, 0.10482025146484375, 0.10791015625]}, "parameters/model.layers.21.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 2.0, 7.0, 4.0, 5.0, 10.0, 14.0, 24.0, 29.0, 30.0, 41.0, 37.0, 71.0, 78.0, 86.0, 142.0, 173.0, 212.0, 268.0, 440.0, 675.0, 2000.0, 12700.0, 89214.0, 472741.0, 1725697.0, 3954388.0, 5120264.0, 3530076.0, 1419828.0, 369555.0, 65044.0, 9487.0, 1641.0, 610.0, 412.0, 302.0, 220.0, 143.0, 129.0, 103.0, 72.0, 62.0, 43.0, 39.0, 28.0, 16.0, 7.0, 10.0, 8.0, 0.0, 8.0, 7.0, 1.0, 4.0, 4.0], "bins": [-0.263671875, -0.2562255859375, -0.248779296875, -0.2413330078125, -0.23388671875, -0.2264404296875, -0.218994140625, -0.2115478515625, -0.2041015625, -0.1966552734375, -0.189208984375, -0.1817626953125, -0.17431640625, -0.1668701171875, -0.159423828125, -0.1519775390625, -0.14453125, -0.1370849609375, -0.129638671875, -0.1221923828125, -0.11474609375, -0.1072998046875, -0.099853515625, -0.0924072265625, -0.0849609375, -0.0775146484375, -0.070068359375, -0.0626220703125, -0.05517578125, -0.0477294921875, -0.040283203125, -0.0328369140625, -0.025390625, -0.0179443359375, -0.010498046875, -0.0030517578125, 0.00439453125, 0.0118408203125, 0.019287109375, 0.0267333984375, 0.0341796875, 0.0416259765625, 0.049072265625, 0.0565185546875, 0.06396484375, 0.0714111328125, 0.078857421875, 0.0863037109375, 0.09375, 0.1011962890625, 0.108642578125, 0.1160888671875, 0.12353515625, 0.1309814453125, 0.138427734375, 0.1458740234375, 0.1533203125, 0.1607666015625, 0.168212890625, 0.1756591796875, 0.18310546875, 0.1905517578125, 0.197998046875, 0.2054443359375, 0.212890625]}, "parameters/model.layers.21.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 4.0, 2.0, 2.0, 5.0, 5.0, 10.0, 10.0, 19.0, 23.0, 45.0, 67.0, 98.0, 163.0, 379.0, 950.0, 3304.0, 17630.0, 160332.0, 2075505.0, 16777216.0, 16777216.0, 16777216.0, 5266521.0, 402527.0, 36923.0, 5683.0, 1462.0, 541.0, 248.0, 130.0, 85.0, 51.0, 32.0, 19.0, 18.0, 11.0, 13.0, 8.0, 7.0, 2.0, 1.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.51171875, -0.496063232421875, -0.48040771484375, -0.464752197265625, -0.4490966796875, -0.433441162109375, -0.41778564453125, -0.402130126953125, -0.386474609375, -0.370819091796875, -0.35516357421875, -0.339508056640625, -0.3238525390625, -0.308197021484375, -0.29254150390625, -0.276885986328125, -0.26123046875, -0.245574951171875, -0.22991943359375, -0.214263916015625, -0.1986083984375, -0.182952880859375, -0.16729736328125, -0.151641845703125, -0.135986328125, -0.120330810546875, -0.10467529296875, -0.089019775390625, -0.0733642578125, -0.057708740234375, -0.04205322265625, -0.026397705078125, -0.0107421875, 0.004913330078125, 0.02056884765625, 0.036224365234375, 0.0518798828125, 0.067535400390625, 0.08319091796875, 0.098846435546875, 0.114501953125, 0.130157470703125, 0.14581298828125, 0.161468505859375, 0.1771240234375, 0.192779541015625, 0.20843505859375, 0.224090576171875, 0.23974609375, 0.255401611328125, 0.27105712890625, 0.286712646484375, 0.3023681640625, 0.318023681640625, 0.33367919921875, 0.349334716796875, 0.364990234375, 0.380645751953125, 0.39630126953125, 0.411956787109375, 0.4276123046875, 0.443267822265625, 0.45892333984375, 0.474578857421875, 0.490234375]}, "parameters/model.layers.21.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 6.0, 7.0, 8.0, 21.0, 29.0, 90.0, 273.0, 2365.0, 577650.0, 16777216.0, 16777216.0, 4625988.0, 37856.0, 555.0, 148.0, 53.0, 26.0, 10.0, 10.0, 9.0, 5.0, 3.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.734375, -0.71466064453125, -0.6949462890625, -0.67523193359375, -0.655517578125, -0.63580322265625, -0.6160888671875, -0.59637451171875, -0.57666015625, -0.55694580078125, -0.5372314453125, -0.51751708984375, -0.497802734375, -0.47808837890625, -0.4583740234375, -0.43865966796875, -0.4189453125, -0.39923095703125, -0.3795166015625, -0.35980224609375, -0.340087890625, -0.32037353515625, -0.3006591796875, -0.28094482421875, -0.26123046875, -0.24151611328125, -0.2218017578125, -0.20208740234375, -0.182373046875, -0.16265869140625, -0.1429443359375, -0.12322998046875, -0.103515625, -0.08380126953125, -0.0640869140625, -0.04437255859375, -0.024658203125, -0.00494384765625, 0.0147705078125, 0.03448486328125, 0.05419921875, 0.07391357421875, 0.0936279296875, 0.11334228515625, 0.133056640625, 0.15277099609375, 0.1724853515625, 0.19219970703125, 0.2119140625, 0.23162841796875, 0.2513427734375, 0.27105712890625, 0.290771484375, 0.31048583984375, 0.3302001953125, 0.34991455078125, 0.36962890625, 0.38934326171875, 0.4090576171875, 0.42877197265625, 0.448486328125, 0.46820068359375, 0.4879150390625, 0.50762939453125, 0.52734375]}, "parameters/model.layers.21.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 2.0, 3.0, 4.0, 4.0, 5.0, 23.0, 384.0, 1382.0, 1364.0, 448.0, 166.0, 62.0, 45.0, 49.0, 23.0, 20.0, 10.0, 8.0, 11.0, 5.0, 4.0, 4.0, 3.0, 4.0, 8.0, 2.0, 0.0, 4.0, 3.0, 4.0, 4.0, 3.0, 2.0, 4.0, 3.0, 2.0, 6.0, 5.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.07666015625, 0.09194183349609375, 0.1072235107421875, 0.12250518798828125, 0.137786865234375, 0.15306854248046875, 0.1683502197265625, 0.18363189697265625, 0.19891357421875, 0.21419525146484375, 0.2294769287109375, 0.24475860595703125, 0.260040283203125, 0.27532196044921875, 0.2906036376953125, 0.30588531494140625, 0.3211669921875, 0.33644866943359375, 0.3517303466796875, 0.36701202392578125, 0.382293701171875, 0.39757537841796875, 0.4128570556640625, 0.42813873291015625, 0.44342041015625, 0.45870208740234375, 0.4739837646484375, 0.48926544189453125, 0.504547119140625, 0.5198287963867188, 0.5351104736328125, 0.5503921508789062, 0.565673828125, 0.5809555053710938, 0.5962371826171875, 0.6115188598632812, 0.626800537109375, 0.6420822143554688, 0.6573638916015625, 0.6726455688476562, 0.68792724609375, 0.7032089233398438, 0.7184906005859375, 0.7337722778320312, 0.749053955078125, 0.7643356323242188, 0.7796173095703125, 0.7948989868164062, 0.8101806640625, 0.8254623413085938, 0.8407440185546875, 0.8560256958007812, 0.871307373046875, 0.8865890502929688, 0.9018707275390625, 0.9171524047851562, 0.93243408203125, 0.9477157592773438, 0.9629974365234375, 0.9782791137695312, 0.993560791015625, 1.0088424682617188, 1.0241241455078125, 1.0394058227539062, 1.0546875]}, "parameters/model.layers.21.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 2.0, 1.0, 0.0, 2.0, 2.0, 3.0, 1.0, 2.0, 3.0, 1.0, 4.0, 3.0, 1.0, 3.0, 1.0, 1.0, 2.0, 4.0, 2.0, 4.0, 1.0, 1.0, 3.0, 1.0, 0.0, 2.0, 3.0, 1.0, 1.0, 2.0, 2.0, 2.0, 4.0, 6.0, 6.0, 13.0, 7.0, 4.0, 14.0, 18.0, 21.0, 15.0, 15.0, 14.0, 20.0, 15.0, 16.0, 27.0, 25.0, 41.0, 64.0, 58.0, 105.0, 286.0, 403.0, 686.0, 1321.0, 616.0, 187.0, 24.0, 1.0, 1.0, 1.0], "bins": [0.08154296875, 0.08805084228515625, 0.0945587158203125, 0.10106658935546875, 0.107574462890625, 0.11408233642578125, 0.1205902099609375, 0.12709808349609375, 0.13360595703125, 0.14011383056640625, 0.1466217041015625, 0.15312957763671875, 0.159637451171875, 0.16614532470703125, 0.1726531982421875, 0.17916107177734375, 0.1856689453125, 0.19217681884765625, 0.1986846923828125, 0.20519256591796875, 0.211700439453125, 0.21820831298828125, 0.2247161865234375, 0.23122406005859375, 0.23773193359375, 0.24423980712890625, 0.2507476806640625, 0.25725555419921875, 0.263763427734375, 0.27027130126953125, 0.2767791748046875, 0.28328704833984375, 0.289794921875, 0.29630279541015625, 0.3028106689453125, 0.30931854248046875, 0.315826416015625, 0.32233428955078125, 0.3288421630859375, 0.33535003662109375, 0.34185791015625, 0.34836578369140625, 0.3548736572265625, 0.36138153076171875, 0.367889404296875, 0.37439727783203125, 0.3809051513671875, 0.38741302490234375, 0.3939208984375, 0.40042877197265625, 0.4069366455078125, 0.41344451904296875, 0.419952392578125, 0.42646026611328125, 0.4329681396484375, 0.43947601318359375, 0.44598388671875, 0.45249176025390625, 0.4589996337890625, 0.46550750732421875, 0.472015380859375, 0.47852325439453125, 0.4850311279296875, 0.49153900146484375, 0.498046875]}, "parameters/model.layers.22.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 4.0, 2.0, 3.0, 4.0, 4.0, 6.0, 20.0, 9.0, 12.0, 20.0, 32.0, 54.0, 67.0, 121.0, 207.0, 365.0, 657.0, 1620.0, 3462.0, 8302.0, 21053.0, 49798.0, 109295.0, 226786.0, 452442.0, 849451.0, 1400552.0, 2183019.0, 2833586.0, 2883189.0, 2297670.0, 1542061.0, 926738.0, 515709.0, 244924.0, 125866.0, 58048.0, 24610.0, 9995.0, 3907.0, 1870.0, 743.0, 366.0, 218.0, 121.0, 76.0, 46.0, 35.0, 23.0, 17.0, 11.0, 6.0, 3.0, 5.0, 0.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.2236328125, -0.217071533203125, -0.21051025390625, -0.203948974609375, -0.1973876953125, -0.190826416015625, -0.18426513671875, -0.177703857421875, -0.171142578125, -0.164581298828125, -0.15802001953125, -0.151458740234375, -0.1448974609375, -0.138336181640625, -0.13177490234375, -0.125213623046875, -0.11865234375, -0.112091064453125, -0.10552978515625, -0.098968505859375, -0.0924072265625, -0.085845947265625, -0.07928466796875, -0.072723388671875, -0.066162109375, -0.059600830078125, -0.05303955078125, -0.046478271484375, -0.0399169921875, -0.033355712890625, -0.02679443359375, -0.020233154296875, -0.013671875, -0.007110595703125, -0.00054931640625, 0.006011962890625, 0.0125732421875, 0.019134521484375, 0.02569580078125, 0.032257080078125, 0.038818359375, 0.045379638671875, 0.05194091796875, 0.058502197265625, 0.0650634765625, 0.071624755859375, 0.07818603515625, 0.084747314453125, 0.09130859375, 0.097869873046875, 0.10443115234375, 0.110992431640625, 0.1175537109375, 0.124114990234375, 0.13067626953125, 0.137237548828125, 0.143798828125, 0.150360107421875, 0.15692138671875, 0.163482666015625, 0.1700439453125, 0.176605224609375, 0.18316650390625, 0.189727783203125, 0.1962890625]}, "parameters/model.layers.22.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 1.0, 2.0, 3.0, 3.0, 3.0, 5.0, 13.0, 3.0, 14.0, 18.0, 29.0, 34.0, 55.0, 77.0, 93.0, 157.0, 297.0, 548.0, 1081.0, 2474.0, 5491.0, 12655.0, 29204.0, 64353.0, 130262.0, 241917.0, 426956.0, 632515.0, 766156.0, 710330.0, 515530.0, 317609.0, 171792.0, 89646.0, 41348.0, 18727.0, 8118.0, 3424.0, 1617.0, 701.0, 370.0, 233.0, 128.0, 100.0, 76.0, 43.0, 30.0, 17.0, 12.0, 13.0, 3.0, 8.0, 2.0, 1.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.33203125, -0.321563720703125, -0.31109619140625, -0.300628662109375, -0.2901611328125, -0.279693603515625, -0.26922607421875, -0.258758544921875, -0.248291015625, -0.237823486328125, -0.22735595703125, -0.216888427734375, -0.2064208984375, -0.195953369140625, -0.18548583984375, -0.175018310546875, -0.16455078125, -0.154083251953125, -0.14361572265625, -0.133148193359375, -0.1226806640625, -0.112213134765625, -0.10174560546875, -0.091278076171875, -0.080810546875, -0.070343017578125, -0.05987548828125, -0.049407958984375, -0.0389404296875, -0.028472900390625, -0.01800537109375, -0.007537841796875, 0.0029296875, 0.013397216796875, 0.02386474609375, 0.034332275390625, 0.0447998046875, 0.055267333984375, 0.06573486328125, 0.076202392578125, 0.086669921875, 0.097137451171875, 0.10760498046875, 0.118072509765625, 0.1285400390625, 0.139007568359375, 0.14947509765625, 0.159942626953125, 0.17041015625, 0.180877685546875, 0.19134521484375, 0.201812744140625, 0.2122802734375, 0.222747802734375, 0.23321533203125, 0.243682861328125, 0.254150390625, 0.264617919921875, 0.27508544921875, 0.285552978515625, 0.2960205078125, 0.306488037109375, 0.31695556640625, 0.327423095703125, 0.337890625]}, "parameters/model.layers.22.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 3.0, 0.0, 6.0, 9.0, 9.0, 16.0, 20.0, 24.0, 40.0, 54.0, 104.0, 144.0, 229.0, 394.0, 737.0, 1230.0, 2633.0, 5311.0, 10360.0, 20041.0, 37879.0, 63130.0, 103072.0, 166287.0, 233408.0, 317897.0, 401331.0, 462818.0, 499978.0, 466916.0, 408115.0, 322910.0, 241116.0, 170260.0, 108330.0, 65822.0, 40122.0, 20879.0, 10977.0, 5655.0, 2921.0, 1350.0, 726.0, 392.0, 219.0, 160.0, 78.0, 63.0, 48.0, 22.0, 22.0, 15.0, 8.0, 8.0, 1.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.0869140625, -0.08406829833984375, -0.0812225341796875, -0.07837677001953125, -0.075531005859375, -0.07268524169921875, -0.0698394775390625, -0.06699371337890625, -0.06414794921875, -0.06130218505859375, -0.0584564208984375, -0.05561065673828125, -0.052764892578125, -0.04991912841796875, -0.0470733642578125, -0.04422760009765625, -0.0413818359375, -0.03853607177734375, -0.0356903076171875, -0.03284454345703125, -0.029998779296875, -0.02715301513671875, -0.0243072509765625, -0.02146148681640625, -0.01861572265625, -0.01576995849609375, -0.0129241943359375, -0.01007843017578125, -0.007232666015625, -0.00438690185546875, -0.0015411376953125, 0.00130462646484375, 0.004150390625, 0.00699615478515625, 0.0098419189453125, 0.01268768310546875, 0.015533447265625, 0.01837921142578125, 0.0212249755859375, 0.02407073974609375, 0.02691650390625, 0.02976226806640625, 0.0326080322265625, 0.03545379638671875, 0.038299560546875, 0.04114532470703125, 0.0439910888671875, 0.04683685302734375, 0.0496826171875, 0.05252838134765625, 0.0553741455078125, 0.05821990966796875, 0.061065673828125, 0.06391143798828125, 0.0667572021484375, 0.06960296630859375, 0.07244873046875, 0.07529449462890625, 0.0781402587890625, 0.08098602294921875, 0.083831787109375, 0.08667755126953125, 0.0895233154296875, 0.09236907958984375, 0.09521484375]}, "parameters/model.layers.22.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 1.0, 2.0, 4.0, 4.0, 2.0, 4.0, 3.0, 6.0, 4.0, 6.0, 4.0, 5.0, 2.0, 7.0, 6.0, 9.0, 6.0, 10.0, 12.0, 17.0, 27.0, 38.0, 84.0, 148.0, 614.0, 5246.0, 59390.0, 482719.0, 2219730.0, 5079166.0, 5457970.0, 2694424.0, 676759.0, 91609.0, 7801.0, 840.0, 232.0, 95.0, 60.0, 24.0, 21.0, 17.0, 12.0, 8.0, 8.0, 9.0, 8.0, 7.0, 5.0, 8.0, 3.0, 2.0, 0.0, 5.0, 1.0, 4.0, 3.0, 1.0, 0.0, 1.0], "bins": [-0.302734375, -0.293853759765625, -0.28497314453125, -0.276092529296875, -0.2672119140625, -0.258331298828125, -0.24945068359375, -0.240570068359375, -0.231689453125, -0.222808837890625, -0.21392822265625, -0.205047607421875, -0.1961669921875, -0.187286376953125, -0.17840576171875, -0.169525146484375, -0.16064453125, -0.151763916015625, -0.14288330078125, -0.134002685546875, -0.1251220703125, -0.116241455078125, -0.10736083984375, -0.098480224609375, -0.089599609375, -0.080718994140625, -0.07183837890625, -0.062957763671875, -0.0540771484375, -0.045196533203125, -0.03631591796875, -0.027435302734375, -0.0185546875, -0.009674072265625, -0.00079345703125, 0.008087158203125, 0.0169677734375, 0.025848388671875, 0.03472900390625, 0.043609619140625, 0.052490234375, 0.061370849609375, 0.07025146484375, 0.079132080078125, 0.0880126953125, 0.096893310546875, 0.10577392578125, 0.114654541015625, 0.12353515625, 0.132415771484375, 0.14129638671875, 0.150177001953125, 0.1590576171875, 0.167938232421875, 0.17681884765625, 0.185699462890625, 0.194580078125, 0.203460693359375, 0.21234130859375, 0.221221923828125, 0.2301025390625, 0.238983154296875, 0.24786376953125, 0.256744384765625, 0.265625]}, "parameters/model.layers.22.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 0.0, 1.0, 1.0, 1.0, 2.0, 3.0, 1.0, 2.0, 5.0, 3.0, 7.0, 6.0, 7.0, 19.0, 23.0, 30.0, 44.0, 57.0, 105.0, 184.0, 459.0, 933.0, 3078.0, 11232.0, 53810.0, 335708.0, 2158653.0, 10858195.0, 16777216.0, 16777216.0, 16777216.0, 6404502.0, 1106195.0, 171240.0, 29280.0, 6600.0, 1894.0, 669.0, 317.0, 155.0, 96.0, 54.0, 45.0, 43.0, 24.0, 15.0, 18.0, 6.0, 8.0, 5.0, 1.0, 4.0, 3.0, 1.0, 1.0, 3.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.369140625, -0.358062744140625, -0.34698486328125, -0.335906982421875, -0.3248291015625, -0.313751220703125, -0.30267333984375, -0.291595458984375, -0.280517578125, -0.269439697265625, -0.25836181640625, -0.247283935546875, -0.2362060546875, -0.225128173828125, -0.21405029296875, -0.202972412109375, -0.19189453125, -0.180816650390625, -0.16973876953125, -0.158660888671875, -0.1475830078125, -0.136505126953125, -0.12542724609375, -0.114349365234375, -0.103271484375, -0.092193603515625, -0.08111572265625, -0.070037841796875, -0.0589599609375, -0.047882080078125, -0.03680419921875, -0.025726318359375, -0.0146484375, -0.003570556640625, 0.00750732421875, 0.018585205078125, 0.0296630859375, 0.040740966796875, 0.05181884765625, 0.062896728515625, 0.073974609375, 0.085052490234375, 0.09613037109375, 0.107208251953125, 0.1182861328125, 0.129364013671875, 0.14044189453125, 0.151519775390625, 0.16259765625, 0.173675537109375, 0.18475341796875, 0.195831298828125, 0.2069091796875, 0.217987060546875, 0.22906494140625, 0.240142822265625, 0.251220703125, 0.262298583984375, 0.27337646484375, 0.284454345703125, 0.2955322265625, 0.306610107421875, 0.31768798828125, 0.328765869140625, 0.33984375]}, "parameters/model.layers.22.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 1.0, 2.0, 1.0, 1.0, 3.0, 3.0, 1.0, 6.0, 11.0, 14.0, 68.0, 181.0, 4545.0, 518316.0, 10229980.0, 16777216.0, 14713113.0, 1072732.0, 12842.0, 261.0, 61.0, 29.0, 17.0, 10.0, 6.0, 3.0, 3.0, 0.0, 2.0, 0.0, 1.0, 1.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.546875, -0.53106689453125, -0.5152587890625, -0.49945068359375, -0.483642578125, -0.46783447265625, -0.4520263671875, -0.43621826171875, -0.42041015625, -0.40460205078125, -0.3887939453125, -0.37298583984375, -0.357177734375, -0.34136962890625, -0.3255615234375, -0.30975341796875, -0.2939453125, -0.27813720703125, -0.2623291015625, -0.24652099609375, -0.230712890625, -0.21490478515625, -0.1990966796875, -0.18328857421875, -0.16748046875, -0.15167236328125, -0.1358642578125, -0.12005615234375, -0.104248046875, -0.08843994140625, -0.0726318359375, -0.05682373046875, -0.041015625, -0.02520751953125, -0.0093994140625, 0.00640869140625, 0.022216796875, 0.03802490234375, 0.0538330078125, 0.06964111328125, 0.08544921875, 0.10125732421875, 0.1170654296875, 0.13287353515625, 0.148681640625, 0.16448974609375, 0.1802978515625, 0.19610595703125, 0.2119140625, 0.22772216796875, 0.2435302734375, 0.25933837890625, 0.275146484375, 0.29095458984375, 0.3067626953125, 0.32257080078125, 0.33837890625, 0.35418701171875, 0.3699951171875, 0.38580322265625, 0.401611328125, 0.41741943359375, 0.4332275390625, 0.44903564453125, 0.46484375]}, "parameters/model.layers.22.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 2.0, 2.0, 5.0, 9.0, 8.0, 87.0, 1097.0, 1779.0, 568.0, 166.0, 69.0, 61.0, 52.0, 27.0, 25.0, 19.0, 19.0, 12.0, 2.0, 7.0, 7.0, 7.0, 6.0, 3.0, 4.0, 3.0, 2.0, 4.0, 4.0, 4.0, 4.0, 2.0, 3.0, 2.0, 6.0, 2.0, 2.0, 4.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.07568359375, 0.09281158447265625, 0.1099395751953125, 0.12706756591796875, 0.144195556640625, 0.16132354736328125, 0.1784515380859375, 0.19557952880859375, 0.21270751953125, 0.22983551025390625, 0.2469635009765625, 0.26409149169921875, 0.281219482421875, 0.29834747314453125, 0.3154754638671875, 0.33260345458984375, 0.3497314453125, 0.36685943603515625, 0.3839874267578125, 0.40111541748046875, 0.418243408203125, 0.43537139892578125, 0.4524993896484375, 0.46962738037109375, 0.48675537109375, 0.5038833618164062, 0.5210113525390625, 0.5381393432617188, 0.555267333984375, 0.5723953247070312, 0.5895233154296875, 0.6066513061523438, 0.623779296875, 0.6409072875976562, 0.6580352783203125, 0.6751632690429688, 0.692291259765625, 0.7094192504882812, 0.7265472412109375, 0.7436752319335938, 0.76080322265625, 0.7779312133789062, 0.7950592041015625, 0.8121871948242188, 0.829315185546875, 0.8464431762695312, 0.8635711669921875, 0.8806991577148438, 0.8978271484375, 0.9149551391601562, 0.9320831298828125, 0.9492111206054688, 0.966339111328125, 0.9834671020507812, 1.0005950927734375, 1.0177230834960938, 1.03485107421875, 1.0519790649414062, 1.0691070556640625, 1.0862350463867188, 1.103363037109375, 1.1204910278320312, 1.1376190185546875, 1.1547470092773438, 1.171875]}, "parameters/model.layers.22.post_attention_layernorm.weight": {"_type": "histogram", "values": [2.0, 0.0, 2.0, 1.0, 4.0, 2.0, 3.0, 0.0, 4.0, 2.0, 3.0, 2.0, 4.0, 0.0, 2.0, 2.0, 3.0, 5.0, 2.0, 0.0, 6.0, 1.0, 0.0, 3.0, 1.0, 1.0, 4.0, 2.0, 5.0, 3.0, 6.0, 3.0, 15.0, 16.0, 7.0, 12.0, 16.0, 8.0, 13.0, 17.0, 7.0, 15.0, 11.0, 17.0, 11.0, 28.0, 16.0, 36.0, 42.0, 45.0, 101.0, 173.0, 242.0, 633.0, 1240.0, 913.0, 359.0, 21.0, 0.0, 3.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.0771484375, 0.0843048095703125, 0.091461181640625, 0.0986175537109375, 0.10577392578125, 0.1129302978515625, 0.120086669921875, 0.1272430419921875, 0.1343994140625, 0.1415557861328125, 0.148712158203125, 0.1558685302734375, 0.16302490234375, 0.1701812744140625, 0.177337646484375, 0.1844940185546875, 0.191650390625, 0.1988067626953125, 0.205963134765625, 0.2131195068359375, 0.22027587890625, 0.2274322509765625, 0.234588623046875, 0.2417449951171875, 0.2489013671875, 0.2560577392578125, 0.263214111328125, 0.2703704833984375, 0.27752685546875, 0.2846832275390625, 0.291839599609375, 0.2989959716796875, 0.30615234375, 0.3133087158203125, 0.320465087890625, 0.3276214599609375, 0.33477783203125, 0.3419342041015625, 0.349090576171875, 0.3562469482421875, 0.3634033203125, 0.3705596923828125, 0.377716064453125, 0.3848724365234375, 0.39202880859375, 0.3991851806640625, 0.406341552734375, 0.4134979248046875, 0.420654296875, 0.4278106689453125, 0.434967041015625, 0.4421234130859375, 0.44927978515625, 0.4564361572265625, 0.463592529296875, 0.4707489013671875, 0.4779052734375, 0.4850616455078125, 0.492218017578125, 0.4993743896484375, 0.50653076171875, 0.5136871337890625, 0.520843505859375, 0.5279998779296875, 0.53515625]}, "parameters/model.layers.23.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 5.0, 7.0, 4.0, 8.0, 11.0, 13.0, 23.0, 18.0, 37.0, 25.0, 70.0, 103.0, 117.0, 209.0, 296.0, 675.0, 1344.0, 3007.0, 8011.0, 20838.0, 53624.0, 131980.0, 282561.0, 593483.0, 1121756.0, 1872858.0, 2741108.0, 3229860.0, 2716916.0, 1849602.0, 1077963.0, 579348.0, 282873.0, 122944.0, 52065.0, 20095.0, 7486.0, 3144.0, 1283.0, 556.0, 323.0, 164.0, 114.0, 93.0, 51.0, 40.0, 32.0, 18.0, 21.0, 14.0, 6.0, 4.0, 3.0, 1.0, 1.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.2314453125, -0.2243194580078125, -0.217193603515625, -0.2100677490234375, -0.20294189453125, -0.1958160400390625, -0.188690185546875, -0.1815643310546875, -0.1744384765625, -0.1673126220703125, -0.160186767578125, -0.1530609130859375, -0.14593505859375, -0.1388092041015625, -0.131683349609375, -0.1245574951171875, -0.117431640625, -0.1103057861328125, -0.103179931640625, -0.0960540771484375, -0.08892822265625, -0.0818023681640625, -0.074676513671875, -0.0675506591796875, -0.0604248046875, -0.0532989501953125, -0.046173095703125, -0.0390472412109375, -0.03192138671875, -0.0247955322265625, -0.017669677734375, -0.0105438232421875, -0.00341796875, 0.0037078857421875, 0.010833740234375, 0.0179595947265625, 0.02508544921875, 0.0322113037109375, 0.039337158203125, 0.0464630126953125, 0.0535888671875, 0.0607147216796875, 0.067840576171875, 0.0749664306640625, 0.08209228515625, 0.0892181396484375, 0.096343994140625, 0.1034698486328125, 0.110595703125, 0.1177215576171875, 0.124847412109375, 0.1319732666015625, 0.13909912109375, 0.1462249755859375, 0.153350830078125, 0.1604766845703125, 0.1676025390625, 0.1747283935546875, 0.181854248046875, 0.1889801025390625, 0.19610595703125, 0.2032318115234375, 0.210357666015625, 0.2174835205078125, 0.224609375]}, "parameters/model.layers.23.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 2.0, 4.0, 8.0, 3.0, 12.0, 13.0, 25.0, 32.0, 43.0, 83.0, 103.0, 162.0, 279.0, 547.0, 1222.0, 3224.0, 8766.0, 21912.0, 59826.0, 140054.0, 289861.0, 531945.0, 794515.0, 867728.0, 685730.0, 419171.0, 207573.0, 98711.0, 38496.0, 14648.0, 5589.0, 2109.0, 854.0, 398.0, 228.0, 158.0, 86.0, 65.0, 37.0, 21.0, 17.0, 21.0, 8.0, 4.0, 2.0, 1.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.3046875, -0.292633056640625, -0.28057861328125, -0.268524169921875, -0.2564697265625, -0.244415283203125, -0.23236083984375, -0.220306396484375, -0.208251953125, -0.196197509765625, -0.18414306640625, -0.172088623046875, -0.1600341796875, -0.147979736328125, -0.13592529296875, -0.123870849609375, -0.11181640625, -0.099761962890625, -0.08770751953125, -0.075653076171875, -0.0635986328125, -0.051544189453125, -0.03948974609375, -0.027435302734375, -0.015380859375, -0.003326416015625, 0.00872802734375, 0.020782470703125, 0.0328369140625, 0.044891357421875, 0.05694580078125, 0.069000244140625, 0.0810546875, 0.093109130859375, 0.10516357421875, 0.117218017578125, 0.1292724609375, 0.141326904296875, 0.15338134765625, 0.165435791015625, 0.177490234375, 0.189544677734375, 0.20159912109375, 0.213653564453125, 0.2257080078125, 0.237762451171875, 0.24981689453125, 0.261871337890625, 0.27392578125, 0.285980224609375, 0.29803466796875, 0.310089111328125, 0.3221435546875, 0.334197998046875, 0.34625244140625, 0.358306884765625, 0.370361328125, 0.382415771484375, 0.39447021484375, 0.406524658203125, 0.4185791015625, 0.430633544921875, 0.44268798828125, 0.454742431640625, 0.466796875]}, "parameters/model.layers.23.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 2.0, 3.0, 0.0, 0.0, 2.0, 9.0, 6.0, 5.0, 4.0, 14.0, 23.0, 22.0, 37.0, 71.0, 118.0, 236.0, 409.0, 727.0, 1574.0, 3427.0, 6743.0, 14550.0, 27804.0, 50557.0, 91023.0, 142603.0, 214396.0, 301017.0, 385725.0, 460052.0, 508974.0, 490139.0, 427985.0, 351920.0, 259363.0, 185769.0, 115771.0, 72176.0, 39630.0, 20846.0, 10634.0, 5109.0, 2257.0, 1230.0, 585.0, 315.0, 166.0, 99.0, 54.0, 47.0, 26.0, 25.0, 6.0, 8.0, 1.0, 1.0, 2.0, 2.0, 3.0], "bins": [-0.11083984375, -0.1077423095703125, -0.104644775390625, -0.1015472412109375, -0.09844970703125, -0.0953521728515625, -0.092254638671875, -0.0891571044921875, -0.0860595703125, -0.0829620361328125, -0.079864501953125, -0.0767669677734375, -0.07366943359375, -0.0705718994140625, -0.067474365234375, -0.0643768310546875, -0.061279296875, -0.0581817626953125, -0.055084228515625, -0.0519866943359375, -0.04888916015625, -0.0457916259765625, -0.042694091796875, -0.0395965576171875, -0.0364990234375, -0.0334014892578125, -0.030303955078125, -0.0272064208984375, -0.02410888671875, -0.0210113525390625, -0.017913818359375, -0.0148162841796875, -0.01171875, -0.0086212158203125, -0.005523681640625, -0.0024261474609375, 0.00067138671875, 0.0037689208984375, 0.006866455078125, 0.0099639892578125, 0.0130615234375, 0.0161590576171875, 0.019256591796875, 0.0223541259765625, 0.02545166015625, 0.0285491943359375, 0.031646728515625, 0.0347442626953125, 0.037841796875, 0.0409393310546875, 0.044036865234375, 0.0471343994140625, 0.05023193359375, 0.0533294677734375, 0.056427001953125, 0.0595245361328125, 0.0626220703125, 0.0657196044921875, 0.068817138671875, 0.0719146728515625, 0.07501220703125, 0.0781097412109375, 0.081207275390625, 0.0843048095703125, 0.08740234375]}, "parameters/model.layers.23.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0, 1.0, 2.0, 0.0, 0.0, 6.0, 10.0, 20.0, 19.0, 41.0, 57.0, 117.0, 329.0, 7217.0, 174097.0, 1694408.0, 5860344.0, 6486791.0, 2264283.0, 275974.0, 12629.0, 539.0, 155.0, 64.0, 33.0, 20.0, 23.0, 8.0, 5.0, 4.0, 3.0, 3.0, 1.0, 2.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.3671875, -0.355743408203125, -0.34429931640625, -0.332855224609375, -0.3214111328125, -0.309967041015625, -0.29852294921875, -0.287078857421875, -0.275634765625, -0.264190673828125, -0.25274658203125, -0.241302490234375, -0.2298583984375, -0.218414306640625, -0.20697021484375, -0.195526123046875, -0.18408203125, -0.172637939453125, -0.16119384765625, -0.149749755859375, -0.1383056640625, -0.126861572265625, -0.11541748046875, -0.103973388671875, -0.092529296875, -0.081085205078125, -0.06964111328125, -0.058197021484375, -0.0467529296875, -0.035308837890625, -0.02386474609375, -0.012420654296875, -0.0009765625, 0.010467529296875, 0.02191162109375, 0.033355712890625, 0.0447998046875, 0.056243896484375, 0.06768798828125, 0.079132080078125, 0.090576171875, 0.102020263671875, 0.11346435546875, 0.124908447265625, 0.1363525390625, 0.147796630859375, 0.15924072265625, 0.170684814453125, 0.18212890625, 0.193572998046875, 0.20501708984375, 0.216461181640625, 0.2279052734375, 0.239349365234375, 0.25079345703125, 0.262237548828125, 0.273681640625, 0.285125732421875, 0.29656982421875, 0.308013916015625, 0.3194580078125, 0.330902099609375, 0.34234619140625, 0.353790283203125, 0.365234375]}, "parameters/model.layers.23.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0, 2.0, 3.0, 5.0, 7.0, 12.0, 7.0, 13.0, 20.0, 33.0, 46.0, 66.0, 110.0, 212.0, 473.0, 1417.0, 4889.0, 21202.0, 117956.0, 808293.0, 5122858.0, 16777216.0, 16777216.0, 16777216.0, 12347145.0, 2424334.0, 346169.0, 54814.0, 10436.0, 2631.0, 860.0, 359.0, 175.0, 91.0, 58.0, 35.0, 21.0, 13.0, 17.0, 8.0, 7.0, 5.0, 3.0, 4.0, 7.0, 1.0, 2.0, 0.0, 0.0, 1.0, 3.0], "bins": [-0.40625, -0.3948974609375, -0.383544921875, -0.3721923828125, -0.36083984375, -0.3494873046875, -0.338134765625, -0.3267822265625, -0.3154296875, -0.3040771484375, -0.292724609375, -0.2813720703125, -0.27001953125, -0.2586669921875, -0.247314453125, -0.2359619140625, -0.224609375, -0.2132568359375, -0.201904296875, -0.1905517578125, -0.17919921875, -0.1678466796875, -0.156494140625, -0.1451416015625, -0.1337890625, -0.1224365234375, -0.111083984375, -0.0997314453125, -0.08837890625, -0.0770263671875, -0.065673828125, -0.0543212890625, -0.04296875, -0.0316162109375, -0.020263671875, -0.0089111328125, 0.00244140625, 0.0137939453125, 0.025146484375, 0.0364990234375, 0.0478515625, 0.0592041015625, 0.070556640625, 0.0819091796875, 0.09326171875, 0.1046142578125, 0.115966796875, 0.1273193359375, 0.138671875, 0.1500244140625, 0.161376953125, 0.1727294921875, 0.18408203125, 0.1954345703125, 0.206787109375, 0.2181396484375, 0.2294921875, 0.2408447265625, 0.252197265625, 0.2635498046875, 0.27490234375, 0.2862548828125, 0.297607421875, 0.3089599609375, 0.3203125]}, "parameters/model.layers.23.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 1.0, 0.0, 2.0, 4.0, 3.0, 6.0, 17.0, 33.0, 101.0, 746.0, 106690.0, 4276792.0, 16777216.0, 16777216.0, 3319807.0, 70562.0, 498.0, 73.0, 35.0, 12.0, 11.0, 1.0, 2.0, 1.0, 4.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.4453125, -0.42938232421875, -0.4134521484375, -0.39752197265625, -0.381591796875, -0.36566162109375, -0.3497314453125, -0.33380126953125, -0.31787109375, -0.30194091796875, -0.2860107421875, -0.27008056640625, -0.254150390625, -0.23822021484375, -0.2222900390625, -0.20635986328125, -0.1904296875, -0.17449951171875, -0.1585693359375, -0.14263916015625, -0.126708984375, -0.11077880859375, -0.0948486328125, -0.07891845703125, -0.06298828125, -0.04705810546875, -0.0311279296875, -0.01519775390625, 0.000732421875, 0.01666259765625, 0.0325927734375, 0.04852294921875, 0.064453125, 0.08038330078125, 0.0963134765625, 0.11224365234375, 0.128173828125, 0.14410400390625, 0.1600341796875, 0.17596435546875, 0.19189453125, 0.20782470703125, 0.2237548828125, 0.23968505859375, 0.255615234375, 0.27154541015625, 0.2874755859375, 0.30340576171875, 0.3193359375, 0.33526611328125, 0.3511962890625, 0.36712646484375, 0.383056640625, 0.39898681640625, 0.4149169921875, 0.43084716796875, 0.44677734375, 0.46270751953125, 0.4786376953125, 0.49456787109375, 0.510498046875, 0.52642822265625, 0.5423583984375, 0.55828857421875, 0.57421875]}, "parameters/model.layers.23.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 0.0, 2.0, 4.0, 2.0, 8.0, 6.0, 11.0, 85.0, 1049.0, 1971.0, 469.0, 130.0, 86.0, 42.0, 42.0, 34.0, 22.0, 14.0, 15.0, 9.0, 7.0, 9.0, 6.0, 6.0, 7.0, 2.0, 6.0, 3.0, 4.0, 2.0, 6.0, 5.0, 2.0, 2.0, 6.0, 2.0, 2.0, 3.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.0703125, 0.085693359375, 0.10107421875, 0.116455078125, 0.1318359375, 0.147216796875, 0.16259765625, 0.177978515625, 0.193359375, 0.208740234375, 0.22412109375, 0.239501953125, 0.2548828125, 0.270263671875, 0.28564453125, 0.301025390625, 0.31640625, 0.331787109375, 0.34716796875, 0.362548828125, 0.3779296875, 0.393310546875, 0.40869140625, 0.424072265625, 0.439453125, 0.454833984375, 0.47021484375, 0.485595703125, 0.5009765625, 0.516357421875, 0.53173828125, 0.547119140625, 0.5625, 0.577880859375, 0.59326171875, 0.608642578125, 0.6240234375, 0.639404296875, 0.65478515625, 0.670166015625, 0.685546875, 0.700927734375, 0.71630859375, 0.731689453125, 0.7470703125, 0.762451171875, 0.77783203125, 0.793212890625, 0.80859375, 0.823974609375, 0.83935546875, 0.854736328125, 0.8701171875, 0.885498046875, 0.90087890625, 0.916259765625, 0.931640625, 0.947021484375, 0.96240234375, 0.977783203125, 0.9931640625, 1.008544921875, 1.02392578125, 1.039306640625, 1.0546875]}, "parameters/model.layers.23.post_attention_layernorm.weight": {"_type": "histogram", "values": [2.0, 1.0, 2.0, 1.0, 4.0, 3.0, 1.0, 1.0, 5.0, 2.0, 1.0, 5.0, 1.0, 1.0, 2.0, 6.0, 1.0, 6.0, 1.0, 4.0, 0.0, 2.0, 0.0, 2.0, 1.0, 5.0, 5.0, 3.0, 5.0, 8.0, 7.0, 14.0, 16.0, 7.0, 8.0, 13.0, 12.0, 8.0, 11.0, 8.0, 15.0, 11.0, 10.0, 18.0, 18.0, 15.0, 23.0, 32.0, 39.0, 48.0, 98.0, 123.0, 292.0, 708.0, 968.0, 1188.0, 299.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.091796875, 0.098968505859375, 0.10614013671875, 0.113311767578125, 0.1204833984375, 0.127655029296875, 0.13482666015625, 0.141998291015625, 0.149169921875, 0.156341552734375, 0.16351318359375, 0.170684814453125, 0.1778564453125, 0.185028076171875, 0.19219970703125, 0.199371337890625, 0.20654296875, 0.213714599609375, 0.22088623046875, 0.228057861328125, 0.2352294921875, 0.242401123046875, 0.24957275390625, 0.256744384765625, 0.263916015625, 0.271087646484375, 0.27825927734375, 0.285430908203125, 0.2926025390625, 0.299774169921875, 0.30694580078125, 0.314117431640625, 0.3212890625, 0.328460693359375, 0.33563232421875, 0.342803955078125, 0.3499755859375, 0.357147216796875, 0.36431884765625, 0.371490478515625, 0.378662109375, 0.385833740234375, 0.39300537109375, 0.400177001953125, 0.4073486328125, 0.414520263671875, 0.42169189453125, 0.428863525390625, 0.43603515625, 0.443206787109375, 0.45037841796875, 0.457550048828125, 0.4647216796875, 0.471893310546875, 0.47906494140625, 0.486236572265625, 0.493408203125, 0.500579833984375, 0.50775146484375, 0.514923095703125, 0.5220947265625, 0.529266357421875, 0.53643798828125, 0.543609619140625, 0.55078125]}, "parameters/model.layers.24.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 3.0, 1.0, 3.0, 3.0, 7.0, 9.0, 10.0, 8.0, 25.0, 33.0, 50.0, 76.0, 96.0, 115.0, 174.0, 352.0, 570.0, 1078.0, 2251.0, 5149.0, 13489.0, 33278.0, 77881.0, 180789.0, 363963.0, 750990.0, 1302015.0, 2139950.0, 2992962.0, 3143754.0, 2427342.0, 1565127.0, 896277.0, 473800.0, 224308.0, 105750.0, 44229.0, 18078.0, 6970.0, 3189.0, 1379.0, 613.0, 345.0, 231.0, 159.0, 85.0, 71.0, 53.0, 39.0, 24.0, 15.0, 9.0, 12.0, 7.0, 5.0, 3.0, 1.0, 2.0, 3.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.2080078125, -0.20111083984375, -0.1942138671875, -0.18731689453125, -0.180419921875, -0.17352294921875, -0.1666259765625, -0.15972900390625, -0.15283203125, -0.14593505859375, -0.1390380859375, -0.13214111328125, -0.125244140625, -0.11834716796875, -0.1114501953125, -0.10455322265625, -0.09765625, -0.09075927734375, -0.0838623046875, -0.07696533203125, -0.070068359375, -0.06317138671875, -0.0562744140625, -0.04937744140625, -0.04248046875, -0.03558349609375, -0.0286865234375, -0.02178955078125, -0.014892578125, -0.00799560546875, -0.0010986328125, 0.00579833984375, 0.0126953125, 0.01959228515625, 0.0264892578125, 0.03338623046875, 0.040283203125, 0.04718017578125, 0.0540771484375, 0.06097412109375, 0.06787109375, 0.07476806640625, 0.0816650390625, 0.08856201171875, 0.095458984375, 0.10235595703125, 0.1092529296875, 0.11614990234375, 0.123046875, 0.12994384765625, 0.1368408203125, 0.14373779296875, 0.150634765625, 0.15753173828125, 0.1644287109375, 0.17132568359375, 0.17822265625, 0.18511962890625, 0.1920166015625, 0.19891357421875, 0.205810546875, 0.21270751953125, 0.2196044921875, 0.22650146484375, 0.2333984375]}, "parameters/model.layers.24.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 5.0, 5.0, 3.0, 2.0, 5.0, 7.0, 13.0, 5.0, 16.0, 32.0, 41.0, 49.0, 56.0, 89.0, 130.0, 152.0, 272.0, 442.0, 896.0, 1997.0, 4321.0, 9421.0, 21026.0, 46331.0, 98442.0, 186449.0, 360315.0, 592274.0, 807365.0, 793091.0, 572863.0, 344393.0, 181913.0, 91390.0, 44078.0, 19829.0, 8776.0, 3869.0, 1727.0, 869.0, 505.0, 259.0, 158.0, 119.0, 84.0, 56.0, 38.0, 33.0, 25.0, 14.0, 17.0, 9.0, 7.0, 7.0, 1.0, 2.0, 0.0, 2.0, 2.0, 1.0, 2.0, 1.0], "bins": [-0.3203125, -0.309967041015625, -0.29962158203125, -0.289276123046875, -0.2789306640625, -0.268585205078125, -0.25823974609375, -0.247894287109375, -0.237548828125, -0.227203369140625, -0.21685791015625, -0.206512451171875, -0.1961669921875, -0.185821533203125, -0.17547607421875, -0.165130615234375, -0.15478515625, -0.144439697265625, -0.13409423828125, -0.123748779296875, -0.1134033203125, -0.103057861328125, -0.09271240234375, -0.082366943359375, -0.072021484375, -0.061676025390625, -0.05133056640625, -0.040985107421875, -0.0306396484375, -0.020294189453125, -0.00994873046875, 0.000396728515625, 0.0107421875, 0.021087646484375, 0.03143310546875, 0.041778564453125, 0.0521240234375, 0.062469482421875, 0.07281494140625, 0.083160400390625, 0.093505859375, 0.103851318359375, 0.11419677734375, 0.124542236328125, 0.1348876953125, 0.145233154296875, 0.15557861328125, 0.165924072265625, 0.17626953125, 0.186614990234375, 0.19696044921875, 0.207305908203125, 0.2176513671875, 0.227996826171875, 0.23834228515625, 0.248687744140625, 0.259033203125, 0.269378662109375, 0.27972412109375, 0.290069580078125, 0.3004150390625, 0.310760498046875, 0.32110595703125, 0.331451416015625, 0.341796875]}, "parameters/model.layers.24.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 0.0, 1.0, 1.0, 9.0, 9.0, 5.0, 16.0, 28.0, 30.0, 30.0, 71.0, 100.0, 145.0, 252.0, 444.0, 864.0, 1383.0, 2488.0, 4706.0, 8886.0, 15208.0, 28768.0, 47866.0, 81302.0, 123339.0, 192651.0, 270116.0, 353101.0, 431513.0, 486844.0, 493097.0, 440970.0, 367036.0, 285713.0, 204839.0, 139895.0, 86145.0, 55490.0, 31166.0, 18381.0, 9418.0, 5295.0, 2856.0, 1668.0, 845.0, 497.0, 319.0, 183.0, 90.0, 80.0, 46.0, 34.0, 18.0, 16.0, 6.0, 13.0, 2.0, 5.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.1005859375, -0.09735107421875, -0.0941162109375, -0.09088134765625, -0.087646484375, -0.08441162109375, -0.0811767578125, -0.07794189453125, -0.07470703125, -0.07147216796875, -0.0682373046875, -0.06500244140625, -0.061767578125, -0.05853271484375, -0.0552978515625, -0.05206298828125, -0.048828125, -0.04559326171875, -0.0423583984375, -0.03912353515625, -0.035888671875, -0.03265380859375, -0.0294189453125, -0.02618408203125, -0.02294921875, -0.01971435546875, -0.0164794921875, -0.01324462890625, -0.010009765625, -0.00677490234375, -0.0035400390625, -0.00030517578125, 0.0029296875, 0.00616455078125, 0.0093994140625, 0.01263427734375, 0.015869140625, 0.01910400390625, 0.0223388671875, 0.02557373046875, 0.02880859375, 0.03204345703125, 0.0352783203125, 0.03851318359375, 0.041748046875, 0.04498291015625, 0.0482177734375, 0.05145263671875, 0.0546875, 0.05792236328125, 0.0611572265625, 0.06439208984375, 0.067626953125, 0.07086181640625, 0.0740966796875, 0.07733154296875, 0.08056640625, 0.08380126953125, 0.0870361328125, 0.09027099609375, 0.093505859375, 0.09674072265625, 0.0999755859375, 0.10321044921875, 0.1064453125]}, "parameters/model.layers.24.self_attn.o_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 1.0, 0.0, 4.0, 2.0, 3.0, 4.0, 1.0, 3.0, 7.0, 11.0, 24.0, 38.0, 40.0, 83.0, 206.0, 897.0, 5000.0, 31346.0, 165337.0, 656875.0, 1829330.0, 3536203.0, 4410069.0, 3497072.0, 1802918.0, 644540.0, 161792.0, 29431.0, 4646.0, 877.0, 205.0, 84.0, 53.0, 32.0, 16.0, 16.0, 12.0, 8.0, 4.0, 4.0, 3.0, 4.0, 1.0, 0.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.1796875, -0.172637939453125, -0.16558837890625, -0.158538818359375, -0.1514892578125, -0.144439697265625, -0.13739013671875, -0.130340576171875, -0.123291015625, -0.116241455078125, -0.10919189453125, -0.102142333984375, -0.0950927734375, -0.088043212890625, -0.08099365234375, -0.073944091796875, -0.06689453125, -0.059844970703125, -0.05279541015625, -0.045745849609375, -0.0386962890625, -0.031646728515625, -0.02459716796875, -0.017547607421875, -0.010498046875, -0.003448486328125, 0.00360107421875, 0.010650634765625, 0.0177001953125, 0.024749755859375, 0.03179931640625, 0.038848876953125, 0.0458984375, 0.052947998046875, 0.05999755859375, 0.067047119140625, 0.0740966796875, 0.081146240234375, 0.08819580078125, 0.095245361328125, 0.102294921875, 0.109344482421875, 0.11639404296875, 0.123443603515625, 0.1304931640625, 0.137542724609375, 0.14459228515625, 0.151641845703125, 0.15869140625, 0.165740966796875, 0.17279052734375, 0.179840087890625, 0.1868896484375, 0.193939208984375, 0.20098876953125, 0.208038330078125, 0.215087890625, 0.222137451171875, 0.22918701171875, 0.236236572265625, 0.2432861328125, 0.250335693359375, 0.25738525390625, 0.264434814453125, 0.271484375]}, "parameters/model.layers.24.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 2.0, 4.0, 1.0, 2.0, 3.0, 8.0, 6.0, 10.0, 9.0, 8.0, 14.0, 25.0, 40.0, 69.0, 118.0, 239.0, 595.0, 2014.0, 8547.0, 50047.0, 387288.0, 3319178.0, 16777216.0, 16777216.0, 16777216.0, 11701406.0, 1761838.0, 204877.0, 27918.0, 5193.0, 1371.0, 422.0, 205.0, 102.0, 58.0, 46.0, 23.0, 9.0, 7.0, 4.0, 2.0, 6.0, 2.0, 6.0, 4.0, 0.0, 1.0, 2.0, 0.0, 2.0], "bins": [-0.48046875, -0.4677734375, -0.455078125, -0.4423828125, -0.4296875, -0.4169921875, -0.404296875, -0.3916015625, -0.37890625, -0.3662109375, -0.353515625, -0.3408203125, -0.328125, -0.3154296875, -0.302734375, -0.2900390625, -0.27734375, -0.2646484375, -0.251953125, -0.2392578125, -0.2265625, -0.2138671875, -0.201171875, -0.1884765625, -0.17578125, -0.1630859375, -0.150390625, -0.1376953125, -0.125, -0.1123046875, -0.099609375, -0.0869140625, -0.07421875, -0.0615234375, -0.048828125, -0.0361328125, -0.0234375, -0.0107421875, 0.001953125, 0.0146484375, 0.02734375, 0.0400390625, 0.052734375, 0.0654296875, 0.078125, 0.0908203125, 0.103515625, 0.1162109375, 0.12890625, 0.1416015625, 0.154296875, 0.1669921875, 0.1796875, 0.1923828125, 0.205078125, 0.2177734375, 0.23046875, 0.2431640625, 0.255859375, 0.2685546875, 0.28125, 0.2939453125, 0.306640625, 0.3193359375, 0.33203125]}, "parameters/model.layers.24.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 2.0, 1.0, 1.0, 0.0, 2.0, 3.0, 1.0, 5.0, 3.0, 15.0, 27.0, 58.0, 411.0, 27676.0, 1077632.0, 10700794.0, 16777216.0, 16717222.0, 2520010.0, 97094.0, 1330.0, 94.0, 44.0, 13.0, 12.0, 7.0, 2.0, 4.0, 3.0, 1.0, 1.0, 0.0, 1.0, 2.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.318359375, -0.304901123046875, -0.29144287109375, -0.277984619140625, -0.2645263671875, -0.251068115234375, -0.23760986328125, -0.224151611328125, -0.210693359375, -0.197235107421875, -0.18377685546875, -0.170318603515625, -0.1568603515625, -0.143402099609375, -0.12994384765625, -0.116485595703125, -0.10302734375, -0.089569091796875, -0.07611083984375, -0.062652587890625, -0.0491943359375, -0.035736083984375, -0.02227783203125, -0.008819580078125, 0.004638671875, 0.018096923828125, 0.03155517578125, 0.045013427734375, 0.0584716796875, 0.071929931640625, 0.08538818359375, 0.098846435546875, 0.1123046875, 0.125762939453125, 0.13922119140625, 0.152679443359375, 0.1661376953125, 0.179595947265625, 0.19305419921875, 0.206512451171875, 0.219970703125, 0.233428955078125, 0.24688720703125, 0.260345458984375, 0.2738037109375, 0.287261962890625, 0.30072021484375, 0.314178466796875, 0.32763671875, 0.341094970703125, 0.35455322265625, 0.368011474609375, 0.3814697265625, 0.394927978515625, 0.40838623046875, 0.421844482421875, 0.435302734375, 0.448760986328125, 0.46221923828125, 0.475677490234375, 0.4891357421875, 0.502593994140625, 0.51605224609375, 0.529510498046875, 0.54296875]}, "parameters/model.layers.24.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 3.0, 0.0, 0.0, 2.0, 6.0, 1.0, 6.0, 4.0, 5.0, 22.0, 510.0, 1523.0, 1226.0, 255.0, 138.0, 85.0, 51.0, 47.0, 31.0, 31.0, 28.0, 8.0, 16.0, 6.0, 11.0, 5.0, 7.0, 7.0, 6.0, 2.0, 4.0, 3.0, 6.0, 2.0, 2.0, 1.0, 1.0, 5.0, 2.0, 7.0, 0.0, 2.0, 4.0, 3.0, 3.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [0.06591796875, 0.08295440673828125, 0.0999908447265625, 0.11702728271484375, 0.134063720703125, 0.15110015869140625, 0.1681365966796875, 0.18517303466796875, 0.20220947265625, 0.21924591064453125, 0.2362823486328125, 0.25331878662109375, 0.270355224609375, 0.28739166259765625, 0.3044281005859375, 0.32146453857421875, 0.3385009765625, 0.35553741455078125, 0.3725738525390625, 0.38961029052734375, 0.406646728515625, 0.42368316650390625, 0.4407196044921875, 0.45775604248046875, 0.47479248046875, 0.49182891845703125, 0.5088653564453125, 0.5259017944335938, 0.542938232421875, 0.5599746704101562, 0.5770111083984375, 0.5940475463867188, 0.611083984375, 0.6281204223632812, 0.6451568603515625, 0.6621932983398438, 0.679229736328125, 0.6962661743164062, 0.7133026123046875, 0.7303390502929688, 0.74737548828125, 0.7644119262695312, 0.7814483642578125, 0.7984848022460938, 0.815521240234375, 0.8325576782226562, 0.8495941162109375, 0.8666305541992188, 0.8836669921875, 0.9007034301757812, 0.9177398681640625, 0.9347763061523438, 0.951812744140625, 0.9688491821289062, 0.9858856201171875, 1.0029220581054688, 1.01995849609375, 1.0369949340820312, 1.0540313720703125, 1.0710678100585938, 1.088104248046875, 1.1051406860351562, 1.1221771240234375, 1.1392135620117188, 1.15625]}, "parameters/model.layers.24.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 3.0, 2.0, 3.0, 3.0, 0.0, 4.0, 1.0, 4.0, 1.0, 5.0, 1.0, 2.0, 2.0, 5.0, 0.0, 4.0, 6.0, 1.0, 1.0, 2.0, 1.0, 2.0, 3.0, 2.0, 3.0, 6.0, 6.0, 10.0, 11.0, 13.0, 16.0, 5.0, 8.0, 14.0, 14.0, 7.0, 10.0, 7.0, 13.0, 15.0, 18.0, 12.0, 24.0, 23.0, 26.0, 43.0, 59.0, 101.0, 137.0, 335.0, 821.0, 1868.0, 388.0, 20.0, 1.0, 1.0, 0.0, 1.0], "bins": [0.0634765625, 0.0710906982421875, 0.078704833984375, 0.0863189697265625, 0.09393310546875, 0.1015472412109375, 0.109161376953125, 0.1167755126953125, 0.1243896484375, 0.1320037841796875, 0.139617919921875, 0.1472320556640625, 0.15484619140625, 0.1624603271484375, 0.170074462890625, 0.1776885986328125, 0.185302734375, 0.1929168701171875, 0.200531005859375, 0.2081451416015625, 0.21575927734375, 0.2233734130859375, 0.230987548828125, 0.2386016845703125, 0.2462158203125, 0.2538299560546875, 0.261444091796875, 0.2690582275390625, 0.27667236328125, 0.2842864990234375, 0.291900634765625, 0.2995147705078125, 0.30712890625, 0.3147430419921875, 0.322357177734375, 0.3299713134765625, 0.33758544921875, 0.3451995849609375, 0.352813720703125, 0.3604278564453125, 0.3680419921875, 0.3756561279296875, 0.383270263671875, 0.3908843994140625, 0.39849853515625, 0.4061126708984375, 0.413726806640625, 0.4213409423828125, 0.428955078125, 0.4365692138671875, 0.444183349609375, 0.4517974853515625, 0.45941162109375, 0.4670257568359375, 0.474639892578125, 0.4822540283203125, 0.4898681640625, 0.4974822998046875, 0.505096435546875, 0.5127105712890625, 0.52032470703125, 0.5279388427734375, 0.535552978515625, 0.5431671142578125, 0.55078125]}, "parameters/model.layers.25.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 1.0, 3.0, 0.0, 4.0, 10.0, 6.0, 10.0, 14.0, 22.0, 32.0, 49.0, 74.0, 89.0, 102.0, 221.0, 266.0, 488.0, 747.0, 1469.0, 3237.0, 7960.0, 21703.0, 55770.0, 125706.0, 293891.0, 638420.0, 1170004.0, 2028862.0, 2904023.0, 3234763.0, 2652513.0, 1738095.0, 996527.0, 508542.0, 221644.0, 102182.0, 41812.0, 16015.0, 6169.0, 2516.0, 1258.0, 692.0, 424.0, 285.0, 156.0, 134.0, 81.0, 58.0, 51.0, 29.0, 23.0, 24.0, 4.0, 8.0, 7.0, 5.0, 6.0, 3.0, 2.0, 0.0, 2.0], "bins": [-0.2275390625, -0.2205047607421875, -0.213470458984375, -0.2064361572265625, -0.19940185546875, -0.1923675537109375, -0.185333251953125, -0.1782989501953125, -0.1712646484375, -0.1642303466796875, -0.157196044921875, -0.1501617431640625, -0.14312744140625, -0.1360931396484375, -0.129058837890625, -0.1220245361328125, -0.114990234375, -0.1079559326171875, -0.100921630859375, -0.0938873291015625, -0.08685302734375, -0.0798187255859375, -0.072784423828125, -0.0657501220703125, -0.0587158203125, -0.0516815185546875, -0.044647216796875, -0.0376129150390625, -0.03057861328125, -0.0235443115234375, -0.016510009765625, -0.0094757080078125, -0.00244140625, 0.0045928955078125, 0.011627197265625, 0.0186614990234375, 0.02569580078125, 0.0327301025390625, 0.039764404296875, 0.0467987060546875, 0.0538330078125, 0.0608673095703125, 0.067901611328125, 0.0749359130859375, 0.08197021484375, 0.0890045166015625, 0.096038818359375, 0.1030731201171875, 0.110107421875, 0.1171417236328125, 0.124176025390625, 0.1312103271484375, 0.13824462890625, 0.1452789306640625, 0.152313232421875, 0.1593475341796875, 0.1663818359375, 0.1734161376953125, 0.180450439453125, 0.1874847412109375, 0.19451904296875, 0.2015533447265625, 0.208587646484375, 0.2156219482421875, 0.22265625]}, "parameters/model.layers.25.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 5.0, 1.0, 3.0, 3.0, 4.0, 8.0, 6.0, 6.0, 20.0, 23.0, 27.0, 51.0, 54.0, 81.0, 147.0, 194.0, 365.0, 870.0, 2288.0, 6497.0, 17695.0, 50702.0, 120175.0, 263678.0, 516685.0, 876491.0, 998733.0, 671104.0, 368474.0, 176437.0, 76805.0, 29472.0, 10631.0, 3896.0, 1244.0, 584.0, 292.0, 164.0, 117.0, 75.0, 55.0, 38.0, 20.0, 19.0, 13.0, 16.0, 15.0, 5.0, 5.0, 4.0, 2.0, 2.0, 1.0], "bins": [-0.458984375, -0.4466552734375, -0.434326171875, -0.4219970703125, -0.40966796875, -0.3973388671875, -0.385009765625, -0.3726806640625, -0.3603515625, -0.3480224609375, -0.335693359375, -0.3233642578125, -0.31103515625, -0.2987060546875, -0.286376953125, -0.2740478515625, -0.26171875, -0.2493896484375, -0.237060546875, -0.2247314453125, -0.21240234375, -0.2000732421875, -0.187744140625, -0.1754150390625, -0.1630859375, -0.1507568359375, -0.138427734375, -0.1260986328125, -0.11376953125, -0.1014404296875, -0.089111328125, -0.0767822265625, -0.064453125, -0.0521240234375, -0.039794921875, -0.0274658203125, -0.01513671875, -0.0028076171875, 0.009521484375, 0.0218505859375, 0.0341796875, 0.0465087890625, 0.058837890625, 0.0711669921875, 0.08349609375, 0.0958251953125, 0.108154296875, 0.1204833984375, 0.1328125, 0.1451416015625, 0.157470703125, 0.1697998046875, 0.18212890625, 0.1944580078125, 0.206787109375, 0.2191162109375, 0.2314453125, 0.2437744140625, 0.256103515625, 0.2684326171875, 0.28076171875, 0.2930908203125, 0.305419921875, 0.3177490234375, 0.330078125]}, "parameters/model.layers.25.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 4.0, 6.0, 8.0, 13.0, 21.0, 19.0, 20.0, 33.0, 59.0, 90.0, 97.0, 146.0, 238.0, 333.0, 513.0, 760.0, 1621.0, 3156.0, 6622.0, 13045.0, 27981.0, 55200.0, 93673.0, 159209.0, 240745.0, 337910.0, 434072.0, 509084.0, 538749.0, 498445.0, 414387.0, 315596.0, 223742.0, 138751.0, 86177.0, 46171.0, 24370.0, 11653.0, 5252.0, 2774.0, 1405.0, 766.0, 408.0, 300.0, 188.0, 154.0, 77.0, 61.0, 63.0, 40.0, 23.0, 19.0, 20.0, 15.0, 5.0, 2.0, 3.0, 1.0, 4.0, 1.0, 2.0], "bins": [-0.11328125, -0.10967254638671875, -0.1060638427734375, -0.10245513916015625, -0.098846435546875, -0.09523773193359375, -0.0916290283203125, -0.08802032470703125, -0.08441162109375, -0.08080291748046875, -0.0771942138671875, -0.07358551025390625, -0.069976806640625, -0.06636810302734375, -0.0627593994140625, -0.05915069580078125, -0.0555419921875, -0.05193328857421875, -0.0483245849609375, -0.04471588134765625, -0.041107177734375, -0.03749847412109375, -0.0338897705078125, -0.03028106689453125, -0.02667236328125, -0.02306365966796875, -0.0194549560546875, -0.01584625244140625, -0.012237548828125, -0.00862884521484375, -0.0050201416015625, -0.00141143798828125, 0.002197265625, 0.00580596923828125, 0.0094146728515625, 0.01302337646484375, 0.016632080078125, 0.02024078369140625, 0.0238494873046875, 0.02745819091796875, 0.03106689453125, 0.03467559814453125, 0.0382843017578125, 0.04189300537109375, 0.045501708984375, 0.04911041259765625, 0.0527191162109375, 0.05632781982421875, 0.0599365234375, 0.06354522705078125, 0.0671539306640625, 0.07076263427734375, 0.074371337890625, 0.07798004150390625, 0.0815887451171875, 0.08519744873046875, 0.08880615234375, 0.09241485595703125, 0.0960235595703125, 0.09963226318359375, 0.103240966796875, 0.10684967041015625, 0.1104583740234375, 0.11406707763671875, 0.11767578125]}, "parameters/model.layers.25.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 7.0, 2.0, 11.0, 17.0, 28.0, 48.0, 81.0, 128.0, 253.0, 526.0, 3064.0, 44836.0, 514332.0, 2748355.0, 6256650.0, 5240437.0, 1703065.0, 245076.0, 18120.0, 1307.0, 382.0, 198.0, 112.0, 62.0, 41.0, 20.0, 17.0, 10.0, 7.0, 3.0, 5.0, 3.0, 2.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.333984375, -0.323150634765625, -0.31231689453125, -0.301483154296875, -0.2906494140625, -0.279815673828125, -0.26898193359375, -0.258148193359375, -0.247314453125, -0.236480712890625, -0.22564697265625, -0.214813232421875, -0.2039794921875, -0.193145751953125, -0.18231201171875, -0.171478271484375, -0.16064453125, -0.149810791015625, -0.13897705078125, -0.128143310546875, -0.1173095703125, -0.106475830078125, -0.09564208984375, -0.084808349609375, -0.073974609375, -0.063140869140625, -0.05230712890625, -0.041473388671875, -0.0306396484375, -0.019805908203125, -0.00897216796875, 0.001861572265625, 0.0126953125, 0.023529052734375, 0.03436279296875, 0.045196533203125, 0.0560302734375, 0.066864013671875, 0.07769775390625, 0.088531494140625, 0.099365234375, 0.110198974609375, 0.12103271484375, 0.131866455078125, 0.1427001953125, 0.153533935546875, 0.16436767578125, 0.175201416015625, 0.18603515625, 0.196868896484375, 0.20770263671875, 0.218536376953125, 0.2293701171875, 0.240203857421875, 0.25103759765625, 0.261871337890625, 0.272705078125, 0.283538818359375, 0.29437255859375, 0.305206298828125, 0.3160400390625, 0.326873779296875, 0.33770751953125, 0.348541259765625, 0.359375]}, "parameters/model.layers.25.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 1.0, 2.0, 0.0, 0.0, 1.0, 5.0, 5.0, 4.0, 3.0, 6.0, 6.0, 11.0, 10.0, 19.0, 17.0, 30.0, 29.0, 51.0, 77.0, 108.0, 206.0, 418.0, 932.0, 2397.0, 6499.0, 23424.0, 94598.0, 424620.0, 2152310.0, 8424633.0, 16777216.0, 16777216.0, 16777216.0, 13817145.0, 4113035.0, 860415.0, 180959.0, 40950.0, 11250.0, 3613.0, 1336.0, 584.0, 283.0, 140.0, 116.0, 61.0, 57.0, 27.0, 22.0, 24.0, 13.0, 7.0, 14.0, 8.0, 4.0, 4.0, 1.0, 2.0, 2.0, 4.0], "bins": [-0.337890625, -0.32843017578125, -0.3189697265625, -0.30950927734375, -0.300048828125, -0.29058837890625, -0.2811279296875, -0.27166748046875, -0.26220703125, -0.25274658203125, -0.2432861328125, -0.23382568359375, -0.224365234375, -0.21490478515625, -0.2054443359375, -0.19598388671875, -0.1865234375, -0.17706298828125, -0.1676025390625, -0.15814208984375, -0.148681640625, -0.13922119140625, -0.1297607421875, -0.12030029296875, -0.11083984375, -0.10137939453125, -0.0919189453125, -0.08245849609375, -0.072998046875, -0.06353759765625, -0.0540771484375, -0.04461669921875, -0.03515625, -0.02569580078125, -0.0162353515625, -0.00677490234375, 0.002685546875, 0.01214599609375, 0.0216064453125, 0.03106689453125, 0.04052734375, 0.04998779296875, 0.0594482421875, 0.06890869140625, 0.078369140625, 0.08782958984375, 0.0972900390625, 0.10675048828125, 0.1162109375, 0.12567138671875, 0.1351318359375, 0.14459228515625, 0.154052734375, 0.16351318359375, 0.1729736328125, 0.18243408203125, 0.19189453125, 0.20135498046875, 0.2108154296875, 0.22027587890625, 0.229736328125, 0.23919677734375, 0.2486572265625, 0.25811767578125, 0.267578125]}, "parameters/model.layers.25.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [3.0, 1.0, 0.0, 1.0, 2.0, 1.0, 0.0, 4.0, 4.0, 6.0, 5.0, 20.0, 36.0, 101.0, 861.0, 63951.0, 2008230.0, 15364183.0, 16777216.0, 11726388.0, 1216248.0, 30572.0, 390.0, 67.0, 28.0, 22.0, 6.0, 6.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.25390625, -0.2401123046875, -0.226318359375, -0.2125244140625, -0.19873046875, -0.1849365234375, -0.171142578125, -0.1573486328125, -0.1435546875, -0.1297607421875, -0.115966796875, -0.1021728515625, -0.08837890625, -0.0745849609375, -0.060791015625, -0.0469970703125, -0.033203125, -0.0194091796875, -0.005615234375, 0.0081787109375, 0.02197265625, 0.0357666015625, 0.049560546875, 0.0633544921875, 0.0771484375, 0.0909423828125, 0.104736328125, 0.1185302734375, 0.13232421875, 0.1461181640625, 0.159912109375, 0.1737060546875, 0.1875, 0.2012939453125, 0.215087890625, 0.2288818359375, 0.24267578125, 0.2564697265625, 0.270263671875, 0.2840576171875, 0.2978515625, 0.3116455078125, 0.325439453125, 0.3392333984375, 0.35302734375, 0.3668212890625, 0.380615234375, 0.3944091796875, 0.408203125, 0.4219970703125, 0.435791015625, 0.4495849609375, 0.46337890625, 0.4771728515625, 0.490966796875, 0.5047607421875, 0.5185546875, 0.5323486328125, 0.546142578125, 0.5599365234375, 0.57373046875, 0.5875244140625, 0.601318359375, 0.6151123046875, 0.62890625]}, "parameters/model.layers.25.input_layernorm.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 0.0, 2.0, 1.0, 3.0, 3.0, 7.0, 9.0, 60.0, 775.0, 1914.0, 558.0, 292.0, 118.0, 52.0, 63.0, 35.0, 39.0, 28.0, 17.0, 14.0, 7.0, 15.0, 5.0, 9.0, 5.0, 4.0, 7.0, 2.0, 2.0, 3.0, 2.0, 2.0, 6.0, 3.0, 4.0, 9.0, 2.0, 2.0, 2.0, 2.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0], "bins": [0.060546875, 0.077789306640625, 0.09503173828125, 0.112274169921875, 0.1295166015625, 0.146759033203125, 0.16400146484375, 0.181243896484375, 0.198486328125, 0.215728759765625, 0.23297119140625, 0.250213623046875, 0.2674560546875, 0.284698486328125, 0.30194091796875, 0.319183349609375, 0.33642578125, 0.353668212890625, 0.37091064453125, 0.388153076171875, 0.4053955078125, 0.422637939453125, 0.43988037109375, 0.457122802734375, 0.474365234375, 0.491607666015625, 0.50885009765625, 0.526092529296875, 0.5433349609375, 0.560577392578125, 0.57781982421875, 0.595062255859375, 0.6123046875, 0.629547119140625, 0.64678955078125, 0.664031982421875, 0.6812744140625, 0.698516845703125, 0.71575927734375, 0.733001708984375, 0.750244140625, 0.767486572265625, 0.78472900390625, 0.801971435546875, 0.8192138671875, 0.836456298828125, 0.85369873046875, 0.870941162109375, 0.88818359375, 0.905426025390625, 0.92266845703125, 0.939910888671875, 0.9571533203125, 0.974395751953125, 0.99163818359375, 1.008880615234375, 1.026123046875, 1.043365478515625, 1.06060791015625, 1.077850341796875, 1.0950927734375, 1.112335205078125, 1.12957763671875, 1.146820068359375, 1.1640625]}, "parameters/model.layers.25.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 4.0, 4.0, 0.0, 2.0, 2.0, 3.0, 2.0, 4.0, 3.0, 1.0, 2.0, 3.0, 2.0, 2.0, 4.0, 3.0, 4.0, 0.0, 3.0, 2.0, 0.0, 2.0, 7.0, 1.0, 6.0, 1.0, 6.0, 14.0, 14.0, 14.0, 10.0, 8.0, 14.0, 12.0, 7.0, 13.0, 9.0, 9.0, 18.0, 15.0, 17.0, 22.0, 21.0, 37.0, 47.0, 72.0, 122.0, 462.0, 917.0, 1451.0, 642.0, 50.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.07177734375, 0.07981109619140625, 0.0878448486328125, 0.09587860107421875, 0.103912353515625, 0.11194610595703125, 0.1199798583984375, 0.12801361083984375, 0.13604736328125, 0.14408111572265625, 0.1521148681640625, 0.16014862060546875, 0.168182373046875, 0.17621612548828125, 0.1842498779296875, 0.19228363037109375, 0.2003173828125, 0.20835113525390625, 0.2163848876953125, 0.22441864013671875, 0.232452392578125, 0.24048614501953125, 0.2485198974609375, 0.25655364990234375, 0.26458740234375, 0.27262115478515625, 0.2806549072265625, 0.28868865966796875, 0.296722412109375, 0.30475616455078125, 0.3127899169921875, 0.32082366943359375, 0.328857421875, 0.33689117431640625, 0.3449249267578125, 0.35295867919921875, 0.360992431640625, 0.36902618408203125, 0.3770599365234375, 0.38509368896484375, 0.39312744140625, 0.40116119384765625, 0.4091949462890625, 0.41722869873046875, 0.425262451171875, 0.43329620361328125, 0.4413299560546875, 0.44936370849609375, 0.4573974609375, 0.46543121337890625, 0.4734649658203125, 0.48149871826171875, 0.489532470703125, 0.49756622314453125, 0.5055999755859375, 0.5136337280273438, 0.52166748046875, 0.5297012329101562, 0.5377349853515625, 0.5457687377929688, 0.553802490234375, 0.5618362426757812, 0.5698699951171875, 0.5779037475585938, 0.5859375]}, "parameters/model.layers.26.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 1.0, 3.0, 2.0, 3.0, 6.0, 6.0, 10.0, 11.0, 20.0, 33.0, 51.0, 84.0, 100.0, 153.0, 210.0, 356.0, 500.0, 1020.0, 1844.0, 4262.0, 10132.0, 25409.0, 60007.0, 128243.0, 276337.0, 584756.0, 1053913.0, 1801876.0, 2630432.0, 3125963.0, 2750549.0, 1941506.0, 1149661.0, 651006.0, 309255.0, 151830.0, 66285.0, 30218.0, 11900.0, 4404.0, 2251.0, 1085.0, 604.0, 294.0, 194.0, 147.0, 98.0, 54.0, 34.0, 32.0, 15.0, 12.0, 9.0, 5.0, 6.0, 5.0, 3.0, 4.0, 2.0, 0.0, 2.0], "bins": [-0.2197265625, -0.212982177734375, -0.20623779296875, -0.199493408203125, -0.1927490234375, -0.186004638671875, -0.17926025390625, -0.172515869140625, -0.165771484375, -0.159027099609375, -0.15228271484375, -0.145538330078125, -0.1387939453125, -0.132049560546875, -0.12530517578125, -0.118560791015625, -0.11181640625, -0.105072021484375, -0.09832763671875, -0.091583251953125, -0.0848388671875, -0.078094482421875, -0.07135009765625, -0.064605712890625, -0.057861328125, -0.051116943359375, -0.04437255859375, -0.037628173828125, -0.0308837890625, -0.024139404296875, -0.01739501953125, -0.010650634765625, -0.00390625, 0.002838134765625, 0.00958251953125, 0.016326904296875, 0.0230712890625, 0.029815673828125, 0.03656005859375, 0.043304443359375, 0.050048828125, 0.056793212890625, 0.06353759765625, 0.070281982421875, 0.0770263671875, 0.083770751953125, 0.09051513671875, 0.097259521484375, 0.10400390625, 0.110748291015625, 0.11749267578125, 0.124237060546875, 0.1309814453125, 0.137725830078125, 0.14447021484375, 0.151214599609375, 0.157958984375, 0.164703369140625, 0.17144775390625, 0.178192138671875, 0.1849365234375, 0.191680908203125, 0.19842529296875, 0.205169677734375, 0.2119140625]}, "parameters/model.layers.26.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 4.0, 3.0, 5.0, 3.0, 6.0, 16.0, 13.0, 30.0, 25.0, 42.0, 59.0, 90.0, 110.0, 194.0, 262.0, 458.0, 875.0, 1921.0, 4311.0, 10289.0, 21174.0, 49077.0, 103940.0, 200500.0, 383930.0, 635364.0, 855256.0, 789410.0, 532118.0, 305337.0, 156415.0, 77122.0, 36581.0, 15775.0, 7195.0, 3280.0, 1443.0, 627.0, 350.0, 198.0, 133.0, 115.0, 73.0, 49.0, 33.0, 32.0, 10.0, 12.0, 10.0, 10.0, 3.0, 4.0, 5.0, 1.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.357421875, -0.346527099609375, -0.33563232421875, -0.324737548828125, -0.3138427734375, -0.302947998046875, -0.29205322265625, -0.281158447265625, -0.270263671875, -0.259368896484375, -0.24847412109375, -0.237579345703125, -0.2266845703125, -0.215789794921875, -0.20489501953125, -0.194000244140625, -0.18310546875, -0.172210693359375, -0.16131591796875, -0.150421142578125, -0.1395263671875, -0.128631591796875, -0.11773681640625, -0.106842041015625, -0.095947265625, -0.085052490234375, -0.07415771484375, -0.063262939453125, -0.0523681640625, -0.041473388671875, -0.03057861328125, -0.019683837890625, -0.0087890625, 0.002105712890625, 0.01300048828125, 0.023895263671875, 0.0347900390625, 0.045684814453125, 0.05657958984375, 0.067474365234375, 0.078369140625, 0.089263916015625, 0.10015869140625, 0.111053466796875, 0.1219482421875, 0.132843017578125, 0.14373779296875, 0.154632568359375, 0.16552734375, 0.176422119140625, 0.18731689453125, 0.198211669921875, 0.2091064453125, 0.220001220703125, 0.23089599609375, 0.241790771484375, 0.252685546875, 0.263580322265625, 0.27447509765625, 0.285369873046875, 0.2962646484375, 0.307159423828125, 0.31805419921875, 0.328948974609375, 0.33984375]}, "parameters/model.layers.26.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 3.0, 10.0, 10.0, 14.0, 28.0, 31.0, 59.0, 75.0, 130.0, 258.0, 558.0, 1439.0, 3706.0, 9545.0, 23970.0, 56663.0, 112648.0, 209039.0, 333217.0, 486589.0, 614953.0, 673160.0, 581716.0, 442662.0, 297133.0, 178693.0, 92569.0, 44036.0, 18960.0, 7542.0, 2858.0, 1101.0, 441.0, 186.0, 99.0, 70.0, 49.0, 30.0, 14.0, 11.0, 11.0, 6.0, 2.0, 2.0, 1.0, 0.0, 2.0], "bins": [-0.1796875, -0.1750030517578125, -0.170318603515625, -0.1656341552734375, -0.16094970703125, -0.1562652587890625, -0.151580810546875, -0.1468963623046875, -0.1422119140625, -0.1375274658203125, -0.132843017578125, -0.1281585693359375, -0.12347412109375, -0.1187896728515625, -0.114105224609375, -0.1094207763671875, -0.104736328125, -0.1000518798828125, -0.095367431640625, -0.0906829833984375, -0.08599853515625, -0.0813140869140625, -0.076629638671875, -0.0719451904296875, -0.0672607421875, -0.0625762939453125, -0.057891845703125, -0.0532073974609375, -0.04852294921875, -0.0438385009765625, -0.039154052734375, -0.0344696044921875, -0.02978515625, -0.0251007080078125, -0.020416259765625, -0.0157318115234375, -0.01104736328125, -0.0063629150390625, -0.001678466796875, 0.0030059814453125, 0.0076904296875, 0.0123748779296875, 0.017059326171875, 0.0217437744140625, 0.02642822265625, 0.0311126708984375, 0.035797119140625, 0.0404815673828125, 0.045166015625, 0.0498504638671875, 0.054534912109375, 0.0592193603515625, 0.06390380859375, 0.0685882568359375, 0.073272705078125, 0.0779571533203125, 0.0826416015625, 0.0873260498046875, 0.092010498046875, 0.0966949462890625, 0.10137939453125, 0.1060638427734375, 0.110748291015625, 0.1154327392578125, 0.1201171875]}, "parameters/model.layers.26.self_attn.o_proj.weight": {"_type": "histogram", "values": [4.0, 0.0, 2.0, 4.0, 1.0, 2.0, 2.0, 6.0, 6.0, 5.0, 8.0, 9.0, 15.0, 16.0, 31.0, 25.0, 30.0, 39.0, 68.0, 93.0, 156.0, 488.0, 4058.0, 36673.0, 271511.0, 1219086.0, 3438299.0, 5304529.0, 4189417.0, 1778704.0, 453617.0, 70447.0, 8373.0, 933.0, 187.0, 95.0, 66.0, 34.0, 35.0, 33.0, 15.0, 18.0, 14.0, 13.0, 10.0, 6.0, 7.0, 6.0, 5.0, 4.0, 3.0, 1.0, 3.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.2421875, -0.233428955078125, -0.22467041015625, -0.215911865234375, -0.2071533203125, -0.198394775390625, -0.18963623046875, -0.180877685546875, -0.172119140625, -0.163360595703125, -0.15460205078125, -0.145843505859375, -0.1370849609375, -0.128326416015625, -0.11956787109375, -0.110809326171875, -0.10205078125, -0.093292236328125, -0.08453369140625, -0.075775146484375, -0.0670166015625, -0.058258056640625, -0.04949951171875, -0.040740966796875, -0.031982421875, -0.023223876953125, -0.01446533203125, -0.005706787109375, 0.0030517578125, 0.011810302734375, 0.02056884765625, 0.029327392578125, 0.0380859375, 0.046844482421875, 0.05560302734375, 0.064361572265625, 0.0731201171875, 0.081878662109375, 0.09063720703125, 0.099395751953125, 0.108154296875, 0.116912841796875, 0.12567138671875, 0.134429931640625, 0.1431884765625, 0.151947021484375, 0.16070556640625, 0.169464111328125, 0.17822265625, 0.186981201171875, 0.19573974609375, 0.204498291015625, 0.2132568359375, 0.222015380859375, 0.23077392578125, 0.239532470703125, 0.248291015625, 0.257049560546875, 0.26580810546875, 0.274566650390625, 0.2833251953125, 0.292083740234375, 0.30084228515625, 0.309600830078125, 0.318359375]}, "parameters/model.layers.26.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 3.0, 1.0, 4.0, 6.0, 7.0, 12.0, 7.0, 13.0, 18.0, 18.0, 39.0, 48.0, 85.0, 151.0, 317.0, 705.0, 2233.0, 8644.0, 33875.0, 178239.0, 1247620.0, 8354644.0, 16777216.0, 16777216.0, 16777216.0, 6166392.0, 847966.0, 128257.0, 25766.0, 6413.0, 1771.0, 631.0, 277.0, 127.0, 56.0, 56.0, 24.0, 20.0, 17.0, 19.0, 16.0, 14.0, 7.0, 3.0, 2.0, 1.0, 1.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.373046875, -0.36077880859375, -0.3485107421875, -0.33624267578125, -0.323974609375, -0.31170654296875, -0.2994384765625, -0.28717041015625, -0.27490234375, -0.26263427734375, -0.2503662109375, -0.23809814453125, -0.225830078125, -0.21356201171875, -0.2012939453125, -0.18902587890625, -0.1767578125, -0.16448974609375, -0.1522216796875, -0.13995361328125, -0.127685546875, -0.11541748046875, -0.1031494140625, -0.09088134765625, -0.07861328125, -0.06634521484375, -0.0540771484375, -0.04180908203125, -0.029541015625, -0.01727294921875, -0.0050048828125, 0.00726318359375, 0.01953125, 0.03179931640625, 0.0440673828125, 0.05633544921875, 0.068603515625, 0.08087158203125, 0.0931396484375, 0.10540771484375, 0.11767578125, 0.12994384765625, 0.1422119140625, 0.15447998046875, 0.166748046875, 0.17901611328125, 0.1912841796875, 0.20355224609375, 0.2158203125, 0.22808837890625, 0.2403564453125, 0.25262451171875, 0.264892578125, 0.27716064453125, 0.2894287109375, 0.30169677734375, 0.31396484375, 0.32623291015625, 0.3385009765625, 0.35076904296875, 0.363037109375, 0.37530517578125, 0.3875732421875, 0.39984130859375, 0.412109375]}, "parameters/model.layers.26.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 2.0, 0.0, 1.0, 1.0, 5.0, 5.0, 4.0, 11.0, 25.0, 51.0, 637.0, 87263.0, 3376128.0, 16777216.0, 16777216.0, 5112869.0, 176303.0, 1361.0, 63.0, 31.0, 15.0, 9.0, 1.0, 3.0, 2.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.376953125, -0.361297607421875, -0.34564208984375, -0.329986572265625, -0.3143310546875, -0.298675537109375, -0.28302001953125, -0.267364501953125, -0.251708984375, -0.236053466796875, -0.22039794921875, -0.204742431640625, -0.1890869140625, -0.173431396484375, -0.15777587890625, -0.142120361328125, -0.12646484375, -0.110809326171875, -0.09515380859375, -0.079498291015625, -0.0638427734375, -0.048187255859375, -0.03253173828125, -0.016876220703125, -0.001220703125, 0.014434814453125, 0.03009033203125, 0.045745849609375, 0.0614013671875, 0.077056884765625, 0.09271240234375, 0.108367919921875, 0.1240234375, 0.139678955078125, 0.15533447265625, 0.170989990234375, 0.1866455078125, 0.202301025390625, 0.21795654296875, 0.233612060546875, 0.249267578125, 0.264923095703125, 0.28057861328125, 0.296234130859375, 0.3118896484375, 0.327545166015625, 0.34320068359375, 0.358856201171875, 0.37451171875, 0.390167236328125, 0.40582275390625, 0.421478271484375, 0.4371337890625, 0.452789306640625, 0.46844482421875, 0.484100341796875, 0.499755859375, 0.515411376953125, 0.53106689453125, 0.546722412109375, 0.5623779296875, 0.578033447265625, 0.59368896484375, 0.609344482421875, 0.625]}, "parameters/model.layers.26.input_layernorm.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 0.0, 1.0, 1.0, 2.0, 2.0, 4.0, 9.0, 169.0, 1412.0, 1410.0, 505.0, 135.0, 101.0, 49.0, 37.0, 32.0, 19.0, 29.0, 19.0, 18.0, 12.0, 25.0, 12.0, 13.0, 5.0, 8.0, 5.0, 8.0, 4.0, 3.0, 3.0, 2.0, 1.0, 6.0, 2.0, 2.0, 3.0, 1.0, 2.0, 2.0, 2.0, 1.0, 4.0, 1.0, 0.0, 2.0, 2.0, 1.0, 0.0, 2.0], "bins": [0.06298828125, 0.08068084716796875, 0.0983734130859375, 0.11606597900390625, 0.133758544921875, 0.15145111083984375, 0.1691436767578125, 0.18683624267578125, 0.20452880859375, 0.22222137451171875, 0.2399139404296875, 0.25760650634765625, 0.275299072265625, 0.29299163818359375, 0.3106842041015625, 0.32837677001953125, 0.3460693359375, 0.36376190185546875, 0.3814544677734375, 0.39914703369140625, 0.416839599609375, 0.43453216552734375, 0.4522247314453125, 0.46991729736328125, 0.48760986328125, 0.5053024291992188, 0.5229949951171875, 0.5406875610351562, 0.558380126953125, 0.5760726928710938, 0.5937652587890625, 0.6114578247070312, 0.629150390625, 0.6468429565429688, 0.6645355224609375, 0.6822280883789062, 0.699920654296875, 0.7176132202148438, 0.7353057861328125, 0.7529983520507812, 0.77069091796875, 0.7883834838867188, 0.8060760498046875, 0.8237686157226562, 0.841461181640625, 0.8591537475585938, 0.8768463134765625, 0.8945388793945312, 0.9122314453125, 0.9299240112304688, 0.9476165771484375, 0.9653091430664062, 0.983001708984375, 1.0006942749023438, 1.0183868408203125, 1.0360794067382812, 1.05377197265625, 1.0714645385742188, 1.0891571044921875, 1.1068496704101562, 1.124542236328125, 1.1422348022460938, 1.1599273681640625, 1.1776199340820312, 1.1953125]}, "parameters/model.layers.26.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 4.0, 3.0, 0.0, 2.0, 2.0, 3.0, 3.0, 3.0, 3.0, 1.0, 3.0, 0.0, 2.0, 3.0, 2.0, 4.0, 3.0, 2.0, 4.0, 0.0, 1.0, 2.0, 3.0, 2.0, 4.0, 2.0, 7.0, 1.0, 5.0, 16.0, 9.0, 15.0, 12.0, 6.0, 12.0, 15.0, 6.0, 13.0, 10.0, 15.0, 19.0, 19.0, 23.0, 27.0, 28.0, 66.0, 98.0, 206.0, 559.0, 1356.0, 1298.0, 185.0, 2.0, 2.0], "bins": [0.06005859375, 0.06803131103515625, 0.0760040283203125, 0.08397674560546875, 0.091949462890625, 0.09992218017578125, 0.1078948974609375, 0.11586761474609375, 0.12384033203125, 0.13181304931640625, 0.1397857666015625, 0.14775848388671875, 0.155731201171875, 0.16370391845703125, 0.1716766357421875, 0.17964935302734375, 0.1876220703125, 0.19559478759765625, 0.2035675048828125, 0.21154022216796875, 0.219512939453125, 0.22748565673828125, 0.2354583740234375, 0.24343109130859375, 0.25140380859375, 0.25937652587890625, 0.2673492431640625, 0.27532196044921875, 0.283294677734375, 0.29126739501953125, 0.2992401123046875, 0.30721282958984375, 0.315185546875, 0.32315826416015625, 0.3311309814453125, 0.33910369873046875, 0.347076416015625, 0.35504913330078125, 0.3630218505859375, 0.37099456787109375, 0.37896728515625, 0.38694000244140625, 0.3949127197265625, 0.40288543701171875, 0.410858154296875, 0.41883087158203125, 0.4268035888671875, 0.43477630615234375, 0.4427490234375, 0.45072174072265625, 0.4586944580078125, 0.46666717529296875, 0.474639892578125, 0.48261260986328125, 0.4905853271484375, 0.49855804443359375, 0.50653076171875, 0.5145034790039062, 0.5224761962890625, 0.5304489135742188, 0.538421630859375, 0.5463943481445312, 0.5543670654296875, 0.5623397827148438, 0.5703125]}, "parameters/model.layers.27.self_attn.q_proj.weight": {"_type": "histogram", "values": [3.0, 0.0, 1.0, 1.0, 3.0, 3.0, 7.0, 9.0, 9.0, 16.0, 14.0, 22.0, 36.0, 46.0, 49.0, 111.0, 202.0, 260.0, 361.0, 602.0, 1097.0, 2168.0, 5052.0, 12966.0, 34093.0, 80283.0, 179170.0, 390994.0, 801389.0, 1413337.0, 2321456.0, 3147763.0, 3132677.0, 2316665.0, 1430026.0, 790608.0, 400245.0, 179427.0, 79615.0, 33172.0, 13134.0, 5042.0, 2241.0, 1107.0, 610.0, 361.0, 242.0, 153.0, 135.0, 66.0, 44.0, 38.0, 26.0, 20.0, 19.0, 3.0, 4.0, 4.0, 3.0, 2.0, 2.0, 1.0, 0.0, 1.0], "bins": [-0.220703125, -0.21380615234375, -0.2069091796875, -0.20001220703125, -0.193115234375, -0.18621826171875, -0.1793212890625, -0.17242431640625, -0.16552734375, -0.15863037109375, -0.1517333984375, -0.14483642578125, -0.137939453125, -0.13104248046875, -0.1241455078125, -0.11724853515625, -0.1103515625, -0.10345458984375, -0.0965576171875, -0.08966064453125, -0.082763671875, -0.07586669921875, -0.0689697265625, -0.06207275390625, -0.05517578125, -0.04827880859375, -0.0413818359375, -0.03448486328125, -0.027587890625, -0.02069091796875, -0.0137939453125, -0.00689697265625, 0.0, 0.00689697265625, 0.0137939453125, 0.02069091796875, 0.027587890625, 0.03448486328125, 0.0413818359375, 0.04827880859375, 0.05517578125, 0.06207275390625, 0.0689697265625, 0.07586669921875, 0.082763671875, 0.08966064453125, 0.0965576171875, 0.10345458984375, 0.1103515625, 0.11724853515625, 0.1241455078125, 0.13104248046875, 0.137939453125, 0.14483642578125, 0.1517333984375, 0.15863037109375, 0.16552734375, 0.17242431640625, 0.1793212890625, 0.18621826171875, 0.193115234375, 0.20001220703125, 0.2069091796875, 0.21380615234375, 0.220703125]}, "parameters/model.layers.27.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 8.0, 6.0, 11.0, 9.0, 25.0, 21.0, 31.0, 50.0, 74.0, 110.0, 138.0, 215.0, 331.0, 570.0, 1070.0, 2629.0, 6054.0, 16368.0, 39345.0, 99514.0, 213778.0, 443474.0, 787668.0, 1023055.0, 764090.0, 431910.0, 202976.0, 96265.0, 38739.0, 14659.0, 5849.0, 2645.0, 1119.0, 555.0, 308.0, 190.0, 133.0, 93.0, 59.0, 44.0, 32.0, 20.0, 18.0, 10.0, 9.0, 6.0, 3.0, 2.0, 3.0, 1.0, 2.0, 0.0, 1.0, 1.0, 2.0], "bins": [-0.40234375, -0.38995361328125, -0.3775634765625, -0.36517333984375, -0.352783203125, -0.34039306640625, -0.3280029296875, -0.31561279296875, -0.30322265625, -0.29083251953125, -0.2784423828125, -0.26605224609375, -0.253662109375, -0.24127197265625, -0.2288818359375, -0.21649169921875, -0.2041015625, -0.19171142578125, -0.1793212890625, -0.16693115234375, -0.154541015625, -0.14215087890625, -0.1297607421875, -0.11737060546875, -0.10498046875, -0.09259033203125, -0.0802001953125, -0.06781005859375, -0.055419921875, -0.04302978515625, -0.0306396484375, -0.01824951171875, -0.005859375, 0.00653076171875, 0.0189208984375, 0.03131103515625, 0.043701171875, 0.05609130859375, 0.0684814453125, 0.08087158203125, 0.09326171875, 0.10565185546875, 0.1180419921875, 0.13043212890625, 0.142822265625, 0.15521240234375, 0.1676025390625, 0.17999267578125, 0.1923828125, 0.20477294921875, 0.2171630859375, 0.22955322265625, 0.241943359375, 0.25433349609375, 0.2667236328125, 0.27911376953125, 0.29150390625, 0.30389404296875, 0.3162841796875, 0.32867431640625, 0.341064453125, 0.35345458984375, 0.3658447265625, 0.37823486328125, 0.390625]}, "parameters/model.layers.27.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 2.0, 7.0, 15.0, 14.0, 32.0, 47.0, 77.0, 119.0, 225.0, 285.0, 532.0, 1011.0, 2126.0, 5409.0, 16561.0, 46673.0, 131801.0, 295699.0, 549952.0, 803925.0, 888147.0, 692744.0, 424959.0, 204351.0, 84705.0, 28169.0, 9802.0, 3606.0, 1458.0, 720.0, 402.0, 297.0, 154.0, 100.0, 62.0, 35.0, 29.0, 19.0, 11.0, 8.0, 3.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.1884765625, -0.1818084716796875, -0.175140380859375, -0.1684722900390625, -0.16180419921875, -0.1551361083984375, -0.148468017578125, -0.1417999267578125, -0.1351318359375, -0.1284637451171875, -0.121795654296875, -0.1151275634765625, -0.10845947265625, -0.1017913818359375, -0.095123291015625, -0.0884552001953125, -0.081787109375, -0.0751190185546875, -0.068450927734375, -0.0617828369140625, -0.05511474609375, -0.0484466552734375, -0.041778564453125, -0.0351104736328125, -0.0284423828125, -0.0217742919921875, -0.015106201171875, -0.0084381103515625, -0.00177001953125, 0.0048980712890625, 0.011566162109375, 0.0182342529296875, 0.02490234375, 0.0315704345703125, 0.038238525390625, 0.0449066162109375, 0.05157470703125, 0.0582427978515625, 0.064910888671875, 0.0715789794921875, 0.0782470703125, 0.0849151611328125, 0.091583251953125, 0.0982513427734375, 0.10491943359375, 0.1115875244140625, 0.118255615234375, 0.1249237060546875, 0.131591796875, 0.1382598876953125, 0.144927978515625, 0.1515960693359375, 0.15826416015625, 0.1649322509765625, 0.171600341796875, 0.1782684326171875, 0.1849365234375, 0.1916046142578125, 0.198272705078125, 0.2049407958984375, 0.21160888671875, 0.2182769775390625, 0.224945068359375, 0.2316131591796875, 0.23828125]}, "parameters/model.layers.27.self_attn.o_proj.weight": {"_type": "histogram", "values": [2.0, 4.0, 7.0, 9.0, 4.0, 7.0, 10.0, 13.0, 13.0, 14.0, 11.0, 10.0, 13.0, 17.0, 28.0, 30.0, 39.0, 57.0, 93.0, 211.0, 1304.0, 21745.0, 292358.0, 1954937.0, 5708648.0, 6041239.0, 2341816.0, 380895.0, 31328.0, 1740.0, 246.0, 95.0, 63.0, 28.0, 27.0, 27.0, 14.0, 10.0, 12.0, 11.0, 11.0, 12.0, 13.0, 8.0, 10.0, 9.0, 3.0, 4.0, 6.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.2890625, -0.27752685546875, -0.2659912109375, -0.25445556640625, -0.242919921875, -0.23138427734375, -0.2198486328125, -0.20831298828125, -0.19677734375, -0.18524169921875, -0.1737060546875, -0.16217041015625, -0.150634765625, -0.13909912109375, -0.1275634765625, -0.11602783203125, -0.1044921875, -0.09295654296875, -0.0814208984375, -0.06988525390625, -0.058349609375, -0.04681396484375, -0.0352783203125, -0.02374267578125, -0.01220703125, -0.00067138671875, 0.0108642578125, 0.02239990234375, 0.033935546875, 0.04547119140625, 0.0570068359375, 0.06854248046875, 0.080078125, 0.09161376953125, 0.1031494140625, 0.11468505859375, 0.126220703125, 0.13775634765625, 0.1492919921875, 0.16082763671875, 0.17236328125, 0.18389892578125, 0.1954345703125, 0.20697021484375, 0.218505859375, 0.23004150390625, 0.2415771484375, 0.25311279296875, 0.2646484375, 0.27618408203125, 0.2877197265625, 0.29925537109375, 0.310791015625, 0.32232666015625, 0.3338623046875, 0.34539794921875, 0.35693359375, 0.36846923828125, 0.3800048828125, 0.39154052734375, 0.403076171875, 0.41461181640625, 0.4261474609375, 0.43768310546875, 0.44921875]}, "parameters/model.layers.27.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 3.0, 0.0, 1.0, 2.0, 0.0, 9.0, 2.0, 8.0, 7.0, 16.0, 21.0, 26.0, 55.0, 75.0, 118.0, 250.0, 672.0, 1906.0, 6063.0, 21443.0, 80795.0, 376492.0, 2268130.0, 11320817.0, 16777216.0, 16777216.0, 16777216.0, 6355820.0, 1106341.0, 196543.0, 47022.0, 12540.0, 3707.0, 1102.0, 396.0, 178.0, 96.0, 58.0, 40.0, 29.0, 20.0, 10.0, 8.0, 2.0, 4.0, 3.0, 2.0, 1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.375, -0.363739013671875, -0.35247802734375, -0.341217041015625, -0.3299560546875, -0.318695068359375, -0.30743408203125, -0.296173095703125, -0.284912109375, -0.273651123046875, -0.26239013671875, -0.251129150390625, -0.2398681640625, -0.228607177734375, -0.21734619140625, -0.206085205078125, -0.19482421875, -0.183563232421875, -0.17230224609375, -0.161041259765625, -0.1497802734375, -0.138519287109375, -0.12725830078125, -0.115997314453125, -0.104736328125, -0.093475341796875, -0.08221435546875, -0.070953369140625, -0.0596923828125, -0.048431396484375, -0.03717041015625, -0.025909423828125, -0.0146484375, -0.003387451171875, 0.00787353515625, 0.019134521484375, 0.0303955078125, 0.041656494140625, 0.05291748046875, 0.064178466796875, 0.075439453125, 0.086700439453125, 0.09796142578125, 0.109222412109375, 0.1204833984375, 0.131744384765625, 0.14300537109375, 0.154266357421875, 0.16552734375, 0.176788330078125, 0.18804931640625, 0.199310302734375, 0.2105712890625, 0.221832275390625, 0.23309326171875, 0.244354248046875, 0.255615234375, 0.266876220703125, 0.27813720703125, 0.289398193359375, 0.3006591796875, 0.311920166015625, 0.32318115234375, 0.334442138671875, 0.345703125]}, "parameters/model.layers.27.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 3.0, 0.0, 1.0, 1.0, 3.0, 0.0, 2.0, 1.0, 6.0, 6.0, 4.0, 16.0, 27.0, 54.0, 157.0, 5404.0, 194823.0, 2937152.0, 15070121.0, 16777216.0, 12842124.0, 2209032.0, 125194.0, 2996.0, 117.0, 71.0, 18.0, 14.0, 7.0, 5.0, 3.0, 4.0, 0.0, 3.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.47265625, -0.460357666015625, -0.44805908203125, -0.435760498046875, -0.4234619140625, -0.411163330078125, -0.39886474609375, -0.386566162109375, -0.374267578125, -0.361968994140625, -0.34967041015625, -0.337371826171875, -0.3250732421875, -0.312774658203125, -0.30047607421875, -0.288177490234375, -0.27587890625, -0.263580322265625, -0.25128173828125, -0.238983154296875, -0.2266845703125, -0.214385986328125, -0.20208740234375, -0.189788818359375, -0.177490234375, -0.165191650390625, -0.15289306640625, -0.140594482421875, -0.1282958984375, -0.115997314453125, -0.10369873046875, -0.091400146484375, -0.0791015625, -0.066802978515625, -0.05450439453125, -0.042205810546875, -0.0299072265625, -0.017608642578125, -0.00531005859375, 0.006988525390625, 0.019287109375, 0.031585693359375, 0.04388427734375, 0.056182861328125, 0.0684814453125, 0.080780029296875, 0.09307861328125, 0.105377197265625, 0.11767578125, 0.129974365234375, 0.14227294921875, 0.154571533203125, 0.1668701171875, 0.179168701171875, 0.19146728515625, 0.203765869140625, 0.216064453125, 0.228363037109375, 0.24066162109375, 0.252960205078125, 0.2652587890625, 0.277557373046875, 0.28985595703125, 0.302154541015625, 0.314453125]}, "parameters/model.layers.27.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 2.0, 1.0, 3.0, 2.0, 5.0, 7.0, 23.0, 529.0, 1733.0, 837.0, 344.0, 161.0, 92.0, 45.0, 37.0, 36.0, 24.0, 13.0, 25.0, 17.0, 18.0, 21.0, 17.0, 8.0, 10.0, 15.0, 7.0, 8.0, 6.0, 6.0, 3.0, 3.0, 5.0, 4.0, 6.0, 4.0, 1.0, 3.0, 1.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.000957489013671875, 0.018710792064666748, 0.03837907314300537, 0.058047354221343994, 0.07771563529968262, 0.09738391637802124, 0.11705219745635986, 0.1367204785346985, 0.1563887596130371, 0.17605704069137573, 0.19572532176971436, 0.21539360284805298, 0.2350618839263916, 0.2547301650047302, 0.27439844608306885, 0.29406672716140747, 0.3137350082397461, 0.3334032893180847, 0.35307157039642334, 0.37273985147476196, 0.3924081325531006, 0.4120764136314392, 0.43174469470977783, 0.45141297578811646, 0.4710812568664551, 0.4907495379447937, 0.5104178190231323, 0.530086100101471, 0.5497543811798096, 0.5694226622581482, 0.5890909433364868, 0.6087592244148254, 0.6284275054931641, 0.6480957865715027, 0.6677640676498413, 0.6874323487281799, 0.7071006298065186, 0.7267689108848572, 0.7464371919631958, 0.7661054730415344, 0.785773754119873, 0.8054420351982117, 0.8251103162765503, 0.8447785973548889, 0.8644468784332275, 0.8841151595115662, 0.9037834405899048, 0.9234517216682434, 0.943120002746582, 0.9627882838249207, 0.9824565649032593, 1.0021247863769531, 1.0217931270599365, 1.04146146774292, 1.0611296892166138, 1.0807979106903076, 1.100466251373291, 1.1201345920562744, 1.1398028135299683, 1.159471035003662, 1.1791393756866455, 1.198807716369629, 1.2184759378433228, 1.2381441593170166, 1.2578125]}, "parameters/model.layers.27.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 3.0, 3.0, 3.0, 1.0, 1.0, 2.0, 2.0, 5.0, 2.0, 3.0, 0.0, 1.0, 5.0, 1.0, 1.0, 3.0, 2.0, 3.0, 2.0, 1.0, 3.0, 4.0, 1.0, 1.0, 1.0, 3.0, 2.0, 5.0, 4.0, 6.0, 1.0, 10.0, 15.0, 12.0, 15.0, 7.0, 5.0, 15.0, 12.0, 7.0, 15.0, 14.0, 18.0, 24.0, 21.0, 31.0, 32.0, 67.0, 113.0, 364.0, 1129.0, 1498.0, 569.0, 22.0, 3.0], "bins": [0.07177734375, 0.07987213134765625, 0.0879669189453125, 0.09606170654296875, 0.104156494140625, 0.11225128173828125, 0.1203460693359375, 0.12844085693359375, 0.13653564453125, 0.14463043212890625, 0.1527252197265625, 0.16082000732421875, 0.168914794921875, 0.17700958251953125, 0.1851043701171875, 0.19319915771484375, 0.2012939453125, 0.20938873291015625, 0.2174835205078125, 0.22557830810546875, 0.233673095703125, 0.24176788330078125, 0.2498626708984375, 0.25795745849609375, 0.26605224609375, 0.27414703369140625, 0.2822418212890625, 0.29033660888671875, 0.298431396484375, 0.30652618408203125, 0.3146209716796875, 0.32271575927734375, 0.330810546875, 0.33890533447265625, 0.3470001220703125, 0.35509490966796875, 0.363189697265625, 0.37128448486328125, 0.3793792724609375, 0.38747406005859375, 0.39556884765625, 0.40366363525390625, 0.4117584228515625, 0.41985321044921875, 0.427947998046875, 0.43604278564453125, 0.4441375732421875, 0.45223236083984375, 0.4603271484375, 0.46842193603515625, 0.4765167236328125, 0.48461151123046875, 0.492706298828125, 0.5008010864257812, 0.5088958740234375, 0.5169906616210938, 0.52508544921875, 0.5331802368164062, 0.5412750244140625, 0.5493698120117188, 0.557464599609375, 0.5655593872070312, 0.5736541748046875, 0.5817489624023438, 0.58984375]}, "parameters/model.layers.28.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 0.0, 6.0, 8.0, 10.0, 15.0, 16.0, 26.0, 25.0, 45.0, 74.0, 113.0, 188.0, 361.0, 680.0, 1597.0, 4187.0, 13887.0, 42626.0, 131421.0, 347705.0, 873074.0, 1790603.0, 3072937.0, 3937444.0, 3155139.0, 1876248.0, 935503.0, 378636.0, 143639.0, 47605.0, 15202.0, 4651.0, 1843.0, 679.0, 404.0, 193.0, 141.0, 81.0, 59.0, 48.0, 38.0, 12.0, 7.0, 11.0, 5.0, 5.0, 5.0, 3.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0], "bins": [-0.265625, -0.2572021484375, -0.248779296875, -0.2403564453125, -0.23193359375, -0.2235107421875, -0.215087890625, -0.2066650390625, -0.1982421875, -0.1898193359375, -0.181396484375, -0.1729736328125, -0.16455078125, -0.1561279296875, -0.147705078125, -0.1392822265625, -0.130859375, -0.1224365234375, -0.114013671875, -0.1055908203125, -0.09716796875, -0.0887451171875, -0.080322265625, -0.0718994140625, -0.0634765625, -0.0550537109375, -0.046630859375, -0.0382080078125, -0.02978515625, -0.0213623046875, -0.012939453125, -0.0045166015625, 0.00390625, 0.0123291015625, 0.020751953125, 0.0291748046875, 0.03759765625, 0.0460205078125, 0.054443359375, 0.0628662109375, 0.0712890625, 0.0797119140625, 0.088134765625, 0.0965576171875, 0.10498046875, 0.1134033203125, 0.121826171875, 0.1302490234375, 0.138671875, 0.1470947265625, 0.155517578125, 0.1639404296875, 0.17236328125, 0.1807861328125, 0.189208984375, 0.1976318359375, 0.2060546875, 0.2144775390625, 0.222900390625, 0.2313232421875, 0.23974609375, 0.2481689453125, 0.256591796875, 0.2650146484375, 0.2734375]}, "parameters/model.layers.28.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 1.0, 3.0, 2.0, 2.0, 9.0, 7.0, 11.0, 24.0, 16.0, 31.0, 36.0, 68.0, 91.0, 113.0, 175.0, 228.0, 320.0, 534.0, 878.0, 1873.0, 3728.0, 8302.0, 18962.0, 41254.0, 83464.0, 155022.0, 281784.0, 431125.0, 605198.0, 696365.0, 646425.0, 494384.0, 329703.0, 191794.0, 103659.0, 52809.0, 25032.0, 10681.0, 4896.0, 2213.0, 1143.0, 654.0, 359.0, 290.0, 172.0, 127.0, 87.0, 65.0, 51.0, 45.0, 25.0, 20.0, 10.0, 9.0, 8.0, 4.0, 2.0, 1.0, 2.0, 1.0, 3.0, 1.0], "bins": [-0.279296875, -0.270477294921875, -0.26165771484375, -0.252838134765625, -0.2440185546875, -0.235198974609375, -0.22637939453125, -0.217559814453125, -0.208740234375, -0.199920654296875, -0.19110107421875, -0.182281494140625, -0.1734619140625, -0.164642333984375, -0.15582275390625, -0.147003173828125, -0.13818359375, -0.129364013671875, -0.12054443359375, -0.111724853515625, -0.1029052734375, -0.094085693359375, -0.08526611328125, -0.076446533203125, -0.067626953125, -0.058807373046875, -0.04998779296875, -0.041168212890625, -0.0323486328125, -0.023529052734375, -0.01470947265625, -0.005889892578125, 0.0029296875, 0.011749267578125, 0.02056884765625, 0.029388427734375, 0.0382080078125, 0.047027587890625, 0.05584716796875, 0.064666748046875, 0.073486328125, 0.082305908203125, 0.09112548828125, 0.099945068359375, 0.1087646484375, 0.117584228515625, 0.12640380859375, 0.135223388671875, 0.14404296875, 0.152862548828125, 0.16168212890625, 0.170501708984375, 0.1793212890625, 0.188140869140625, 0.19696044921875, 0.205780029296875, 0.214599609375, 0.223419189453125, 0.23223876953125, 0.241058349609375, 0.2498779296875, 0.258697509765625, 0.26751708984375, 0.276336669921875, 0.28515625]}, "parameters/model.layers.28.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0, 0.0, 2.0, 2.0, 5.0, 5.0, 7.0, 16.0, 15.0, 48.0, 32.0, 55.0, 109.0, 145.0, 233.0, 533.0, 1066.0, 2364.0, 5238.0, 10866.0, 21814.0, 42340.0, 80632.0, 137356.0, 215178.0, 315535.0, 424136.0, 515424.0, 574380.0, 524523.0, 434854.0, 334431.0, 229366.0, 141680.0, 87695.0, 47697.0, 23845.0, 11934.0, 5655.0, 2595.0, 1231.0, 516.0, 294.0, 163.0, 99.0, 56.0, 35.0, 29.0, 20.0, 12.0, 12.0, 8.0, 3.0, 4.0, 3.0, 1.0, 0.0, 3.0], "bins": [-0.1513671875, -0.1469879150390625, -0.142608642578125, -0.1382293701171875, -0.13385009765625, -0.1294708251953125, -0.125091552734375, -0.1207122802734375, -0.1163330078125, -0.1119537353515625, -0.107574462890625, -0.1031951904296875, -0.09881591796875, -0.0944366455078125, -0.090057373046875, -0.0856781005859375, -0.081298828125, -0.0769195556640625, -0.072540283203125, -0.0681610107421875, -0.06378173828125, -0.0594024658203125, -0.055023193359375, -0.0506439208984375, -0.0462646484375, -0.0418853759765625, -0.037506103515625, -0.0331268310546875, -0.02874755859375, -0.0243682861328125, -0.019989013671875, -0.0156097412109375, -0.01123046875, -0.0068511962890625, -0.002471923828125, 0.0019073486328125, 0.00628662109375, 0.0106658935546875, 0.015045166015625, 0.0194244384765625, 0.0238037109375, 0.0281829833984375, 0.032562255859375, 0.0369415283203125, 0.04132080078125, 0.0457000732421875, 0.050079345703125, 0.0544586181640625, 0.058837890625, 0.0632171630859375, 0.067596435546875, 0.0719757080078125, 0.07635498046875, 0.0807342529296875, 0.085113525390625, 0.0894927978515625, 0.0938720703125, 0.0982513427734375, 0.102630615234375, 0.1070098876953125, 0.11138916015625, 0.1157684326171875, 0.120147705078125, 0.1245269775390625, 0.12890625]}, "parameters/model.layers.28.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 2.0, 2.0, 1.0, 3.0, 5.0, 2.0, 2.0, 4.0, 5.0, 3.0, 6.0, 7.0, 8.0, 13.0, 19.0, 26.0, 63.0, 134.0, 384.0, 11100.0, 363321.0, 3982012.0, 8955447.0, 3206453.0, 250421.0, 7196.0, 268.0, 113.0, 60.0, 40.0, 13.0, 16.0, 9.0, 2.0, 7.0, 7.0, 8.0, 5.0, 0.0, 3.0, 4.0, 3.0, 6.0, 4.0, 2.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.482421875, -0.466033935546875, -0.44964599609375, -0.433258056640625, -0.4168701171875, -0.400482177734375, -0.38409423828125, -0.367706298828125, -0.351318359375, -0.334930419921875, -0.31854248046875, -0.302154541015625, -0.2857666015625, -0.269378662109375, -0.25299072265625, -0.236602783203125, -0.22021484375, -0.203826904296875, -0.18743896484375, -0.171051025390625, -0.1546630859375, -0.138275146484375, -0.12188720703125, -0.105499267578125, -0.089111328125, -0.072723388671875, -0.05633544921875, -0.039947509765625, -0.0235595703125, -0.007171630859375, 0.00921630859375, 0.025604248046875, 0.0419921875, 0.058380126953125, 0.07476806640625, 0.091156005859375, 0.1075439453125, 0.123931884765625, 0.14031982421875, 0.156707763671875, 0.173095703125, 0.189483642578125, 0.20587158203125, 0.222259521484375, 0.2386474609375, 0.255035400390625, 0.27142333984375, 0.287811279296875, 0.30419921875, 0.320587158203125, 0.33697509765625, 0.353363037109375, 0.3697509765625, 0.386138916015625, 0.40252685546875, 0.418914794921875, 0.435302734375, 0.451690673828125, 0.46807861328125, 0.484466552734375, 0.5008544921875, 0.517242431640625, 0.53363037109375, 0.550018310546875, 0.56640625]}, "parameters/model.layers.28.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 3.0, 5.0, 10.0, 9.0, 19.0, 14.0, 30.0, 35.0, 70.0, 124.0, 288.0, 733.0, 3059.0, 13649.0, 62418.0, 385684.0, 3378126.0, 16777216.0, 16777216.0, 16777216.0, 7719137.0, 896600.0, 122861.0, 24754.0, 5066.0, 1314.0, 364.0, 153.0, 111.0, 56.0, 27.0, 26.0, 12.0, 7.0, 6.0, 6.0, 6.0, 3.0, 1.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.416015625, -0.402008056640625, -0.38800048828125, -0.373992919921875, -0.3599853515625, -0.345977783203125, -0.33197021484375, -0.317962646484375, -0.303955078125, -0.289947509765625, -0.27593994140625, -0.261932373046875, -0.2479248046875, -0.233917236328125, -0.21990966796875, -0.205902099609375, -0.19189453125, -0.177886962890625, -0.16387939453125, -0.149871826171875, -0.1358642578125, -0.121856689453125, -0.10784912109375, -0.093841552734375, -0.079833984375, -0.065826416015625, -0.05181884765625, -0.037811279296875, -0.0238037109375, -0.009796142578125, 0.00421142578125, 0.018218994140625, 0.0322265625, 0.046234130859375, 0.06024169921875, 0.074249267578125, 0.0882568359375, 0.102264404296875, 0.11627197265625, 0.130279541015625, 0.144287109375, 0.158294677734375, 0.17230224609375, 0.186309814453125, 0.2003173828125, 0.214324951171875, 0.22833251953125, 0.242340087890625, 0.25634765625, 0.270355224609375, 0.28436279296875, 0.298370361328125, 0.3123779296875, 0.326385498046875, 0.34039306640625, 0.354400634765625, 0.368408203125, 0.382415771484375, 0.39642333984375, 0.410430908203125, 0.4244384765625, 0.438446044921875, 0.45245361328125, 0.466461181640625, 0.48046875]}, "parameters/model.layers.28.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 3.0, 0.0, 1.0, 1.0, 4.0, 4.0, 10.0, 27.0, 93.0, 952.0, 175283.0, 6484490.0, 16777216.0, 16777216.0, 1398053.0, 14669.0, 228.0, 64.0, 22.0, 12.0, 5.0, 2.0, 1.0, 1.0, 3.0, 0.0, 1.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.6171875, -0.59942626953125, -0.5816650390625, -0.56390380859375, -0.546142578125, -0.52838134765625, -0.5106201171875, -0.49285888671875, -0.47509765625, -0.45733642578125, -0.4395751953125, -0.42181396484375, -0.404052734375, -0.38629150390625, -0.3685302734375, -0.35076904296875, -0.3330078125, -0.31524658203125, -0.2974853515625, -0.27972412109375, -0.261962890625, -0.24420166015625, -0.2264404296875, -0.20867919921875, -0.19091796875, -0.17315673828125, -0.1553955078125, -0.13763427734375, -0.119873046875, -0.10211181640625, -0.0843505859375, -0.06658935546875, -0.048828125, -0.03106689453125, -0.0133056640625, 0.00445556640625, 0.022216796875, 0.03997802734375, 0.0577392578125, 0.07550048828125, 0.09326171875, 0.11102294921875, 0.1287841796875, 0.14654541015625, 0.164306640625, 0.18206787109375, 0.1998291015625, 0.21759033203125, 0.2353515625, 0.25311279296875, 0.2708740234375, 0.28863525390625, 0.306396484375, 0.32415771484375, 0.3419189453125, 0.35968017578125, 0.37744140625, 0.39520263671875, 0.4129638671875, 0.43072509765625, 0.448486328125, 0.46624755859375, 0.4840087890625, 0.50177001953125, 0.51953125]}, "parameters/model.layers.28.input_layernorm.weight": {"_type": "histogram", "values": [2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 4.0, 1.0, 2.0, 11.0, 6.0, 26.0, 492.0, 2301.0, 667.0, 183.0, 81.0, 46.0, 34.0, 18.0, 29.0, 24.0, 18.0, 20.0, 17.0, 17.0, 8.0, 10.0, 9.0, 8.0, 3.0, 4.0, 4.0, 4.0, 4.0, 1.0, 2.0, 2.0, 5.0, 1.0, 6.0, 1.0, 3.0, 4.0, 2.0, 1.0, 2.0, 1.0, 3.0, 1.0, 3.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.10009765625, 0.11867523193359375, 0.1372528076171875, 0.15583038330078125, 0.174407958984375, 0.19298553466796875, 0.2115631103515625, 0.23014068603515625, 0.24871826171875, 0.26729583740234375, 0.2858734130859375, 0.30445098876953125, 0.323028564453125, 0.34160614013671875, 0.3601837158203125, 0.37876129150390625, 0.3973388671875, 0.41591644287109375, 0.4344940185546875, 0.45307159423828125, 0.471649169921875, 0.49022674560546875, 0.5088043212890625, 0.5273818969726562, 0.54595947265625, 0.5645370483398438, 0.5831146240234375, 0.6016921997070312, 0.620269775390625, 0.6388473510742188, 0.6574249267578125, 0.6760025024414062, 0.694580078125, 0.7131576538085938, 0.7317352294921875, 0.7503128051757812, 0.768890380859375, 0.7874679565429688, 0.8060455322265625, 0.8246231079101562, 0.84320068359375, 0.8617782592773438, 0.8803558349609375, 0.8989334106445312, 0.917510986328125, 0.9360885620117188, 0.9546661376953125, 0.9732437133789062, 0.9918212890625, 1.0103988647460938, 1.0289764404296875, 1.0475540161132812, 1.066131591796875, 1.0847091674804688, 1.1032867431640625, 1.1218643188476562, 1.14044189453125, 1.1590194702148438, 1.1775970458984375, 1.1961746215820312, 1.214752197265625, 1.2333297729492188, 1.2519073486328125, 1.2704849243164062, 1.2890625]}, "parameters/model.layers.28.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 1.0, 3.0, 1.0, 1.0, 2.0, 2.0, 3.0, 2.0, 2.0, 2.0, 5.0, 1.0, 2.0, 0.0, 0.0, 2.0, 5.0, 1.0, 1.0, 3.0, 2.0, 1.0, 3.0, 2.0, 1.0, 2.0, 1.0, 2.0, 6.0, 5.0, 1.0, 6.0, 4.0, 12.0, 14.0, 14.0, 20.0, 16.0, 14.0, 14.0, 21.0, 18.0, 23.0, 38.0, 68.0, 105.0, 883.0, 1632.0, 1008.0, 98.0, 14.0, 3.0, 1.0], "bins": [0.083984375, 0.092437744140625, 0.10089111328125, 0.109344482421875, 0.1177978515625, 0.126251220703125, 0.13470458984375, 0.143157958984375, 0.151611328125, 0.160064697265625, 0.16851806640625, 0.176971435546875, 0.1854248046875, 0.193878173828125, 0.20233154296875, 0.210784912109375, 0.21923828125, 0.227691650390625, 0.23614501953125, 0.244598388671875, 0.2530517578125, 0.261505126953125, 0.26995849609375, 0.278411865234375, 0.286865234375, 0.295318603515625, 0.30377197265625, 0.312225341796875, 0.3206787109375, 0.329132080078125, 0.33758544921875, 0.346038818359375, 0.3544921875, 0.362945556640625, 0.37139892578125, 0.379852294921875, 0.3883056640625, 0.396759033203125, 0.40521240234375, 0.413665771484375, 0.422119140625, 0.430572509765625, 0.43902587890625, 0.447479248046875, 0.4559326171875, 0.464385986328125, 0.47283935546875, 0.481292724609375, 0.48974609375, 0.498199462890625, 0.50665283203125, 0.515106201171875, 0.5235595703125, 0.532012939453125, 0.54046630859375, 0.548919677734375, 0.557373046875, 0.565826416015625, 0.57427978515625, 0.582733154296875, 0.5911865234375, 0.599639892578125, 0.60809326171875, 0.616546630859375, 0.625]}, "parameters/model.layers.29.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 2.0, 3.0, 4.0, 10.0, 6.0, 6.0, 17.0, 21.0, 40.0, 74.0, 86.0, 143.0, 266.0, 381.0, 650.0, 1246.0, 2476.0, 6493.0, 20526.0, 64898.0, 190062.0, 523472.0, 1228703.0, 2434885.0, 3685184.0, 3754943.0, 2591145.0, 1351001.0, 588867.0, 218419.0, 75486.0, 24194.0, 7515.0, 2870.0, 1305.0, 733.0, 370.0, 279.0, 156.0, 84.0, 61.0, 38.0, 32.0, 13.0, 15.0, 5.0, 12.0, 7.0, 1.0, 2.0, 2.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.291015625, -0.282470703125, -0.27392578125, -0.265380859375, -0.2568359375, -0.248291015625, -0.23974609375, -0.231201171875, -0.22265625, -0.214111328125, -0.20556640625, -0.197021484375, -0.1884765625, -0.179931640625, -0.17138671875, -0.162841796875, -0.154296875, -0.145751953125, -0.13720703125, -0.128662109375, -0.1201171875, -0.111572265625, -0.10302734375, -0.094482421875, -0.0859375, -0.077392578125, -0.06884765625, -0.060302734375, -0.0517578125, -0.043212890625, -0.03466796875, -0.026123046875, -0.017578125, -0.009033203125, -0.00048828125, 0.008056640625, 0.0166015625, 0.025146484375, 0.03369140625, 0.042236328125, 0.05078125, 0.059326171875, 0.06787109375, 0.076416015625, 0.0849609375, 0.093505859375, 0.10205078125, 0.110595703125, 0.119140625, 0.127685546875, 0.13623046875, 0.144775390625, 0.1533203125, 0.161865234375, 0.17041015625, 0.178955078125, 0.1875, 0.196044921875, 0.20458984375, 0.213134765625, 0.2216796875, 0.230224609375, 0.23876953125, 0.247314453125, 0.255859375]}, "parameters/model.layers.29.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 4.0, 5.0, 4.0, 6.0, 9.0, 16.0, 22.0, 34.0, 40.0, 79.0, 77.0, 154.0, 220.0, 400.0, 1002.0, 2351.0, 6021.0, 15569.0, 40344.0, 98715.0, 208794.0, 429125.0, 694438.0, 886989.0, 787879.0, 521903.0, 277157.0, 130811.0, 55297.0, 22014.0, 8687.0, 3431.0, 1294.0, 605.0, 290.0, 139.0, 108.0, 60.0, 66.0, 33.0, 28.0, 28.0, 11.0, 14.0, 11.0, 6.0, 2.0, 3.0, 2.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.404296875, -0.39263916015625, -0.3809814453125, -0.36932373046875, -0.357666015625, -0.34600830078125, -0.3343505859375, -0.32269287109375, -0.31103515625, -0.29937744140625, -0.2877197265625, -0.27606201171875, -0.264404296875, -0.25274658203125, -0.2410888671875, -0.22943115234375, -0.2177734375, -0.20611572265625, -0.1944580078125, -0.18280029296875, -0.171142578125, -0.15948486328125, -0.1478271484375, -0.13616943359375, -0.12451171875, -0.11285400390625, -0.1011962890625, -0.08953857421875, -0.077880859375, -0.06622314453125, -0.0545654296875, -0.04290771484375, -0.03125, -0.01959228515625, -0.0079345703125, 0.00372314453125, 0.015380859375, 0.02703857421875, 0.0386962890625, 0.05035400390625, 0.06201171875, 0.07366943359375, 0.0853271484375, 0.09698486328125, 0.108642578125, 0.12030029296875, 0.1319580078125, 0.14361572265625, 0.1552734375, 0.16693115234375, 0.1785888671875, 0.19024658203125, 0.201904296875, 0.21356201171875, 0.2252197265625, 0.23687744140625, 0.24853515625, 0.26019287109375, 0.2718505859375, 0.28350830078125, 0.295166015625, 0.30682373046875, 0.3184814453125, 0.33013916015625, 0.341796875]}, "parameters/model.layers.29.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 2.0, 3.0, 2.0, 6.0, 3.0, 9.0, 15.0, 20.0, 29.0, 47.0, 57.0, 72.0, 111.0, 146.0, 215.0, 362.0, 854.0, 2322.0, 6794.0, 19865.0, 53595.0, 131167.0, 260380.0, 458482.0, 648792.0, 783742.0, 706725.0, 522149.0, 314387.0, 168823.0, 71483.0, 28407.0, 9581.0, 3188.0, 1109.0, 493.0, 267.0, 178.0, 126.0, 81.0, 63.0, 51.0, 31.0, 19.0, 16.0, 5.0, 10.0, 5.0, 5.0, 3.0, 2.0, 0.0, 1.0], "bins": [-0.234375, -0.2279815673828125, -0.221588134765625, -0.2151947021484375, -0.20880126953125, -0.2024078369140625, -0.196014404296875, -0.1896209716796875, -0.1832275390625, -0.1768341064453125, -0.170440673828125, -0.1640472412109375, -0.15765380859375, -0.1512603759765625, -0.144866943359375, -0.1384735107421875, -0.132080078125, -0.1256866455078125, -0.119293212890625, -0.1128997802734375, -0.10650634765625, -0.1001129150390625, -0.093719482421875, -0.0873260498046875, -0.0809326171875, -0.0745391845703125, -0.068145751953125, -0.0617523193359375, -0.05535888671875, -0.0489654541015625, -0.042572021484375, -0.0361785888671875, -0.02978515625, -0.0233917236328125, -0.016998291015625, -0.0106048583984375, -0.00421142578125, 0.0021820068359375, 0.008575439453125, 0.0149688720703125, 0.0213623046875, 0.0277557373046875, 0.034149169921875, 0.0405426025390625, 0.04693603515625, 0.0533294677734375, 0.059722900390625, 0.0661163330078125, 0.072509765625, 0.0789031982421875, 0.085296630859375, 0.0916900634765625, 0.09808349609375, 0.1044769287109375, 0.110870361328125, 0.1172637939453125, 0.1236572265625, 0.1300506591796875, 0.136444091796875, 0.1428375244140625, 0.14923095703125, 0.1556243896484375, 0.162017822265625, 0.1684112548828125, 0.1748046875]}, "parameters/model.layers.29.self_attn.o_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 0.0, 1.0, 1.0, 2.0, 0.0, 1.0, 2.0, 6.0, 5.0, 6.0, 6.0, 8.0, 15.0, 24.0, 20.0, 39.0, 64.0, 56.0, 88.0, 227.0, 1530.0, 15425.0, 136757.0, 812246.0, 2813659.0, 5311814.0, 4824129.0, 2200675.0, 564967.0, 84713.0, 9346.0, 888.0, 170.0, 81.0, 51.0, 42.0, 36.0, 24.0, 16.0, 20.0, 11.0, 11.0, 3.0, 9.0, 4.0, 3.0, 2.0, 3.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.2890625, -0.279052734375, -0.26904296875, -0.259033203125, -0.2490234375, -0.239013671875, -0.22900390625, -0.218994140625, -0.208984375, -0.198974609375, -0.18896484375, -0.178955078125, -0.1689453125, -0.158935546875, -0.14892578125, -0.138916015625, -0.12890625, -0.118896484375, -0.10888671875, -0.098876953125, -0.0888671875, -0.078857421875, -0.06884765625, -0.058837890625, -0.048828125, -0.038818359375, -0.02880859375, -0.018798828125, -0.0087890625, 0.001220703125, 0.01123046875, 0.021240234375, 0.03125, 0.041259765625, 0.05126953125, 0.061279296875, 0.0712890625, 0.081298828125, 0.09130859375, 0.101318359375, 0.111328125, 0.121337890625, 0.13134765625, 0.141357421875, 0.1513671875, 0.161376953125, 0.17138671875, 0.181396484375, 0.19140625, 0.201416015625, 0.21142578125, 0.221435546875, 0.2314453125, 0.241455078125, 0.25146484375, 0.261474609375, 0.271484375, 0.281494140625, 0.29150390625, 0.301513671875, 0.3115234375, 0.321533203125, 0.33154296875, 0.341552734375, 0.3515625]}, "parameters/model.layers.29.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 3.0, 2.0, 3.0, 8.0, 5.0, 6.0, 14.0, 16.0, 21.0, 29.0, 49.0, 91.0, 202.0, 419.0, 1480.0, 5584.0, 22370.0, 101963.0, 591655.0, 4967750.0, 16777216.0, 16777216.0, 16777216.0, 6696739.0, 783300.0, 130669.0, 29133.0, 6738.0, 1629.0, 492.0, 210.0, 98.0, 71.0, 47.0, 17.0, 21.0, 13.0, 10.0, 3.0, 1.0, 2.0, 0.0, 5.0, 1.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.443359375, -0.42974853515625, -0.4161376953125, -0.40252685546875, -0.388916015625, -0.37530517578125, -0.3616943359375, -0.34808349609375, -0.33447265625, -0.32086181640625, -0.3072509765625, -0.29364013671875, -0.280029296875, -0.26641845703125, -0.2528076171875, -0.23919677734375, -0.2255859375, -0.21197509765625, -0.1983642578125, -0.18475341796875, -0.171142578125, -0.15753173828125, -0.1439208984375, -0.13031005859375, -0.11669921875, -0.10308837890625, -0.0894775390625, -0.07586669921875, -0.062255859375, -0.04864501953125, -0.0350341796875, -0.02142333984375, -0.0078125, 0.00579833984375, 0.0194091796875, 0.03302001953125, 0.046630859375, 0.06024169921875, 0.0738525390625, 0.08746337890625, 0.10107421875, 0.11468505859375, 0.1282958984375, 0.14190673828125, 0.155517578125, 0.16912841796875, 0.1827392578125, 0.19635009765625, 0.2099609375, 0.22357177734375, 0.2371826171875, 0.25079345703125, 0.264404296875, 0.27801513671875, 0.2916259765625, 0.30523681640625, 0.31884765625, 0.33245849609375, 0.3460693359375, 0.35968017578125, 0.373291015625, 0.38690185546875, 0.4005126953125, 0.41412353515625, 0.427734375]}, "parameters/model.layers.29.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 0.0, 1.0, 0.0, 1.0, 0.0, 4.0, 5.0, 3.0, 12.0, 31.0, 97.0, 430.0, 49075.0, 3968654.0, 16777216.0, 16777216.0, 1287486.0, 8930.0, 258.0, 71.0, 30.0, 14.0, 5.0, 6.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.59375, -0.5738525390625, -0.553955078125, -0.5340576171875, -0.51416015625, -0.4942626953125, -0.474365234375, -0.4544677734375, -0.4345703125, -0.4146728515625, -0.394775390625, -0.3748779296875, -0.35498046875, -0.3350830078125, -0.315185546875, -0.2952880859375, -0.275390625, -0.2554931640625, -0.235595703125, -0.2156982421875, -0.19580078125, -0.1759033203125, -0.156005859375, -0.1361083984375, -0.1162109375, -0.0963134765625, -0.076416015625, -0.0565185546875, -0.03662109375, -0.0167236328125, 0.003173828125, 0.0230712890625, 0.04296875, 0.0628662109375, 0.082763671875, 0.1026611328125, 0.12255859375, 0.1424560546875, 0.162353515625, 0.1822509765625, 0.2021484375, 0.2220458984375, 0.241943359375, 0.2618408203125, 0.28173828125, 0.3016357421875, 0.321533203125, 0.3414306640625, 0.361328125, 0.3812255859375, 0.401123046875, 0.4210205078125, 0.44091796875, 0.4608154296875, 0.480712890625, 0.5006103515625, 0.5205078125, 0.5404052734375, 0.560302734375, 0.5802001953125, 0.60009765625, 0.6199951171875, 0.639892578125, 0.6597900390625, 0.6796875]}, "parameters/model.layers.29.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 2.0, 3.0, 1.0, 2.0, 11.0, 17.0, 47.0, 142.0, 534.0, 1051.0, 1000.0, 728.0, 216.0, 72.0, 42.0, 32.0, 23.0, 30.0, 26.0, 13.0, 17.0, 13.0, 7.0, 4.0, 3.0, 13.0, 4.0, 2.0, 2.0, 4.0, 5.0, 2.0, 3.0, 3.0, 1.0, 2.0, 0.0, 2.0, 2.0, 4.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0022735595703125, 0.016927003860473633, 0.036127567291259766, 0.0553281307220459, 0.07452869415283203, 0.09372925758361816, 0.1129298210144043, 0.13213038444519043, 0.15133094787597656, 0.1705315113067627, 0.18973207473754883, 0.20893263816833496, 0.2281332015991211, 0.24733376502990723, 0.26653432846069336, 0.2857348918914795, 0.3049354553222656, 0.32413601875305176, 0.3433365821838379, 0.362537145614624, 0.38173770904541016, 0.4009382724761963, 0.4201388359069824, 0.43933939933776855, 0.4585399627685547, 0.4777405261993408, 0.49694108963012695, 0.5161416530609131, 0.5353422164916992, 0.5545427799224854, 0.5737433433532715, 0.5929439067840576, 0.6121444702148438, 0.6313450336456299, 0.650545597076416, 0.6697461605072021, 0.6889467239379883, 0.7081472873687744, 0.7273478507995605, 0.7465484142303467, 0.7657489776611328, 0.784949541091919, 0.8041501045227051, 0.8233506679534912, 0.8425512313842773, 0.8617517948150635, 0.8809523582458496, 0.9001529216766357, 0.9193534851074219, 0.938554048538208, 0.9577546119689941, 0.9769551753997803, 0.9961557388305664, 1.0153563022613525, 1.0345568656921387, 1.0537574291229248, 1.072957992553711, 1.092158555984497, 1.1113591194152832, 1.1305596828460693, 1.1497602462768555, 1.1689608097076416, 1.1881613731384277, 1.2073619365692139, 1.2265625]}, "parameters/model.layers.29.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 4.0, 0.0, 1.0, 3.0, 1.0, 1.0, 3.0, 2.0, 2.0, 0.0, 3.0, 2.0, 2.0, 3.0, 2.0, 3.0, 1.0, 2.0, 4.0, 2.0, 0.0, 3.0, 2.0, 1.0, 2.0, 2.0, 2.0, 3.0, 4.0, 6.0, 4.0, 4.0, 6.0, 6.0, 11.0, 19.0, 14.0, 16.0, 26.0, 8.0, 23.0, 30.0, 34.0, 90.0, 217.0, 913.0, 1648.0, 777.0, 162.0, 16.0, 1.0, 1.0, 1.0], "bins": [0.09619140625, 0.10482025146484375, 0.1134490966796875, 0.12207794189453125, 0.130706787109375, 0.13933563232421875, 0.1479644775390625, 0.15659332275390625, 0.16522216796875, 0.17385101318359375, 0.1824798583984375, 0.19110870361328125, 0.199737548828125, 0.20836639404296875, 0.2169952392578125, 0.22562408447265625, 0.2342529296875, 0.24288177490234375, 0.2515106201171875, 0.26013946533203125, 0.268768310546875, 0.27739715576171875, 0.2860260009765625, 0.29465484619140625, 0.30328369140625, 0.31191253662109375, 0.3205413818359375, 0.32917022705078125, 0.337799072265625, 0.34642791748046875, 0.3550567626953125, 0.36368560791015625, 0.372314453125, 0.38094329833984375, 0.3895721435546875, 0.39820098876953125, 0.406829833984375, 0.41545867919921875, 0.4240875244140625, 0.43271636962890625, 0.44134521484375, 0.44997406005859375, 0.4586029052734375, 0.46723175048828125, 0.475860595703125, 0.48448944091796875, 0.4931182861328125, 0.5017471313476562, 0.5103759765625, 0.5190048217773438, 0.5276336669921875, 0.5362625122070312, 0.544891357421875, 0.5535202026367188, 0.5621490478515625, 0.5707778930664062, 0.57940673828125, 0.5880355834960938, 0.5966644287109375, 0.6052932739257812, 0.613922119140625, 0.6225509643554688, 0.6311798095703125, 0.6398086547851562, 0.6484375]}, "parameters/model.layers.30.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 2.0, 3.0, 2.0, 7.0, 7.0, 22.0, 25.0, 35.0, 47.0, 64.0, 109.0, 170.0, 254.0, 489.0, 777.0, 1375.0, 2617.0, 5764.0, 14675.0, 37891.0, 93872.0, 236965.0, 574051.0, 1199575.0, 2248655.0, 3682709.0, 3828934.0, 2448433.0, 1303769.0, 646126.0, 266545.0, 111297.0, 42286.0, 16334.0, 6787.0, 2874.0, 1529.0, 840.0, 439.0, 315.0, 213.0, 126.0, 78.0, 33.0, 33.0, 19.0, 14.0, 7.0, 7.0, 5.0, 3.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.22265625, -0.214996337890625, -0.20733642578125, -0.199676513671875, -0.1920166015625, -0.184356689453125, -0.17669677734375, -0.169036865234375, -0.161376953125, -0.153717041015625, -0.14605712890625, -0.138397216796875, -0.1307373046875, -0.123077392578125, -0.11541748046875, -0.107757568359375, -0.10009765625, -0.092437744140625, -0.08477783203125, -0.077117919921875, -0.0694580078125, -0.061798095703125, -0.05413818359375, -0.046478271484375, -0.038818359375, -0.031158447265625, -0.02349853515625, -0.015838623046875, -0.0081787109375, -0.000518798828125, 0.00714111328125, 0.014801025390625, 0.0224609375, 0.030120849609375, 0.03778076171875, 0.045440673828125, 0.0531005859375, 0.060760498046875, 0.06842041015625, 0.076080322265625, 0.083740234375, 0.091400146484375, 0.09906005859375, 0.106719970703125, 0.1143798828125, 0.122039794921875, 0.12969970703125, 0.137359619140625, 0.14501953125, 0.152679443359375, 0.16033935546875, 0.167999267578125, 0.1756591796875, 0.183319091796875, 0.19097900390625, 0.198638916015625, 0.206298828125, 0.213958740234375, 0.22161865234375, 0.229278564453125, 0.2369384765625, 0.244598388671875, 0.25225830078125, 0.259918212890625, 0.267578125]}, "parameters/model.layers.30.self_attn.k_proj.weight": {"_type": "histogram", "values": [6.0, 2.0, 9.0, 4.0, 5.0, 9.0, 24.0, 13.0, 31.0, 42.0, 64.0, 75.0, 94.0, 122.0, 146.0, 243.0, 267.0, 411.0, 621.0, 999.0, 1808.0, 3608.0, 8201.0, 19508.0, 50573.0, 119452.0, 266849.0, 522775.0, 832983.0, 933269.0, 702346.0, 398209.0, 188890.0, 83266.0, 33487.0, 13357.0, 5573.0, 2681.0, 1439.0, 836.0, 492.0, 418.0, 278.0, 197.0, 155.0, 137.0, 84.0, 68.0, 37.0, 39.0, 30.0, 23.0, 16.0, 16.0, 3.0, 1.0, 4.0, 1.0, 1.0, 1.0, 3.0, 1.0, 1.0, 1.0], "bins": [-0.3046875, -0.294281005859375, -0.28387451171875, -0.273468017578125, -0.2630615234375, -0.252655029296875, -0.24224853515625, -0.231842041015625, -0.221435546875, -0.211029052734375, -0.20062255859375, -0.190216064453125, -0.1798095703125, -0.169403076171875, -0.15899658203125, -0.148590087890625, -0.13818359375, -0.127777099609375, -0.11737060546875, -0.106964111328125, -0.0965576171875, -0.086151123046875, -0.07574462890625, -0.065338134765625, -0.054931640625, -0.044525146484375, -0.03411865234375, -0.023712158203125, -0.0133056640625, -0.002899169921875, 0.00750732421875, 0.017913818359375, 0.0283203125, 0.038726806640625, 0.04913330078125, 0.059539794921875, 0.0699462890625, 0.080352783203125, 0.09075927734375, 0.101165771484375, 0.111572265625, 0.121978759765625, 0.13238525390625, 0.142791748046875, 0.1531982421875, 0.163604736328125, 0.17401123046875, 0.184417724609375, 0.19482421875, 0.205230712890625, 0.21563720703125, 0.226043701171875, 0.2364501953125, 0.246856689453125, 0.25726318359375, 0.267669677734375, 0.278076171875, 0.288482666015625, 0.29888916015625, 0.309295654296875, 0.3197021484375, 0.330108642578125, 0.34051513671875, 0.350921630859375, 0.361328125]}, "parameters/model.layers.30.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 3.0, 5.0, 11.0, 10.0, 13.0, 17.0, 26.0, 42.0, 61.0, 81.0, 113.0, 196.0, 217.0, 375.0, 554.0, 929.0, 1484.0, 3122.0, 6761.0, 13624.0, 26690.0, 54191.0, 96895.0, 171124.0, 257462.0, 387212.0, 498422.0, 595436.0, 588857.0, 493468.0, 376475.0, 258799.0, 162781.0, 94003.0, 52284.0, 25913.0, 13035.0, 6538.0, 2965.0, 1596.0, 891.0, 503.0, 301.0, 251.0, 164.0, 112.0, 102.0, 64.0, 30.0, 27.0, 16.0, 20.0, 11.0, 11.0, 1.0, 5.0, 0.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.1767578125, -0.171051025390625, -0.16534423828125, -0.159637451171875, -0.1539306640625, -0.148223876953125, -0.14251708984375, -0.136810302734375, -0.131103515625, -0.125396728515625, -0.11968994140625, -0.113983154296875, -0.1082763671875, -0.102569580078125, -0.09686279296875, -0.091156005859375, -0.08544921875, -0.079742431640625, -0.07403564453125, -0.068328857421875, -0.0626220703125, -0.056915283203125, -0.05120849609375, -0.045501708984375, -0.039794921875, -0.034088134765625, -0.02838134765625, -0.022674560546875, -0.0169677734375, -0.011260986328125, -0.00555419921875, 0.000152587890625, 0.005859375, 0.011566162109375, 0.01727294921875, 0.022979736328125, 0.0286865234375, 0.034393310546875, 0.04010009765625, 0.045806884765625, 0.051513671875, 0.057220458984375, 0.06292724609375, 0.068634033203125, 0.0743408203125, 0.080047607421875, 0.08575439453125, 0.091461181640625, 0.09716796875, 0.102874755859375, 0.10858154296875, 0.114288330078125, 0.1199951171875, 0.125701904296875, 0.13140869140625, 0.137115478515625, 0.142822265625, 0.148529052734375, 0.15423583984375, 0.159942626953125, 0.1656494140625, 0.171356201171875, 0.17706298828125, 0.182769775390625, 0.1884765625]}, "parameters/model.layers.30.self_attn.o_proj.weight": {"_type": "histogram", "values": [5.0, 1.0, 4.0, 2.0, 1.0, 4.0, 4.0, 11.0, 12.0, 13.0, 10.0, 9.0, 12.0, 18.0, 15.0, 13.0, 38.0, 37.0, 35.0, 45.0, 73.0, 136.0, 244.0, 783.0, 3942.0, 18143.0, 82680.0, 308817.0, 935946.0, 2176884.0, 3597512.0, 4070917.0, 3098662.0, 1610552.0, 627781.0, 185046.0, 45462.0, 10248.0, 2040.0, 468.0, 187.0, 101.0, 51.0, 48.0, 35.0, 31.0, 11.0, 17.0, 11.0, 9.0, 7.0, 11.0, 11.0, 7.0, 6.0, 12.0, 8.0, 8.0, 3.0, 6.0, 2.0, 2.0, 6.0, 1.0], "bins": [-0.2412109375, -0.2335052490234375, -0.225799560546875, -0.2180938720703125, -0.21038818359375, -0.2026824951171875, -0.194976806640625, -0.1872711181640625, -0.1795654296875, -0.1718597412109375, -0.164154052734375, -0.1564483642578125, -0.14874267578125, -0.1410369873046875, -0.133331298828125, -0.1256256103515625, -0.117919921875, -0.1102142333984375, -0.102508544921875, -0.0948028564453125, -0.08709716796875, -0.0793914794921875, -0.071685791015625, -0.0639801025390625, -0.0562744140625, -0.0485687255859375, -0.040863037109375, -0.0331573486328125, -0.02545166015625, -0.0177459716796875, -0.010040283203125, -0.0023345947265625, 0.00537109375, 0.0130767822265625, 0.020782470703125, 0.0284881591796875, 0.03619384765625, 0.0438995361328125, 0.051605224609375, 0.0593109130859375, 0.0670166015625, 0.0747222900390625, 0.082427978515625, 0.0901336669921875, 0.09783935546875, 0.1055450439453125, 0.113250732421875, 0.1209564208984375, 0.128662109375, 0.1363677978515625, 0.144073486328125, 0.1517791748046875, 0.15948486328125, 0.1671905517578125, 0.174896240234375, 0.1826019287109375, 0.1903076171875, 0.1980133056640625, 0.205718994140625, 0.2134246826171875, 0.22113037109375, 0.2288360595703125, 0.236541748046875, 0.2442474365234375, 0.251953125]}, "parameters/model.layers.30.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0, 5.0, 3.0, 4.0, 8.0, 23.0, 22.0, 42.0, 77.0, 215.0, 672.0, 3745.0, 20706.0, 117905.0, 830701.0, 8566695.0, 16777216.0, 16777216.0, 13916784.0, 1446158.0, 181862.0, 30920.0, 5578.0, 967.0, 253.0, 76.0, 47.0, 28.0, 8.0, 14.0, 3.0, 5.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.5234375, -0.50714111328125, -0.4908447265625, -0.47454833984375, -0.458251953125, -0.44195556640625, -0.4256591796875, -0.40936279296875, -0.39306640625, -0.37677001953125, -0.3604736328125, -0.34417724609375, -0.327880859375, -0.31158447265625, -0.2952880859375, -0.27899169921875, -0.2626953125, -0.24639892578125, -0.2301025390625, -0.21380615234375, -0.197509765625, -0.18121337890625, -0.1649169921875, -0.14862060546875, -0.13232421875, -0.11602783203125, -0.0997314453125, -0.08343505859375, -0.067138671875, -0.05084228515625, -0.0345458984375, -0.01824951171875, -0.001953125, 0.01434326171875, 0.0306396484375, 0.04693603515625, 0.063232421875, 0.07952880859375, 0.0958251953125, 0.11212158203125, 0.12841796875, 0.14471435546875, 0.1610107421875, 0.17730712890625, 0.193603515625, 0.20989990234375, 0.2261962890625, 0.24249267578125, 0.2587890625, 0.27508544921875, 0.2913818359375, 0.30767822265625, 0.323974609375, 0.34027099609375, 0.3565673828125, 0.37286376953125, 0.38916015625, 0.40545654296875, 0.4217529296875, 0.43804931640625, 0.454345703125, 0.47064208984375, 0.4869384765625, 0.50323486328125, 0.51953125]}, "parameters/model.layers.30.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 2.0, 4.0, 3.0, 10.0, 10.0, 22.0, 39.0, 49.0, 64.0, 154.0, 399.0, 1197.0, 15779.0, 345386.0, 3793397.0, 16777216.0, 16777216.0, 10806127.0, 1691314.0, 109923.0, 4766.0, 734.0, 248.0, 126.0, 73.0, 38.0, 22.0, 12.0, 12.0, 4.0, 5.0, 3.0, 2.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.3984375, -0.385711669921875, -0.37298583984375, -0.360260009765625, -0.3475341796875, -0.334808349609375, -0.32208251953125, -0.309356689453125, -0.296630859375, -0.283905029296875, -0.27117919921875, -0.258453369140625, -0.2457275390625, -0.233001708984375, -0.22027587890625, -0.207550048828125, -0.19482421875, -0.182098388671875, -0.16937255859375, -0.156646728515625, -0.1439208984375, -0.131195068359375, -0.11846923828125, -0.105743408203125, -0.093017578125, -0.080291748046875, -0.06756591796875, -0.054840087890625, -0.0421142578125, -0.029388427734375, -0.01666259765625, -0.003936767578125, 0.0087890625, 0.021514892578125, 0.03424072265625, 0.046966552734375, 0.0596923828125, 0.072418212890625, 0.08514404296875, 0.097869873046875, 0.110595703125, 0.123321533203125, 0.13604736328125, 0.148773193359375, 0.1614990234375, 0.174224853515625, 0.18695068359375, 0.199676513671875, 0.21240234375, 0.225128173828125, 0.23785400390625, 0.250579833984375, 0.2633056640625, 0.276031494140625, 0.28875732421875, 0.301483154296875, 0.314208984375, 0.326934814453125, 0.33966064453125, 0.352386474609375, 0.3651123046875, 0.377838134765625, 0.39056396484375, 0.403289794921875, 0.416015625]}, "parameters/model.layers.30.input_layernorm.weight": {"_type": "histogram", "values": [2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 3.0, 2.0, 4.0, 11.0, 36.0, 169.0, 920.0, 1439.0, 778.0, 205.0, 100.0, 77.0, 40.0, 38.0, 21.0, 20.0, 21.0, 15.0, 10.0, 16.0, 10.0, 15.0, 17.0, 15.0, 14.0, 9.0, 14.0, 6.0, 6.0, 5.0, 2.0, 7.0, 4.0, 3.0, 2.0, 1.0, 3.0, 1.0, 4.0, 2.0, 4.0, 1.0, 2.0, 5.0, 2.0, 1.0, 3.0, 1.0, 2.0, 1.0, 1.0, 0.0, 2.0], "bins": [0.09326171875, 0.11438751220703125, 0.1355133056640625, 0.15663909912109375, 0.177764892578125, 0.19889068603515625, 0.2200164794921875, 0.24114227294921875, 0.26226806640625, 0.28339385986328125, 0.3045196533203125, 0.32564544677734375, 0.346771240234375, 0.36789703369140625, 0.3890228271484375, 0.41014862060546875, 0.4312744140625, 0.45240020751953125, 0.4735260009765625, 0.49465179443359375, 0.515777587890625, 0.5369033813476562, 0.5580291748046875, 0.5791549682617188, 0.60028076171875, 0.6214065551757812, 0.6425323486328125, 0.6636581420898438, 0.684783935546875, 0.7059097290039062, 0.7270355224609375, 0.7481613159179688, 0.769287109375, 0.7904129028320312, 0.8115386962890625, 0.8326644897460938, 0.853790283203125, 0.8749160766601562, 0.8960418701171875, 0.9171676635742188, 0.93829345703125, 0.9594192504882812, 0.9805450439453125, 1.0016708374023438, 1.022796630859375, 1.0439224243164062, 1.0650482177734375, 1.0861740112304688, 1.1072998046875, 1.1284255981445312, 1.1495513916015625, 1.1706771850585938, 1.191802978515625, 1.2129287719726562, 1.2340545654296875, 1.2551803588867188, 1.27630615234375, 1.2974319458007812, 1.3185577392578125, 1.3396835327148438, 1.360809326171875, 1.3819351196289062, 1.4030609130859375, 1.4241867065429688, 1.4453125]}, "parameters/model.layers.30.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 4.0, 0.0, 2.0, 7.0, 2.0, 5.0, 3.0, 0.0, 4.0, 5.0, 4.0, 6.0, 5.0, 3.0, 10.0, 4.0, 7.0, 4.0, 4.0, 9.0, 16.0, 5.0, 8.0, 16.0, 23.0, 28.0, 29.0, 35.0, 97.0, 287.0, 1909.0, 1110.0, 371.0, 49.0, 14.0, 0.0, 4.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.08349609375, 0.09714508056640625, 0.1107940673828125, 0.12444305419921875, 0.138092041015625, 0.15174102783203125, 0.1653900146484375, 0.17903900146484375, 0.19268798828125, 0.20633697509765625, 0.2199859619140625, 0.23363494873046875, 0.247283935546875, 0.26093292236328125, 0.2745819091796875, 0.28823089599609375, 0.3018798828125, 0.31552886962890625, 0.3291778564453125, 0.34282684326171875, 0.356475830078125, 0.37012481689453125, 0.3837738037109375, 0.39742279052734375, 0.41107177734375, 0.42472076416015625, 0.4383697509765625, 0.45201873779296875, 0.465667724609375, 0.47931671142578125, 0.4929656982421875, 0.5066146850585938, 0.520263671875, 0.5339126586914062, 0.5475616455078125, 0.5612106323242188, 0.574859619140625, 0.5885086059570312, 0.6021575927734375, 0.6158065795898438, 0.62945556640625, 0.6431045532226562, 0.6567535400390625, 0.6704025268554688, 0.684051513671875, 0.6977005004882812, 0.7113494873046875, 0.7249984741210938, 0.7386474609375, 0.7522964477539062, 0.7659454345703125, 0.7795944213867188, 0.793243408203125, 0.8068923950195312, 0.8205413818359375, 0.8341903686523438, 0.84783935546875, 0.8614883422851562, 0.8751373291015625, 0.8887863159179688, 0.902435302734375, 0.9160842895507812, 0.9297332763671875, 0.9433822631835938, 0.95703125]}, "parameters/model.layers.31.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 8.0, 4.0, 5.0, 8.0, 11.0, 13.0, 21.0, 30.0, 46.0, 48.0, 71.0, 118.0, 178.0, 307.0, 609.0, 1167.0, 3003.0, 8634.0, 34448.0, 140510.0, 546728.0, 1808309.0, 4347519.0, 5461133.0, 2975286.0, 1060770.0, 288675.0, 73117.0, 17613.0, 5023.0, 1844.0, 807.0, 466.0, 242.0, 150.0, 81.0, 66.0, 45.0, 32.0, 17.0, 11.0, 7.0, 5.0, 3.0, 4.0, 2.0, 1.0, 2.0, 2.0, 2.0, 2.0, 1.0, 2.0, 1.0], "bins": [-0.4140625, -0.401611328125, -0.38916015625, -0.376708984375, -0.3642578125, -0.351806640625, -0.33935546875, -0.326904296875, -0.314453125, -0.302001953125, -0.28955078125, -0.277099609375, -0.2646484375, -0.252197265625, -0.23974609375, -0.227294921875, -0.21484375, -0.202392578125, -0.18994140625, -0.177490234375, -0.1650390625, -0.152587890625, -0.14013671875, -0.127685546875, -0.115234375, -0.102783203125, -0.09033203125, -0.077880859375, -0.0654296875, -0.052978515625, -0.04052734375, -0.028076171875, -0.015625, -0.003173828125, 0.00927734375, 0.021728515625, 0.0341796875, 0.046630859375, 0.05908203125, 0.071533203125, 0.083984375, 0.096435546875, 0.10888671875, 0.121337890625, 0.1337890625, 0.146240234375, 0.15869140625, 0.171142578125, 0.18359375, 0.196044921875, 0.20849609375, 0.220947265625, 0.2333984375, 0.245849609375, 0.25830078125, 0.270751953125, 0.283203125, 0.295654296875, 0.30810546875, 0.320556640625, 0.3330078125, 0.345458984375, 0.35791015625, 0.370361328125, 0.3828125]}, "parameters/model.layers.31.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 4.0, 3.0, 4.0, 9.0, 5.0, 10.0, 18.0, 28.0, 31.0, 38.0, 68.0, 101.0, 111.0, 190.0, 302.0, 510.0, 1112.0, 2484.0, 8235.0, 28681.0, 100763.0, 279750.0, 658644.0, 1070692.0, 1042005.0, 613672.0, 260026.0, 89730.0, 25349.0, 7063.0, 2402.0, 951.0, 471.0, 247.0, 201.0, 106.0, 79.0, 56.0, 44.0, 26.0, 22.0, 11.0, 13.0, 6.0, 4.0, 7.0, 1.0, 6.0, 2.0, 1.0, 2.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.373046875, -0.358673095703125, -0.34429931640625, -0.329925537109375, -0.3155517578125, -0.301177978515625, -0.28680419921875, -0.272430419921875, -0.258056640625, -0.243682861328125, -0.22930908203125, -0.214935302734375, -0.2005615234375, -0.186187744140625, -0.17181396484375, -0.157440185546875, -0.14306640625, -0.128692626953125, -0.11431884765625, -0.099945068359375, -0.0855712890625, -0.071197509765625, -0.05682373046875, -0.042449951171875, -0.028076171875, -0.013702392578125, 0.00067138671875, 0.015045166015625, 0.0294189453125, 0.043792724609375, 0.05816650390625, 0.072540283203125, 0.0869140625, 0.101287841796875, 0.11566162109375, 0.130035400390625, 0.1444091796875, 0.158782958984375, 0.17315673828125, 0.187530517578125, 0.201904296875, 0.216278076171875, 0.23065185546875, 0.245025634765625, 0.2593994140625, 0.273773193359375, 0.28814697265625, 0.302520751953125, 0.31689453125, 0.331268310546875, 0.34564208984375, 0.360015869140625, 0.3743896484375, 0.388763427734375, 0.40313720703125, 0.417510986328125, 0.431884765625, 0.446258544921875, 0.46063232421875, 0.475006103515625, 0.4893798828125, 0.503753662109375, 0.51812744140625, 0.532501220703125, 0.546875]}, "parameters/model.layers.31.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 2.0, 1.0, 3.0, 3.0, 4.0, 3.0, 9.0, 11.0, 25.0, 26.0, 52.0, 97.0, 107.0, 221.0, 383.0, 741.0, 1616.0, 4043.0, 10547.0, 28818.0, 75602.0, 180624.0, 372991.0, 644238.0, 860329.0, 832626.0, 591112.0, 332146.0, 157195.0, 61401.0, 24294.0, 8927.0, 3201.0, 1420.0, 621.0, 302.0, 204.0, 119.0, 91.0, 64.0, 26.0, 23.0, 6.0, 10.0, 5.0, 7.0, 3.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.28515625, -0.2776336669921875, -0.270111083984375, -0.2625885009765625, -0.25506591796875, -0.2475433349609375, -0.240020751953125, -0.2324981689453125, -0.2249755859375, -0.2174530029296875, -0.209930419921875, -0.2024078369140625, -0.19488525390625, -0.1873626708984375, -0.179840087890625, -0.1723175048828125, -0.164794921875, -0.1572723388671875, -0.149749755859375, -0.1422271728515625, -0.13470458984375, -0.1271820068359375, -0.119659423828125, -0.1121368408203125, -0.1046142578125, -0.0970916748046875, -0.089569091796875, -0.0820465087890625, -0.07452392578125, -0.0670013427734375, -0.059478759765625, -0.0519561767578125, -0.04443359375, -0.0369110107421875, -0.029388427734375, -0.0218658447265625, -0.01434326171875, -0.0068206787109375, 0.000701904296875, 0.0082244873046875, 0.0157470703125, 0.0232696533203125, 0.030792236328125, 0.0383148193359375, 0.04583740234375, 0.0533599853515625, 0.060882568359375, 0.0684051513671875, 0.075927734375, 0.0834503173828125, 0.090972900390625, 0.0984954833984375, 0.10601806640625, 0.1135406494140625, 0.121063232421875, 0.1285858154296875, 0.1361083984375, 0.1436309814453125, 0.151153564453125, 0.1586761474609375, 0.16619873046875, 0.1737213134765625, 0.181243896484375, 0.1887664794921875, 0.1962890625]}, "parameters/model.layers.31.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 2.0, 0.0, 1.0, 2.0, 0.0, 0.0, 2.0, 3.0, 5.0, 3.0, 5.0, 5.0, 5.0, 13.0, 11.0, 26.0, 33.0, 50.0, 76.0, 101.0, 202.0, 286.0, 497.0, 891.0, 2378.0, 13579.0, 115008.0, 840095.0, 3582325.0, 6635087.0, 4256471.0, 1138484.0, 166212.0, 19554.0, 3160.0, 1163.0, 575.0, 339.0, 188.0, 114.0, 86.0, 58.0, 29.0, 24.0, 23.0, 12.0, 9.0, 4.0, 5.0, 2.0, 1.0, 5.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0], "bins": [-0.4140625, -0.4017333984375, -0.389404296875, -0.3770751953125, -0.36474609375, -0.3524169921875, -0.340087890625, -0.3277587890625, -0.3154296875, -0.3031005859375, -0.290771484375, -0.2784423828125, -0.26611328125, -0.2537841796875, -0.241455078125, -0.2291259765625, -0.216796875, -0.2044677734375, -0.192138671875, -0.1798095703125, -0.16748046875, -0.1551513671875, -0.142822265625, -0.1304931640625, -0.1181640625, -0.1058349609375, -0.093505859375, -0.0811767578125, -0.06884765625, -0.0565185546875, -0.044189453125, -0.0318603515625, -0.01953125, -0.0072021484375, 0.005126953125, 0.0174560546875, 0.02978515625, 0.0421142578125, 0.054443359375, 0.0667724609375, 0.0791015625, 0.0914306640625, 0.103759765625, 0.1160888671875, 0.12841796875, 0.1407470703125, 0.153076171875, 0.1654052734375, 0.177734375, 0.1900634765625, 0.202392578125, 0.2147216796875, 0.22705078125, 0.2393798828125, 0.251708984375, 0.2640380859375, 0.2763671875, 0.2886962890625, 0.301025390625, 0.3133544921875, 0.32568359375, 0.3380126953125, 0.350341796875, 0.3626708984375, 0.375]}, "parameters/model.layers.31.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 2.0, 2.0, 1.0, 4.0, 2.0, 1.0, 18.0, 22.0, 40.0, 48.0, 78.0, 101.0, 228.0, 366.0, 734.0, 1842.0, 7741.0, 45613.0, 345771.0, 3962516.0, 16777216.0, 16777216.0, 16777216.0, 1721793.0, 172738.0, 25435.0, 4607.0, 1341.0, 604.0, 328.0, 168.0, 101.0, 57.0, 42.0, 22.0, 21.0, 11.0, 4.0, 2.0, 1.0, 2.0, 3.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.68359375, -0.6630859375, -0.642578125, -0.6220703125, -0.6015625, -0.5810546875, -0.560546875, -0.5400390625, -0.51953125, -0.4990234375, -0.478515625, -0.4580078125, -0.4375, -0.4169921875, -0.396484375, -0.3759765625, -0.35546875, -0.3349609375, -0.314453125, -0.2939453125, -0.2734375, -0.2529296875, -0.232421875, -0.2119140625, -0.19140625, -0.1708984375, -0.150390625, -0.1298828125, -0.109375, -0.0888671875, -0.068359375, -0.0478515625, -0.02734375, -0.0068359375, 0.013671875, 0.0341796875, 0.0546875, 0.0751953125, 0.095703125, 0.1162109375, 0.13671875, 0.1572265625, 0.177734375, 0.1982421875, 0.21875, 0.2392578125, 0.259765625, 0.2802734375, 0.30078125, 0.3212890625, 0.341796875, 0.3623046875, 0.3828125, 0.4033203125, 0.423828125, 0.4443359375, 0.46484375, 0.4853515625, 0.505859375, 0.5263671875, 0.546875, 0.5673828125, 0.587890625, 0.6083984375, 0.62890625]}, "parameters/model.layers.31.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 1.0, 2.0, 2.0, 0.0, 1.0, 1.0, 1.0, 3.0, 2.0, 2.0, 1.0, 1.0, 1.0, 2.0, 1.0, 1.0, 15.0, 11.0, 19.0, 23.0, 46.0, 72.0, 156.0, 267.0, 660.0, 2160.0, 29508.0, 599932.0, 6116689.0, 16777216.0, 16777216.0, 5057109.0, 463227.0, 22595.0, 1803.0, 608.0, 249.0, 132.0, 48.0, 47.0, 25.0, 16.0, 14.0, 5.0, 4.0, 2.0, 0.0, 0.0, 1.0, 1.0, 1.0, 3.0, 1.0, 2.0, 1.0, 1.0, 3.0, 2.0, 0.0, 1.0, 3.0], "bins": [-0.474609375, -0.460205078125, -0.44580078125, -0.431396484375, -0.4169921875, -0.402587890625, -0.38818359375, -0.373779296875, -0.359375, -0.344970703125, -0.33056640625, -0.316162109375, -0.3017578125, -0.287353515625, -0.27294921875, -0.258544921875, -0.244140625, -0.229736328125, -0.21533203125, -0.200927734375, -0.1865234375, -0.172119140625, -0.15771484375, -0.143310546875, -0.12890625, -0.114501953125, -0.10009765625, -0.085693359375, -0.0712890625, -0.056884765625, -0.04248046875, -0.028076171875, -0.013671875, 0.000732421875, 0.01513671875, 0.029541015625, 0.0439453125, 0.058349609375, 0.07275390625, 0.087158203125, 0.1015625, 0.115966796875, 0.13037109375, 0.144775390625, 0.1591796875, 0.173583984375, 0.18798828125, 0.202392578125, 0.216796875, 0.231201171875, 0.24560546875, 0.260009765625, 0.2744140625, 0.288818359375, 0.30322265625, 0.317626953125, 0.33203125, 0.346435546875, 0.36083984375, 0.375244140625, 0.3896484375, 0.404052734375, 0.41845703125, 0.432861328125, 0.447265625]}, "parameters/model.layers.31.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 2.0, 1.0, 2.0, 2.0, 4.0, 3.0, 1.0, 11.0, 5.0, 4.0, 6.0, 9.0, 5.0, 18.0, 29.0, 35.0, 63.0, 111.0, 155.0, 219.0, 310.0, 501.0, 615.0, 568.0, 557.0, 307.0, 180.0, 65.0, 28.0, 38.0, 23.0, 11.0, 11.0, 14.0, 13.0, 7.0, 12.0, 8.0, 15.0, 13.0, 7.0, 11.0, 12.0, 8.0, 10.0, 11.0, 8.0, 7.0, 10.0, 4.0, 2.0, 2.0, 7.0, 3.0, 0.0, 0.0, 1.0, 3.0, 3.0, 1.0, 1.0, 1.0, 2.0], "bins": [0.06884765625, 0.08437347412109375, 0.0998992919921875, 0.11542510986328125, 0.130950927734375, 0.14647674560546875, 0.1620025634765625, 0.17752838134765625, 0.19305419921875, 0.20858001708984375, 0.2241058349609375, 0.23963165283203125, 0.255157470703125, 0.27068328857421875, 0.2862091064453125, 0.30173492431640625, 0.3172607421875, 0.33278656005859375, 0.3483123779296875, 0.36383819580078125, 0.379364013671875, 0.39488983154296875, 0.4104156494140625, 0.42594146728515625, 0.44146728515625, 0.45699310302734375, 0.4725189208984375, 0.48804473876953125, 0.503570556640625, 0.5190963745117188, 0.5346221923828125, 0.5501480102539062, 0.565673828125, 0.5811996459960938, 0.5967254638671875, 0.6122512817382812, 0.627777099609375, 0.6433029174804688, 0.6588287353515625, 0.6743545532226562, 0.68988037109375, 0.7054061889648438, 0.7209320068359375, 0.7364578247070312, 0.751983642578125, 0.7675094604492188, 0.7830352783203125, 0.7985610961914062, 0.8140869140625, 0.8296127319335938, 0.8451385498046875, 0.8606643676757812, 0.876190185546875, 0.8917160034179688, 0.9072418212890625, 0.9227676391601562, 0.93829345703125, 0.9538192749023438, 0.9693450927734375, 0.9848709106445312, 1.000396728515625, 1.0159225463867188, 1.0314483642578125, 1.0469741821289062, 1.0625]}, "parameters/model.layers.31.post_attention_layernorm.weight": {"_type": "histogram", "values": [2.0, 2.0, 4.0, 7.0, 10.0, 10.0, 9.0, 17.0, 11.0, 6.0, 18.0, 17.0, 27.0, 42.0, 84.0, 131.0, 206.0, 326.0, 696.0, 1178.0, 946.0, 250.0, 54.0, 13.0, 6.0, 5.0, 2.0, 3.0, 3.0, 2.0, 1.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [0.042724609375, 0.06659317016601562, 0.09046173095703125, 0.11433029174804688, 0.1381988525390625, 0.16206741333007812, 0.18593597412109375, 0.20980453491210938, 0.233673095703125, 0.2575416564941406, 0.28141021728515625, 0.3052787780761719, 0.3291473388671875, 0.3530158996582031, 0.37688446044921875, 0.4007530212402344, 0.42462158203125, 0.4484901428222656, 0.47235870361328125, 0.4962272644042969, 0.5200958251953125, 0.5439643859863281, 0.5678329467773438, 0.5917015075683594, 0.615570068359375, 0.6394386291503906, 0.6633071899414062, 0.6871757507324219, 0.7110443115234375, 0.7349128723144531, 0.7587814331054688, 0.7826499938964844, 0.8065185546875, 0.8303871154785156, 0.8542556762695312, 0.8781242370605469, 0.9019927978515625, 0.9258613586425781, 0.9497299194335938, 0.9735984802246094, 0.997467041015625, 1.0213356018066406, 1.0452041625976562, 1.0690727233886719, 1.0929412841796875, 1.1168098449707031, 1.1406784057617188, 1.1645469665527344, 1.18841552734375, 1.2122840881347656, 1.2361526489257812, 1.2600212097167969, 1.2838897705078125, 1.3077583312988281, 1.3316268920898438, 1.3554954528808594, 1.379364013671875, 1.4032325744628906, 1.4271011352539062, 1.4509696960449219, 1.4748382568359375, 1.4987068176269531, 1.5225753784179688, 1.5464439392089844, 1.5703125]}, "parameters/model.norm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 4.0, 2.0, 3.0, 2.0, 3.0, 1.0, 3.0, 5.0, 5.0, 9.0, 3.0, 8.0, 7.0, 6.0, 7.0, 5.0, 5.0, 6.0, 5.0, 15.0, 17.0, 7.0, 30.0, 45.0, 65.0, 110.0, 111.0, 170.0, 255.0, 399.0, 725.0, 460.0, 895.0, 512.0, 138.0, 23.0, 5.0, 11.0, 0.0, 2.0, 3.0, 2.0, 0.0, 1.0, 0.0, 2.0], "bins": [0.421875, 0.465576171875, 0.50927734375, 0.552978515625, 0.5966796875, 0.640380859375, 0.68408203125, 0.727783203125, 0.771484375, 0.815185546875, 0.85888671875, 0.902587890625, 0.9462890625, 0.989990234375, 1.03369140625, 1.077392578125, 1.12109375, 1.164794921875, 1.20849609375, 1.252197265625, 1.2958984375, 1.339599609375, 1.38330078125, 1.427001953125, 1.470703125, 1.514404296875, 1.55810546875, 1.601806640625, 1.6455078125, 1.689208984375, 1.73291015625, 1.776611328125, 1.8203125, 1.864013671875, 1.90771484375, 1.951416015625, 1.9951171875, 2.038818359375, 2.08251953125, 2.126220703125, 2.169921875, 2.213623046875, 2.25732421875, 2.301025390625, 2.3447265625, 2.388427734375, 2.43212890625, 2.475830078125, 2.51953125, 2.563232421875, 2.60693359375, 2.650634765625, 2.6943359375, 2.738037109375, 2.78173828125, 2.825439453125, 2.869140625, 2.912841796875, 2.95654296875, 3.000244140625, 3.0439453125, 3.087646484375, 3.13134765625, 3.175048828125, 3.21875]}, "parameters/lm_head.weight": {"_type": "histogram", "values": [1.0, 3.0, 6.0, 6.0, 5.0, 6.0, 2.0, 1.0, 2.0, 6.0, 9.0, 17.0, 25.0, 56.0, 72.0, 163.0, 326.0, 599.0, 1207.0, 2155.0, 4749.0, 10393.0, 20501.0, 39261.0, 81776.0, 174933.0, 461362.0, 2127354.0, 11818193.0, 16777216.0, 16777216.0, 16777216.0, 16777216.0, 16777216.0, 7650735.0, 1248089.0, 219933.0, 58803.0, 26709.0, 14579.0, 8837.0, 5203.0, 3021.0, 1693.0, 1005.0, 652.0, 332.0, 193.0, 142.0, 66.0, 39.0, 23.0, 10.0, 19.0, 6.0, 5.0, 1.0, 4.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.34375, -0.332794189453125, -0.32183837890625, -0.310882568359375, -0.2999267578125, -0.288970947265625, -0.27801513671875, -0.267059326171875, -0.256103515625, -0.245147705078125, -0.23419189453125, -0.223236083984375, -0.2122802734375, -0.201324462890625, -0.19036865234375, -0.179412841796875, -0.16845703125, -0.157501220703125, -0.14654541015625, -0.135589599609375, -0.1246337890625, -0.113677978515625, -0.10272216796875, -0.091766357421875, -0.080810546875, -0.069854736328125, -0.05889892578125, -0.047943115234375, -0.0369873046875, -0.026031494140625, -0.01507568359375, -0.004119873046875, 0.0068359375, 0.017791748046875, 0.02874755859375, 0.039703369140625, 0.0506591796875, 0.061614990234375, 0.07257080078125, 0.083526611328125, 0.094482421875, 0.105438232421875, 0.11639404296875, 0.127349853515625, 0.1383056640625, 0.149261474609375, 0.16021728515625, 0.171173095703125, 0.18212890625, 0.193084716796875, 0.20404052734375, 0.214996337890625, 0.2259521484375, 0.236907958984375, 0.24786376953125, 0.258819580078125, 0.269775390625, 0.280731201171875, 0.29168701171875, 0.302642822265625, 0.3135986328125, 0.324554443359375, 0.33551025390625, 0.346466064453125, 0.357421875]}, "gradients/lm_head.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 2.0, 2.0, 1.0, 1.0, 0.0, 2.0, 3.0, 5.0, 4.0, 12.0, 12.0, 38.0, 111.0, 366.0, 1766.0, 124887.0, 16777216.0, 3769.0, 508.0, 155.0, 49.0, 20.0, 6.0, 6.0, 1.0, 3.0, 3.0, 2.0, 0.0, 0.0, 2.0, 1.0, 0.0, 1.0, 2.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.1728515625, -0.167938232421875, -0.16302490234375, -0.158111572265625, -0.1531982421875, -0.148284912109375, -0.14337158203125, -0.138458251953125, -0.133544921875, -0.128631591796875, -0.12371826171875, -0.118804931640625, -0.1138916015625, -0.108978271484375, -0.10406494140625, -0.099151611328125, -0.09423828125, -0.089324951171875, -0.08441162109375, -0.079498291015625, -0.0745849609375, -0.069671630859375, -0.06475830078125, -0.059844970703125, -0.054931640625, -0.050018310546875, -0.04510498046875, -0.040191650390625, -0.0352783203125, -0.030364990234375, -0.02545166015625, -0.020538330078125, -0.015625, -0.010711669921875, -0.00579833984375, -0.000885009765625, 0.0040283203125, 0.008941650390625, 0.01385498046875, 0.018768310546875, 0.023681640625, 0.028594970703125, 0.03350830078125, 0.038421630859375, 0.0433349609375, 0.048248291015625, 0.05316162109375, 0.058074951171875, 0.06298828125, 0.067901611328125, 0.07281494140625, 0.077728271484375, 0.0826416015625, 0.087554931640625, 0.09246826171875, 0.097381591796875, 0.102294921875, 0.107208251953125, 0.11212158203125, 0.117034912109375, 0.1219482421875, 0.126861572265625, 0.13177490234375, 0.136688232421875, 0.1416015625]}, "gradients/model.norm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 2.0, 4.0, 0.0, 4.0, 2.0, 27.0, 374.0, 2480.0, 1086.0, 83.0, 7.0, 2.0, 2.0, 4.0, 1.0, 0.0, 0.0, 1.0, 2.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00201416015625, -0.0018773078918457031, -0.0017404556274414062, -0.0016036033630371094, -0.0014667510986328125, -0.0013298988342285156, -0.0011930465698242188, -0.0010561943054199219, -0.000919342041015625, -0.0007824897766113281, -0.0006456375122070312, -0.0005087852478027344, -0.0003719329833984375, -0.00023508071899414062, -9.822845458984375e-05, 3.8623809814453125e-05, 0.00017547607421875, 0.0003123283386230469, 0.00044918060302734375, 0.0005860328674316406, 0.0007228851318359375, 0.0008597373962402344, 0.0009965896606445312, 0.0011334419250488281, 0.001270294189453125, 0.0014071464538574219, 0.0015439987182617188, 0.0016808509826660156, 0.0018177032470703125, 0.0019545555114746094, 0.0020914077758789062, 0.002228260040283203, 0.0023651123046875, 0.002501964569091797, 0.0026388168334960938, 0.0027756690979003906, 0.0029125213623046875, 0.0030493736267089844, 0.0031862258911132812, 0.003323078155517578, 0.003459930419921875, 0.003596782684326172, 0.0037336349487304688, 0.0038704872131347656, 0.0040073394775390625, 0.004144191741943359, 0.004281044006347656, 0.004417896270751953, 0.00455474853515625, 0.004691600799560547, 0.004828453063964844, 0.004965305328369141, 0.0051021575927734375, 0.005239009857177734, 0.005375862121582031, 0.005512714385986328, 0.005649566650390625, 0.005786418914794922, 0.005923271179199219, 0.006060123443603516, 0.0061969757080078125, 0.006333827972412109, 0.006470680236816406, 0.006607532501220703, 0.006744384765625]}, "gradients/model.layers.31.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 1.0, 4.0, 3.0, 5.0, 9.0, 8.0, 10.0, 18.0, 24.0, 35.0, 35.0, 61.0, 65.0, 75.0, 82.0, 128.0, 173.0, 240.0, 335.0, 528.0, 929.0, 1760.0, 5525.0, 67335.0, 16777216.0, 1516659.0, 30196.0, 3653.0, 1379.0, 777.0, 429.0, 245.0, 204.0, 141.0, 110.0, 88.0, 72.0, 62.0, 51.0, 37.0, 27.0, 14.0, 12.0, 10.0, 8.0, 3.0, 4.0, 1.0, 4.0, 3.0, 0.0, 2.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.036376953125, -0.035198211669921875, -0.03401947021484375, -0.032840728759765625, -0.0316619873046875, -0.030483245849609375, -0.02930450439453125, -0.028125762939453125, -0.026947021484375, -0.025768280029296875, -0.02458953857421875, -0.023410797119140625, -0.0222320556640625, -0.021053314208984375, -0.01987457275390625, -0.018695831298828125, -0.01751708984375, -0.016338348388671875, -0.01515960693359375, -0.013980865478515625, -0.0128021240234375, -0.011623382568359375, -0.01044464111328125, -0.009265899658203125, -0.008087158203125, -0.006908416748046875, -0.00572967529296875, -0.004550933837890625, -0.0033721923828125, -0.002193450927734375, -0.00101470947265625, 0.000164031982421875, 0.0013427734375, 0.002521514892578125, 0.00370025634765625, 0.004878997802734375, 0.0060577392578125, 0.007236480712890625, 0.00841522216796875, 0.009593963623046875, 0.010772705078125, 0.011951446533203125, 0.01313018798828125, 0.014308929443359375, 0.0154876708984375, 0.016666412353515625, 0.01784515380859375, 0.019023895263671875, 0.02020263671875, 0.021381378173828125, 0.02256011962890625, 0.023738861083984375, 0.0249176025390625, 0.026096343994140625, 0.02727508544921875, 0.028453826904296875, 0.029632568359375, 0.030811309814453125, 0.03199005126953125, 0.033168792724609375, 0.0343475341796875, 0.035526275634765625, 0.03670501708984375, 0.037883758544921875, 0.0390625]}, "gradients/model.layers.31.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 3.0, 3.0, 8.0, 12.0, 9.0, 11.0, 14.0, 18.0, 35.0, 45.0, 61.0, 140.0, 249.0, 602.0, 1439.0, 3906.0, 13089.0, 48393.0, 190572.0, 16777216.0, 16777216.0, 220461.0, 54860.0, 14918.0, 4407.0, 1521.0, 637.0, 275.0, 143.0, 85.0, 54.0, 34.0, 19.0, 18.0, 11.0, 7.0, 1.0, 3.0, 0.0, 6.0, 4.0, 1.0, 0.0, 1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 2.0], "bins": [-0.0106201171875, -0.010298728942871094, -0.009977340698242188, -0.009655952453613281, -0.009334564208984375, -0.009013175964355469, -0.008691787719726562, -0.008370399475097656, -0.00804901123046875, -0.007727622985839844, -0.0074062347412109375, -0.007084846496582031, -0.006763458251953125, -0.006442070007324219, -0.0061206817626953125, -0.005799293518066406, -0.0054779052734375, -0.005156517028808594, -0.0048351287841796875, -0.004513740539550781, -0.004192352294921875, -0.0038709640502929688, -0.0035495758056640625, -0.0032281875610351562, -0.00290679931640625, -0.0025854110717773438, -0.0022640228271484375, -0.0019426345825195312, -0.001621246337890625, -0.0012998580932617188, -0.0009784698486328125, -0.0006570816040039062, -0.000335693359375, -1.430511474609375e-05, 0.0003070831298828125, 0.0006284713745117188, 0.000949859619140625, 0.0012712478637695312, 0.0015926361083984375, 0.0019140243530273438, 0.00223541259765625, 0.0025568008422851562, 0.0028781890869140625, 0.0031995773315429688, 0.003520965576171875, 0.0038423538208007812, 0.0041637420654296875, 0.004485130310058594, 0.0048065185546875, 0.005127906799316406, 0.0054492950439453125, 0.005770683288574219, 0.006092071533203125, 0.006413459777832031, 0.0067348480224609375, 0.007056236267089844, 0.00737762451171875, 0.007699012756347656, 0.008020401000976562, 0.008341789245605469, 0.008663177490234375, 0.008984565734863281, 0.009305953979492188, 0.009627342224121094, 0.00994873046875]}, "gradients/model.layers.31.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 0.0, 0.0, 1.0, 1.0, 2.0, 6.0, 5.0, 9.0, 41.0, 1580.0, 2333.0, 69.0, 15.0, 10.0, 4.0, 1.0, 1.0, 1.0, 0.0, 3.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.010009765625, -0.009487152099609375, -0.00896453857421875, -0.008441925048828125, -0.0079193115234375, -0.007396697998046875, -0.00687408447265625, -0.006351470947265625, -0.005828857421875, -0.005306243896484375, -0.00478363037109375, -0.004261016845703125, -0.0037384033203125, -0.003215789794921875, -0.00269317626953125, -0.002170562744140625, -0.00164794921875, -0.001125335693359375, -0.00060272216796875, -8.0108642578125e-05, 0.0004425048828125, 0.000965118408203125, 0.00148773193359375, 0.002010345458984375, 0.002532958984375, 0.003055572509765625, 0.00357818603515625, 0.004100799560546875, 0.0046234130859375, 0.005146026611328125, 0.00566864013671875, 0.006191253662109375, 0.0067138671875, 0.007236480712890625, 0.00775909423828125, 0.008281707763671875, 0.0088043212890625, 0.009326934814453125, 0.00984954833984375, 0.010372161865234375, 0.010894775390625, 0.011417388916015625, 0.01194000244140625, 0.012462615966796875, 0.0129852294921875, 0.013507843017578125, 0.01403045654296875, 0.014553070068359375, 0.01507568359375, 0.015598297119140625, 0.01612091064453125, 0.016643524169921875, 0.0171661376953125, 0.017688751220703125, 0.01821136474609375, 0.018733978271484375, 0.019256591796875, 0.019779205322265625, 0.02030181884765625, 0.020824432373046875, 0.0213470458984375, 0.021869659423828125, 0.02239227294921875, 0.022914886474609375, 0.0234375]}, "gradients/model.layers.31.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 2.0, 2.0, 3.0, 6.0, 13.0, 28.0, 41.0, 75.0, 162.0, 346.0, 769.0, 2187.0, 7464.0, 71791.0, 16585009.0, 96941.0, 8476.0, 2360.0, 825.0, 347.0, 178.0, 86.0, 50.0, 18.0, 11.0, 5.0, 5.0, 2.0, 2.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.014404296875, -0.014039993286132812, -0.013675689697265625, -0.013311386108398438, -0.01294708251953125, -0.012582778930664062, -0.012218475341796875, -0.011854171752929688, -0.0114898681640625, -0.011125564575195312, -0.010761260986328125, -0.010396957397460938, -0.01003265380859375, -0.009668350219726562, -0.009304046630859375, -0.008939743041992188, -0.008575439453125, -0.008211135864257812, -0.007846832275390625, -0.0074825286865234375, -0.00711822509765625, -0.0067539215087890625, -0.006389617919921875, -0.0060253143310546875, -0.0056610107421875, -0.0052967071533203125, -0.004932403564453125, -0.0045680999755859375, -0.00420379638671875, -0.0038394927978515625, -0.003475189208984375, -0.0031108856201171875, -0.00274658203125, -0.0023822784423828125, -0.002017974853515625, -0.0016536712646484375, -0.00128936767578125, -0.0009250640869140625, -0.000560760498046875, -0.0001964569091796875, 0.0001678466796875, 0.0005321502685546875, 0.000896453857421875, 0.0012607574462890625, 0.00162506103515625, 0.0019893646240234375, 0.002353668212890625, 0.0027179718017578125, 0.003082275390625, 0.0034465789794921875, 0.003810882568359375, 0.0041751861572265625, 0.00453948974609375, 0.0049037933349609375, 0.005268096923828125, 0.0056324005126953125, 0.0059967041015625, 0.0063610076904296875, 0.006725311279296875, 0.0070896148681640625, 0.00745391845703125, 0.007818222045898438, 0.008182525634765625, 0.008546829223632812, 0.0089111328125]}, "gradients/model.layers.31.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 4.0, 6.0, 4.0, 4.0, 19.0, 23.0, 40.0, 73.0, 135.0, 357.0, 1395.0, 21159.0, 4118770.0, 49676.0, 1833.0, 412.0, 159.0, 109.0, 42.0, 26.0, 14.0, 11.0, 7.0, 6.0, 4.0, 4.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.01385498046875, -0.013513565063476562, -0.013172149658203125, -0.012830734252929688, -0.01248931884765625, -0.012147903442382812, -0.011806488037109375, -0.011465072631835938, -0.0111236572265625, -0.010782241821289062, -0.010440826416015625, -0.010099411010742188, -0.00975799560546875, -0.009416580200195312, -0.009075164794921875, -0.008733749389648438, -0.008392333984375, -0.008050918579101562, -0.007709503173828125, -0.0073680877685546875, -0.00702667236328125, -0.0066852569580078125, -0.006343841552734375, -0.0060024261474609375, -0.0056610107421875, -0.0053195953369140625, -0.004978179931640625, -0.0046367645263671875, -0.00429534912109375, -0.0039539337158203125, -0.003612518310546875, -0.0032711029052734375, -0.0029296875, -0.0025882720947265625, -0.002246856689453125, -0.0019054412841796875, -0.00156402587890625, -0.0012226104736328125, -0.000881195068359375, -0.0005397796630859375, -0.0001983642578125, 0.0001430511474609375, 0.000484466552734375, 0.0008258819580078125, 0.00116729736328125, 0.0015087127685546875, 0.001850128173828125, 0.0021915435791015625, 0.002532958984375, 0.0028743743896484375, 0.003215789794921875, 0.0035572052001953125, 0.00389862060546875, 0.0042400360107421875, 0.004581451416015625, 0.0049228668212890625, 0.0052642822265625, 0.0056056976318359375, 0.005947113037109375, 0.0062885284423828125, 0.00662994384765625, 0.0069713592529296875, 0.007312774658203125, 0.0076541900634765625, 0.00799560546875]}, "gradients/model.layers.31.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 2.0, 0.0, 1.0, 1.0, 2.0, 0.0, 2.0, 3.0, 9.0, 11.0, 16.0, 27.0, 31.0, 30.0, 84.0, 152.0, 374.0, 1153.0, 3726.0, 20643.0, 3492196.0, 659287.0, 12305.0, 2741.0, 875.0, 314.0, 119.0, 60.0, 39.0, 33.0, 18.0, 15.0, 10.0, 6.0, 4.0, 2.0, 2.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0022430419921875, -0.0021491050720214844, -0.0020551681518554688, -0.001961231231689453, -0.0018672943115234375, -0.0017733573913574219, -0.0016794204711914062, -0.0015854835510253906, -0.001491546630859375, -0.0013976097106933594, -0.0013036727905273438, -0.0012097358703613281, -0.0011157989501953125, -0.0010218620300292969, -0.0009279251098632812, -0.0008339881896972656, -0.00074005126953125, -0.0006461143493652344, -0.0005521774291992188, -0.0004582405090332031, -0.0003643035888671875, -0.0002703666687011719, -0.00017642974853515625, -8.249282836914062e-05, 1.1444091796875e-05, 0.00010538101196289062, 0.00019931793212890625, 0.0002932548522949219, 0.0003871917724609375, 0.0004811286926269531, 0.0005750656127929688, 0.0006690025329589844, 0.000762939453125, 0.0008568763732910156, 0.0009508132934570312, 0.0010447502136230469, 0.0011386871337890625, 0.0012326240539550781, 0.0013265609741210938, 0.0014204978942871094, 0.001514434814453125, 0.0016083717346191406, 0.0017023086547851562, 0.0017962455749511719, 0.0018901824951171875, 0.001984119415283203, 0.0020780563354492188, 0.0021719932556152344, 0.00226593017578125, 0.0023598670959472656, 0.0024538040161132812, 0.002547740936279297, 0.0026416778564453125, 0.002735614776611328, 0.0028295516967773438, 0.0029234886169433594, 0.003017425537109375, 0.0031113624572753906, 0.0032052993774414062, 0.003299236297607422, 0.0033931732177734375, 0.003487110137939453, 0.0035810470581054688, 0.0036749839782714844, 0.0037689208984375]}, "gradients/model.layers.31.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 3.0, 3.0, 2.0, 9.0, 8.0, 12.0, 12.0, 33.0, 34.0, 51.0, 116.0, 205.0, 481.0, 1299.0, 3807.0, 18413.0, 457533.0, 16198769.0, 83109.0, 9243.0, 2479.0, 860.0, 325.0, 152.0, 82.0, 58.0, 36.0, 20.0, 13.0, 15.0, 10.0, 7.0, 1.0, 1.0, 5.0, 2.0, 2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0019989013671875, -0.001916646957397461, -0.0018343925476074219, -0.0017521381378173828, -0.0016698837280273438, -0.0015876293182373047, -0.0015053749084472656, -0.0014231204986572266, -0.0013408660888671875, -0.0012586116790771484, -0.0011763572692871094, -0.0010941028594970703, -0.0010118484497070312, -0.0009295940399169922, -0.0008473396301269531, -0.0007650852203369141, -0.000682830810546875, -0.0006005764007568359, -0.0005183219909667969, -0.0004360675811767578, -0.00035381317138671875, -0.0002715587615966797, -0.00018930435180664062, -0.00010704994201660156, -2.47955322265625e-05, 5.745887756347656e-05, 0.00013971328735351562, 0.0002219676971435547, 0.00030422210693359375, 0.0003864765167236328, 0.0004687309265136719, 0.0005509853363037109, 0.00063323974609375, 0.0007154941558837891, 0.0007977485656738281, 0.0008800029754638672, 0.0009622573852539062, 0.0010445117950439453, 0.0011267662048339844, 0.0012090206146240234, 0.0012912750244140625, 0.0013735294342041016, 0.0014557838439941406, 0.0015380382537841797, 0.0016202926635742188, 0.0017025470733642578, 0.0017848014831542969, 0.001867055892944336, 0.001949310302734375, 0.002031564712524414, 0.002113819122314453, 0.002196073532104492, 0.0022783279418945312, 0.0023605823516845703, 0.0024428367614746094, 0.0025250911712646484, 0.0026073455810546875, 0.0026895999908447266, 0.0027718544006347656, 0.0028541088104248047, 0.0029363632202148438, 0.003018617630004883, 0.003100872039794922, 0.003183126449584961, 0.003265380859375]}, "gradients/model.layers.31.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 8.0, 2915.0, 1161.0, 6.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.032470703125, -0.031879425048828125, -0.03128814697265625, -0.030696868896484375, -0.0301055908203125, -0.029514312744140625, -0.02892303466796875, -0.028331756591796875, -0.027740478515625, -0.027149200439453125, -0.02655792236328125, -0.025966644287109375, -0.0253753662109375, -0.024784088134765625, -0.02419281005859375, -0.023601531982421875, -0.02301025390625, -0.022418975830078125, -0.02182769775390625, -0.021236419677734375, -0.0206451416015625, -0.020053863525390625, -0.01946258544921875, -0.018871307373046875, -0.018280029296875, -0.017688751220703125, -0.01709747314453125, -0.016506195068359375, -0.0159149169921875, -0.015323638916015625, -0.01473236083984375, -0.014141082763671875, -0.0135498046875, -0.012958526611328125, -0.01236724853515625, -0.011775970458984375, -0.0111846923828125, -0.010593414306640625, -0.01000213623046875, -0.009410858154296875, -0.008819580078125, -0.008228302001953125, -0.00763702392578125, -0.007045745849609375, -0.0064544677734375, -0.005863189697265625, -0.00527191162109375, -0.004680633544921875, -0.00408935546875, -0.003498077392578125, -0.00290679931640625, -0.002315521240234375, -0.0017242431640625, -0.001132965087890625, -0.00054168701171875, 4.9591064453125e-05, 0.000640869140625, 0.001232147216796875, 0.00182342529296875, 0.002414703369140625, 0.0030059814453125, 0.003597259521484375, 0.00418853759765625, 0.004779815673828125, 0.00537109375]}, "gradients/model.layers.30.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 3.0, 2.0, 4.0, 8.0, 12.0, 23.0, 35.0, 54.0, 99.0, 93.0, 140.0, 273.0, 426.0, 1156.0, 7060.0, 112015.0, 16777216.0, 392609.0, 12465.0, 1647.0, 557.0, 292.0, 188.0, 137.0, 84.0, 55.0, 43.0, 26.0, 18.0, 13.0, 10.0, 3.0, 5.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00787353515625, -0.007525444030761719, -0.0071773529052734375, -0.006829261779785156, -0.006481170654296875, -0.006133079528808594, -0.0057849884033203125, -0.005436897277832031, -0.00508880615234375, -0.004740715026855469, -0.0043926239013671875, -0.004044532775878906, -0.003696441650390625, -0.0033483505249023438, -0.0030002593994140625, -0.0026521682739257812, -0.0023040771484375, -0.0019559860229492188, -0.0016078948974609375, -0.0012598037719726562, -0.000911712646484375, -0.0005636215209960938, -0.0002155303955078125, 0.00013256072998046875, 0.00048065185546875, 0.0008287429809570312, 0.0011768341064453125, 0.0015249252319335938, 0.001873016357421875, 0.0022211074829101562, 0.0025691986083984375, 0.0029172897338867188, 0.003265380859375, 0.0036134719848632812, 0.0039615631103515625, 0.004309654235839844, 0.004657745361328125, 0.005005836486816406, 0.0053539276123046875, 0.005702018737792969, 0.00605010986328125, 0.006398200988769531, 0.0067462921142578125, 0.007094383239746094, 0.007442474365234375, 0.007790565490722656, 0.008138656616210938, 0.008486747741699219, 0.0088348388671875, 0.009182929992675781, 0.009531021118164062, 0.009879112243652344, 0.010227203369140625, 0.010575294494628906, 0.010923385620117188, 0.011271476745605469, 0.01161956787109375, 0.011967658996582031, 0.012315750122070312, 0.012663841247558594, 0.013011932373046875, 0.013360023498535156, 0.013708114624023438, 0.014056205749511719, 0.014404296875]}, "gradients/model.layers.30.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 1.0, 0.0, 1.0, 2.0, 2.0, 6.0, 10.0, 12.0, 9.0, 37.0, 71.0, 213.0, 659.0, 2901.0, 40108.0, 16777216.0, 311183.0, 8230.0, 1165.0, 378.0, 110.0, 43.0, 16.0, 24.0, 15.0, 2.0, 2.0, 2.0, 2.0, 1.0, 1.0, 3.0, 1.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.01708984375, -0.016597747802734375, -0.01610565185546875, -0.015613555908203125, -0.0151214599609375, -0.014629364013671875, -0.01413726806640625, -0.013645172119140625, -0.013153076171875, -0.012660980224609375, -0.01216888427734375, -0.011676788330078125, -0.0111846923828125, -0.010692596435546875, -0.01020050048828125, -0.009708404541015625, -0.00921630859375, -0.008724212646484375, -0.00823211669921875, -0.007740020751953125, -0.0072479248046875, -0.006755828857421875, -0.00626373291015625, -0.005771636962890625, -0.005279541015625, -0.004787445068359375, -0.00429534912109375, -0.003803253173828125, -0.0033111572265625, -0.002819061279296875, -0.00232696533203125, -0.001834869384765625, -0.0013427734375, -0.000850677490234375, -0.00035858154296875, 0.000133514404296875, 0.0006256103515625, 0.001117706298828125, 0.00160980224609375, 0.002101898193359375, 0.002593994140625, 0.003086090087890625, 0.00357818603515625, 0.004070281982421875, 0.0045623779296875, 0.005054473876953125, 0.00554656982421875, 0.006038665771484375, 0.00653076171875, 0.007022857666015625, 0.00751495361328125, 0.008007049560546875, 0.0084991455078125, 0.008991241455078125, 0.00948333740234375, 0.009975433349609375, 0.010467529296875, 0.010959625244140625, 0.01145172119140625, 0.011943817138671875, 0.0124359130859375, 0.012928009033203125, 0.01342010498046875, 0.013912200927734375, 0.014404296875]}, "gradients/model.layers.30.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 9.0, 467.0, 3542.0, 55.0, 11.0, 6.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.018310546875, -0.01797318458557129, -0.017635822296142578, -0.017298460006713867, -0.016961097717285156, -0.016623735427856445, -0.016286373138427734, -0.015949010848999023, -0.015611648559570312, -0.015274286270141602, -0.01493692398071289, -0.01459956169128418, -0.014262199401855469, -0.013924837112426758, -0.013587474822998047, -0.013250112533569336, -0.012912750244140625, -0.012575387954711914, -0.012238025665283203, -0.011900663375854492, -0.011563301086425781, -0.01122593879699707, -0.01088857650756836, -0.010551214218139648, -0.010213851928710938, -0.009876489639282227, -0.009539127349853516, -0.009201765060424805, -0.008864402770996094, -0.008527040481567383, -0.008189678192138672, -0.007852315902709961, -0.00751495361328125, -0.007177591323852539, -0.006840229034423828, -0.006502866744995117, -0.006165504455566406, -0.005828142166137695, -0.005490779876708984, -0.0051534175872802734, -0.0048160552978515625, -0.0044786930084228516, -0.004141330718994141, -0.0038039684295654297, -0.0034666061401367188, -0.003129243850708008, -0.002791881561279297, -0.002454519271850586, -0.002117156982421875, -0.001779794692993164, -0.0014424324035644531, -0.0011050701141357422, -0.0007677078247070312, -0.0004303455352783203, -9.298324584960938e-05, 0.00024437904357910156, 0.0005817413330078125, 0.0009191036224365234, 0.0012564659118652344, 0.0015938282012939453, 0.0019311904907226562, 0.002268552780151367, 0.002605915069580078, 0.002943277359008789, 0.0032806396484375]}, "gradients/model.layers.30.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 4.0, 4.0, 11.0, 10.0, 9.0, 27.0, 54.0, 82.0, 121.0, 224.0, 361.0, 711.0, 1438.0, 3211.0, 9496.0, 38471.0, 247152.0, 3180706.0, 11471838.0, 1637164.0, 148025.0, 26085.0, 6999.0, 2533.0, 1167.0, 545.0, 334.0, 176.0, 116.0, 48.0, 32.0, 18.0, 11.0, 7.0, 5.0, 3.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.00131988525390625, -0.0012791156768798828, -0.0012383460998535156, -0.0011975765228271484, -0.0011568069458007812, -0.001116037368774414, -0.0010752677917480469, -0.0010344982147216797, -0.0009937286376953125, -0.0009529590606689453, -0.0009121894836425781, -0.0008714199066162109, -0.0008306503295898438, -0.0007898807525634766, -0.0007491111755371094, -0.0007083415985107422, -0.000667572021484375, -0.0006268024444580078, -0.0005860328674316406, -0.0005452632904052734, -0.0005044937133789062, -0.00046372413635253906, -0.0004229545593261719, -0.0003821849822998047, -0.0003414154052734375, -0.0003006458282470703, -0.0002598762512207031, -0.00021910667419433594, -0.00017833709716796875, -0.00013756752014160156, -9.679794311523438e-05, -5.602836608886719e-05, -1.52587890625e-05, 2.5510787963867188e-05, 6.628036499023438e-05, 0.00010704994201660156, 0.00014781951904296875, 0.00018858909606933594, 0.00022935867309570312, 0.0002701282501220703, 0.0003108978271484375, 0.0003516674041748047, 0.0003924369812011719, 0.00043320655822753906, 0.00047397613525390625, 0.0005147457122802734, 0.0005555152893066406, 0.0005962848663330078, 0.000637054443359375, 0.0006778240203857422, 0.0007185935974121094, 0.0007593631744384766, 0.0008001327514648438, 0.0008409023284912109, 0.0008816719055175781, 0.0009224414825439453, 0.0009632110595703125, 0.0010039806365966797, 0.0010447502136230469, 0.001085519790649414, 0.0011262893676757812, 0.0011670589447021484, 0.0012078285217285156, 0.0012485980987548828, 0.00128936767578125]}, "gradients/model.layers.30.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 3.0, 2.0, 7.0, 3.0, 6.0, 2.0, 8.0, 17.0, 14.0, 27.0, 34.0, 46.0, 61.0, 114.0, 149.0, 227.0, 356.0, 633.0, 1324.0, 3442.0, 12726.0, 724709.0, 3411229.0, 29923.0, 5221.0, 1848.0, 822.0, 455.0, 297.0, 193.0, 117.0, 82.0, 66.0, 33.0, 30.0, 18.0, 13.0, 10.0, 11.0, 8.0, 6.0, 2.0, 0.0, 2.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00347900390625, -0.0033597946166992188, -0.0032405853271484375, -0.0031213760375976562, -0.003002166748046875, -0.0028829574584960938, -0.0027637481689453125, -0.0026445388793945312, -0.00252532958984375, -0.0024061203002929688, -0.0022869110107421875, -0.0021677017211914062, -0.002048492431640625, -0.0019292831420898438, -0.0018100738525390625, -0.0016908645629882812, -0.0015716552734375, -0.0014524459838867188, -0.0013332366943359375, -0.0012140274047851562, -0.001094818115234375, -0.0009756088256835938, -0.0008563995361328125, -0.0007371902465820312, -0.00061798095703125, -0.0004987716674804688, -0.0003795623779296875, -0.00026035308837890625, -0.000141143798828125, -2.193450927734375e-05, 9.72747802734375e-05, 0.00021648406982421875, 0.000335693359375, 0.00045490264892578125, 0.0005741119384765625, 0.0006933212280273438, 0.000812530517578125, 0.0009317398071289062, 0.0010509490966796875, 0.0011701583862304688, 0.00128936767578125, 0.0014085769653320312, 0.0015277862548828125, 0.0016469955444335938, 0.001766204833984375, 0.0018854141235351562, 0.0020046234130859375, 0.0021238327026367188, 0.0022430419921875, 0.0023622512817382812, 0.0024814605712890625, 0.0026006698608398438, 0.002719879150390625, 0.0028390884399414062, 0.0029582977294921875, 0.0030775070190429688, 0.00319671630859375, 0.0033159255981445312, 0.0034351348876953125, 0.0035543441772460938, 0.003673553466796875, 0.0037927627563476562, 0.0039119720458984375, 0.004031181335449219, 0.004150390625]}, "gradients/model.layers.30.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 5.0, 3.0, 7.0, 5.0, 16.0, 7.0, 17.0, 17.0, 27.0, 26.0, 34.0, 52.0, 67.0, 102.0, 237.0, 396.0, 1109.0, 2880.0, 11080.0, 109515.0, 3969348.0, 84571.0, 10022.0, 2793.0, 944.0, 402.0, 184.0, 122.0, 60.0, 64.0, 40.0, 34.0, 20.0, 20.0, 20.0, 7.0, 11.0, 8.0, 4.0, 2.0, 3.0, 1.0, 2.0, 1.0, 2.0, 2.0, 1.0, 2.0, 1.0, 2.0, 2.0, 1.0], "bins": [-0.00159454345703125, -0.0015454292297363281, -0.0014963150024414062, -0.0014472007751464844, -0.0013980865478515625, -0.0013489723205566406, -0.0012998580932617188, -0.0012507438659667969, -0.001201629638671875, -0.0011525154113769531, -0.0011034011840820312, -0.0010542869567871094, -0.0010051727294921875, -0.0009560585021972656, -0.0009069442749023438, -0.0008578300476074219, -0.0008087158203125, -0.0007596015930175781, -0.0007104873657226562, -0.0006613731384277344, -0.0006122589111328125, -0.0005631446838378906, -0.0005140304565429688, -0.0004649162292480469, -0.000415802001953125, -0.0003666877746582031, -0.00031757354736328125, -0.0002684593200683594, -0.0002193450927734375, -0.00017023086547851562, -0.00012111663818359375, -7.200241088867188e-05, -2.288818359375e-05, 2.6226043701171875e-05, 7.534027099609375e-05, 0.00012445449829101562, 0.0001735687255859375, 0.00022268295288085938, 0.00027179718017578125, 0.0003209114074707031, 0.000370025634765625, 0.0004191398620605469, 0.00046825408935546875, 0.0005173683166503906, 0.0005664825439453125, 0.0006155967712402344, 0.0006647109985351562, 0.0007138252258300781, 0.000762939453125, 0.0008120536804199219, 0.0008611679077148438, 0.0009102821350097656, 0.0009593963623046875, 0.0010085105895996094, 0.0010576248168945312, 0.0011067390441894531, 0.001155853271484375, 0.0012049674987792969, 0.0012540817260742188, 0.0013031959533691406, 0.0013523101806640625, 0.0014014244079589844, 0.0014505386352539062, 0.0014996528625488281, 0.00154876708984375]}, "gradients/model.layers.30.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 1.0, 3.0, 0.0, 3.0, 3.0, 0.0, 4.0, 5.0, 9.0, 10.0, 10.0, 9.0, 22.0, 20.0, 37.0, 58.0, 70.0, 129.0, 258.0, 600.0, 1417.0, 4336.0, 18310.0, 222105.0, 16106319.0, 392214.0, 23040.0, 5098.0, 1650.0, 695.0, 326.0, 136.0, 101.0, 76.0, 41.0, 23.0, 13.0, 14.0, 14.0, 2.0, 5.0, 7.0, 1.0, 3.0, 0.0, 3.0, 4.0, 0.0, 0.0, 2.0, 2.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0], "bins": [-0.0013275146484375, -0.0012793540954589844, -0.0012311935424804688, -0.0011830329895019531, -0.0011348724365234375, -0.0010867118835449219, -0.0010385513305664062, -0.0009903907775878906, -0.000942230224609375, -0.0008940696716308594, -0.0008459091186523438, -0.0007977485656738281, -0.0007495880126953125, -0.0007014274597167969, -0.0006532669067382812, -0.0006051063537597656, -0.00055694580078125, -0.0005087852478027344, -0.00046062469482421875, -0.0004124641418457031, -0.0003643035888671875, -0.0003161430358886719, -0.00026798248291015625, -0.00021982192993164062, -0.000171661376953125, -0.00012350082397460938, -7.534027099609375e-05, -2.7179718017578125e-05, 2.09808349609375e-05, 6.914138793945312e-05, 0.00011730194091796875, 0.00016546249389648438, 0.000213623046875, 0.0002617835998535156, 0.00030994415283203125, 0.0003581047058105469, 0.0004062652587890625, 0.0004544258117675781, 0.0005025863647460938, 0.0005507469177246094, 0.000598907470703125, 0.0006470680236816406, 0.0006952285766601562, 0.0007433891296386719, 0.0007915496826171875, 0.0008397102355957031, 0.0008878707885742188, 0.0009360313415527344, 0.00098419189453125, 0.0010323524475097656, 0.0010805130004882812, 0.0011286735534667969, 0.0011768341064453125, 0.0012249946594238281, 0.0012731552124023438, 0.0013213157653808594, 0.001369476318359375, 0.0014176368713378906, 0.0014657974243164062, 0.0015139579772949219, 0.0015621185302734375, 0.0016102790832519531, 0.0016584396362304688, 0.0017066001892089844, 0.0017547607421875]}, "gradients/model.layers.30.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 4.0, 14.0, 31.0, 639.0, 3081.0, 300.0, 15.0, 5.0, 1.0, 3.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00127410888671875, -0.0011855363845825195, -0.001096963882446289, -0.0010083913803100586, -0.0009198188781738281, -0.0008312463760375977, -0.0007426738739013672, -0.0006541013717651367, -0.0005655288696289062, -0.0004769563674926758, -0.0003883838653564453, -0.00029981136322021484, -0.00021123886108398438, -0.0001226663589477539, -3.409385681152344e-05, 5.447864532470703e-05, 0.0001430511474609375, 0.00023162364959716797, 0.00032019615173339844, 0.0004087686538696289, 0.0004973411560058594, 0.0005859136581420898, 0.0006744861602783203, 0.0007630586624145508, 0.0008516311645507812, 0.0009402036666870117, 0.0010287761688232422, 0.0011173486709594727, 0.0012059211730957031, 0.0012944936752319336, 0.001383066177368164, 0.0014716386795043945, 0.001560211181640625, 0.0016487836837768555, 0.001737356185913086, 0.0018259286880493164, 0.0019145011901855469, 0.0020030736923217773, 0.002091646194458008, 0.0021802186965942383, 0.0022687911987304688, 0.0023573637008666992, 0.0024459362030029297, 0.00253450870513916, 0.0026230812072753906, 0.002711653709411621, 0.0028002262115478516, 0.002888798713684082, 0.0029773712158203125, 0.003065943717956543, 0.0031545162200927734, 0.003243088722229004, 0.0033316612243652344, 0.003420233726501465, 0.0035088062286376953, 0.0035973787307739258, 0.0036859512329101562, 0.0037745237350463867, 0.003863096237182617, 0.003951668739318848, 0.004040241241455078, 0.004128813743591309, 0.004217386245727539, 0.0043059587478637695, 0.00439453125]}, "gradients/model.layers.29.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 0.0, 1.0, 2.0, 1.0, 3.0, 3.0, 10.0, 22.0, 46.0, 92.0, 205.0, 385.0, 901.0, 2151.0, 6041.0, 41168.0, 3478111.0, 16777216.0, 182469.0, 11651.0, 3243.0, 1286.0, 571.0, 259.0, 119.0, 53.0, 31.0, 12.0, 4.0, 5.0, 5.0, 0.0, 2.0, 2.0, 2.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00494384765625, -0.004799365997314453, -0.004654884338378906, -0.004510402679443359, -0.0043659210205078125, -0.004221439361572266, -0.004076957702636719, -0.003932476043701172, -0.003787994384765625, -0.003643512725830078, -0.0034990310668945312, -0.0033545494079589844, -0.0032100677490234375, -0.0030655860900878906, -0.0029211044311523438, -0.002776622772216797, -0.00263214111328125, -0.002487659454345703, -0.0023431777954101562, -0.0021986961364746094, -0.0020542144775390625, -0.0019097328186035156, -0.0017652511596679688, -0.0016207695007324219, -0.001476287841796875, -0.0013318061828613281, -0.0011873245239257812, -0.0010428428649902344, -0.0008983612060546875, -0.0007538795471191406, -0.0006093978881835938, -0.0004649162292480469, -0.0003204345703125, -0.00017595291137695312, -3.147125244140625e-05, 0.00011301040649414062, 0.0002574920654296875, 0.0004019737243652344, 0.0005464553833007812, 0.0006909370422363281, 0.000835418701171875, 0.0009799003601074219, 0.0011243820190429688, 0.0012688636779785156, 0.0014133453369140625, 0.0015578269958496094, 0.0017023086547851562, 0.0018467903137207031, 0.00199127197265625, 0.002135753631591797, 0.0022802352905273438, 0.0024247169494628906, 0.0025691986083984375, 0.0027136802673339844, 0.0028581619262695312, 0.003002643585205078, 0.003147125244140625, 0.003291606903076172, 0.0034360885620117188, 0.0035805702209472656, 0.0037250518798828125, 0.0038695335388183594, 0.004014015197753906, 0.004158496856689453, 0.004302978515625]}, "gradients/model.layers.29.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 3.0, 2.0, 3.0, 5.0, 8.0, 9.0, 10.0, 26.0, 20.0, 60.0, 147.0, 375.0, 1131.0, 5024.0, 44157.0, 16777216.0, 13428132.0, 30103.0, 3928.0, 1015.0, 288.0, 146.0, 57.0, 30.0, 19.0, 10.0, 4.0, 8.0, 1.0, 3.0, 2.0, 4.0, 2.0, 0.0, 2.0, 2.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.006591796875, -0.006378650665283203, -0.006165504455566406, -0.005952358245849609, -0.0057392120361328125, -0.005526065826416016, -0.005312919616699219, -0.005099773406982422, -0.004886627197265625, -0.004673480987548828, -0.004460334777832031, -0.004247188568115234, -0.0040340423583984375, -0.0038208961486816406, -0.0036077499389648438, -0.003394603729248047, -0.00318145751953125, -0.002968311309814453, -0.0027551651000976562, -0.0025420188903808594, -0.0023288726806640625, -0.0021157264709472656, -0.0019025802612304688, -0.0016894340515136719, -0.001476287841796875, -0.0012631416320800781, -0.0010499954223632812, -0.0008368492126464844, -0.0006237030029296875, -0.0004105567932128906, -0.00019741058349609375, 1.5735626220703125e-05, 0.0002288818359375, 0.0004420280456542969, 0.0006551742553710938, 0.0008683204650878906, 0.0010814666748046875, 0.0012946128845214844, 0.0015077590942382812, 0.0017209053039550781, 0.001934051513671875, 0.002147197723388672, 0.0023603439331054688, 0.0025734901428222656, 0.0027866363525390625, 0.0029997825622558594, 0.0032129287719726562, 0.003426074981689453, 0.00363922119140625, 0.003852367401123047, 0.004065513610839844, 0.004278659820556641, 0.0044918060302734375, 0.004704952239990234, 0.004918098449707031, 0.005131244659423828, 0.005344390869140625, 0.005557537078857422, 0.005770683288574219, 0.005983829498291016, 0.0061969757080078125, 0.006410121917724609, 0.006623268127441406, 0.006836414337158203, 0.007049560546875]}, "gradients/model.layers.29.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 0.0, 1.0, 1.0, 2.0, 7.0, 32.0, 173.0, 638.0, 1326.0, 1175.0, 544.0, 141.0, 24.0, 10.0, 8.0, 4.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.001007080078125, -0.0009505748748779297, -0.0008940696716308594, -0.0008375644683837891, -0.0007810592651367188, -0.0007245540618896484, -0.0006680488586425781, -0.0006115436553955078, -0.0005550384521484375, -0.0004985332489013672, -0.0004420280456542969, -0.00038552284240722656, -0.00032901763916015625, -0.00027251243591308594, -0.00021600723266601562, -0.0001595020294189453, -0.000102996826171875, -4.649162292480469e-05, 1.0013580322265625e-05, 6.651878356933594e-05, 0.00012302398681640625, 0.00017952919006347656, 0.00023603439331054688, 0.0002925395965576172, 0.0003490447998046875, 0.0004055500030517578, 0.0004620552062988281, 0.0005185604095458984, 0.0005750656127929688, 0.0006315708160400391, 0.0006880760192871094, 0.0007445812225341797, 0.00080108642578125, 0.0008575916290283203, 0.0009140968322753906, 0.0009706020355224609, 0.0010271072387695312, 0.0010836124420166016, 0.0011401176452636719, 0.0011966228485107422, 0.0012531280517578125, 0.0013096332550048828, 0.0013661384582519531, 0.0014226436614990234, 0.0014791488647460938, 0.001535654067993164, 0.0015921592712402344, 0.0016486644744873047, 0.001705169677734375, 0.0017616748809814453, 0.0018181800842285156, 0.001874685287475586, 0.0019311904907226562, 0.0019876956939697266, 0.002044200897216797, 0.002100706100463867, 0.0021572113037109375, 0.002213716506958008, 0.002270221710205078, 0.0023267269134521484, 0.0023832321166992188, 0.002439737319946289, 0.0024962425231933594, 0.0025527477264404297, 0.0026092529296875]}, "gradients/model.layers.29.self_attn.o_proj.weight": {"_type": "histogram", "values": [2.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 3.0, 2.0, 2.0, 3.0, 8.0, 10.0, 9.0, 17.0, 28.0, 63.0, 76.0, 153.0, 272.0, 464.0, 805.0, 1484.0, 3072.0, 6781.0, 26861.0, 369925.0, 14394284.0, 1882925.0, 69823.0, 11432.0, 4105.0, 1983.0, 1126.0, 592.0, 381.0, 206.0, 116.0, 80.0, 43.0, 29.0, 18.0, 11.0, 5.0, 3.0, 4.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0016021728515625, -0.0015463829040527344, -0.0014905929565429688, -0.0014348030090332031, -0.0013790130615234375, -0.0013232231140136719, -0.0012674331665039062, -0.0012116432189941406, -0.001155853271484375, -0.0011000633239746094, -0.0010442733764648438, -0.0009884834289550781, -0.0009326934814453125, -0.0008769035339355469, -0.0008211135864257812, -0.0007653236389160156, -0.00070953369140625, -0.0006537437438964844, -0.0005979537963867188, -0.0005421638488769531, -0.0004863739013671875, -0.0004305839538574219, -0.00037479400634765625, -0.0003190040588378906, -0.000263214111328125, -0.00020742416381835938, -0.00015163421630859375, -9.584426879882812e-05, -4.00543212890625e-05, 1.5735626220703125e-05, 7.152557373046875e-05, 0.00012731552124023438, 0.00018310546875, 0.00023889541625976562, 0.00029468536376953125, 0.0003504753112792969, 0.0004062652587890625, 0.0004620552062988281, 0.0005178451538085938, 0.0005736351013183594, 0.000629425048828125, 0.0006852149963378906, 0.0007410049438476562, 0.0007967948913574219, 0.0008525848388671875, 0.0009083747863769531, 0.0009641647338867188, 0.0010199546813964844, 0.00107574462890625, 0.0011315345764160156, 0.0011873245239257812, 0.0012431144714355469, 0.0012989044189453125, 0.0013546943664550781, 0.0014104843139648438, 0.0014662742614746094, 0.001522064208984375, 0.0015778541564941406, 0.0016336441040039062, 0.0016894340515136719, 0.0017452239990234375, 0.0018010139465332031, 0.0018568038940429688, 0.0019125938415527344, 0.0019683837890625]}, "gradients/model.layers.29.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 3.0, 7.0, 7.0, 5.0, 9.0, 13.0, 15.0, 33.0, 28.0, 41.0, 72.0, 69.0, 82.0, 135.0, 187.0, 264.0, 480.0, 1310.0, 5966.0, 498327.0, 3670253.0, 13103.0, 2134.0, 629.0, 345.0, 217.0, 132.0, 116.0, 78.0, 57.0, 37.0, 34.0, 30.0, 14.0, 15.0, 20.0, 10.0, 6.0, 4.0, 3.0, 0.0, 4.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0037841796875, -0.00365447998046875, -0.0035247802734375, -0.00339508056640625, -0.003265380859375, -0.00313568115234375, -0.0030059814453125, -0.00287628173828125, -0.00274658203125, -0.00261688232421875, -0.0024871826171875, -0.00235748291015625, -0.002227783203125, -0.00209808349609375, -0.0019683837890625, -0.00183868408203125, -0.001708984375, -0.00157928466796875, -0.0014495849609375, -0.00131988525390625, -0.001190185546875, -0.00106048583984375, -0.0009307861328125, -0.00080108642578125, -0.00067138671875, -0.00054168701171875, -0.0004119873046875, -0.00028228759765625, -0.000152587890625, -2.288818359375e-05, 0.0001068115234375, 0.00023651123046875, 0.0003662109375, 0.00049591064453125, 0.0006256103515625, 0.00075531005859375, 0.000885009765625, 0.00101470947265625, 0.0011444091796875, 0.00127410888671875, 0.00140380859375, 0.00153350830078125, 0.0016632080078125, 0.00179290771484375, 0.001922607421875, 0.00205230712890625, 0.0021820068359375, 0.00231170654296875, 0.00244140625, 0.00257110595703125, 0.0027008056640625, 0.00283050537109375, 0.002960205078125, 0.00308990478515625, 0.0032196044921875, 0.00334930419921875, 0.00347900390625, 0.00360870361328125, 0.0037384033203125, 0.00386810302734375, 0.003997802734375, 0.00412750244140625, 0.0042572021484375, 0.00438690185546875, 0.0045166015625]}, "gradients/model.layers.29.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 3.0, 2.0, 5.0, 5.0, 12.0, 20.0, 37.0, 57.0, 119.0, 218.0, 571.0, 1885.0, 9897.0, 437227.0, 3719932.0, 19996.0, 2915.0, 780.0, 305.0, 133.0, 78.0, 42.0, 21.0, 12.0, 4.0, 6.0, 2.0, 3.0, 2.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00173187255859375, -0.0016744136810302734, -0.0016169548034667969, -0.0015594959259033203, -0.0015020370483398438, -0.0014445781707763672, -0.0013871192932128906, -0.001329660415649414, -0.0012722015380859375, -0.001214742660522461, -0.0011572837829589844, -0.0010998249053955078, -0.0010423660278320312, -0.0009849071502685547, -0.0009274482727050781, -0.0008699893951416016, -0.000812530517578125, -0.0007550716400146484, -0.0006976127624511719, -0.0006401538848876953, -0.0005826950073242188, -0.0005252361297607422, -0.0004677772521972656, -0.00041031837463378906, -0.0003528594970703125, -0.00029540061950683594, -0.00023794174194335938, -0.0001804828643798828, -0.00012302398681640625, -6.556510925292969e-05, -8.106231689453125e-06, 4.935264587402344e-05, 0.0001068115234375, 0.00016427040100097656, 0.00022172927856445312, 0.0002791881561279297, 0.00033664703369140625, 0.0003941059112548828, 0.0004515647888183594, 0.0005090236663818359, 0.0005664825439453125, 0.0006239414215087891, 0.0006814002990722656, 0.0007388591766357422, 0.0007963180541992188, 0.0008537769317626953, 0.0009112358093261719, 0.0009686946868896484, 0.001026153564453125, 0.0010836124420166016, 0.0011410713195800781, 0.0011985301971435547, 0.0012559890747070312, 0.0013134479522705078, 0.0013709068298339844, 0.001428365707397461, 0.0014858245849609375, 0.001543283462524414, 0.0016007423400878906, 0.0016582012176513672, 0.0017156600952148438, 0.0017731189727783203, 0.0018305778503417969, 0.0018880367279052734, 0.00194549560546875]}, "gradients/model.layers.29.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 0.0, 5.0, 2.0, 2.0, 4.0, 2.0, 6.0, 10.0, 7.0, 19.0, 20.0, 43.0, 92.0, 322.0, 992.0, 5043.0, 16761049.0, 7786.0, 1200.0, 361.0, 119.0, 36.0, 19.0, 18.0, 11.0, 9.0, 7.0, 3.0, 5.0, 3.0, 0.0, 5.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.024658203125, -0.023944854736328125, -0.02323150634765625, -0.022518157958984375, -0.0218048095703125, -0.021091461181640625, -0.02037811279296875, -0.019664764404296875, -0.018951416015625, -0.018238067626953125, -0.01752471923828125, -0.016811370849609375, -0.0160980224609375, -0.015384674072265625, -0.01467132568359375, -0.013957977294921875, -0.01324462890625, -0.012531280517578125, -0.01181793212890625, -0.011104583740234375, -0.0103912353515625, -0.009677886962890625, -0.00896453857421875, -0.008251190185546875, -0.007537841796875, -0.006824493408203125, -0.00611114501953125, -0.005397796630859375, -0.0046844482421875, -0.003971099853515625, -0.00325775146484375, -0.002544403076171875, -0.0018310546875, -0.001117706298828125, -0.00040435791015625, 0.000308990478515625, 0.0010223388671875, 0.001735687255859375, 0.00244903564453125, 0.003162384033203125, 0.003875732421875, 0.004589080810546875, 0.00530242919921875, 0.006015777587890625, 0.0067291259765625, 0.007442474365234375, 0.00815582275390625, 0.008869171142578125, 0.00958251953125, 0.010295867919921875, 0.01100921630859375, 0.011722564697265625, 0.0124359130859375, 0.013149261474609375, 0.01386260986328125, 0.014575958251953125, 0.015289306640625, 0.016002655029296875, 0.01671600341796875, 0.017429351806640625, 0.0181427001953125, 0.018856048583984375, 0.01956939697265625, 0.020282745361328125, 0.02099609375]}, "gradients/model.layers.29.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 5.0, 139.0, 3928.0, 19.0], "bins": [-0.0087890625, -0.008648142218589783, -0.008507221937179565, -0.008366301655769348, -0.00822538137435913, -0.008084461092948914, -0.007943540811538696, -0.007802620530128479, -0.007661700248718262, -0.007520779967308044, -0.007379859685897827, -0.00723893940448761, -0.007098019123077393, -0.006957098841667175, -0.006816178560256958, -0.006675258278846741, -0.0065343379974365234, -0.006393417716026306, -0.006252497434616089, -0.006111577153205872, -0.005970656871795654, -0.005829736590385437, -0.00568881630897522, -0.0055478960275650024, -0.005406975746154785, -0.005266055464744568, -0.005125135183334351, -0.004984214901924133, -0.004843294620513916, -0.004702374339103699, -0.0045614540576934814, -0.004420533776283264, -0.004279613494873047, -0.00413869321346283, -0.003997772932052612, -0.003856852650642395, -0.0037159323692321777, -0.0035750120878219604, -0.003434091806411743, -0.003293171525001526, -0.0031522512435913086, -0.0030113309621810913, -0.002870410680770874, -0.0027294903993606567, -0.0025885701179504395, -0.002447649836540222, -0.002306729555130005, -0.0021658092737197876, -0.0020248889923095703, -0.001883968710899353, -0.0017430484294891357, -0.0016021281480789185, -0.0014612078666687012, -0.0013202875852584839, -0.0011793673038482666, -0.0010384470224380493, -0.000897526741027832, -0.0007566064596176147, -0.0006156861782073975, -0.0004747658967971802, -0.0003338456153869629, -0.0001929253339767456, -5.200505256652832e-05, 8.891522884368896e-05, 0.00022983551025390625]}, "gradients/model.layers.28.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 2.0, 5.0, 7.0, 3.0, 8.0, 18.0, 18.0, 19.0, 31.0, 55.0, 64.0, 105.0, 186.0, 292.0, 533.0, 988.0, 1840.0, 3937.0, 8604.0, 21654.0, 59617.0, 187748.0, 765226.0, 11322629.0, 16777216.0, 5024159.0, 504457.0, 138241.0, 46559.0, 16921.0, 7086.0, 3248.0, 1571.0, 829.0, 464.0, 256.0, 143.0, 107.0, 63.0, 56.0, 21.0, 19.0, 20.0, 9.0, 6.0, 9.0, 4.0, 1.0, 5.0, 3.0, 2.0, 0.0, 0.0, 1.0, 2.0, 0.0, 1.0], "bins": [-0.00113677978515625, -0.0011005401611328125, -0.001064300537109375, -0.0010280609130859375, -0.0009918212890625, -0.0009555816650390625, -0.000919342041015625, -0.0008831024169921875, -0.00084686279296875, -0.0008106231689453125, -0.000774383544921875, -0.0007381439208984375, -0.000701904296875, -0.0006656646728515625, -0.000629425048828125, -0.0005931854248046875, -0.00055694580078125, -0.0005207061767578125, -0.000484466552734375, -0.0004482269287109375, -0.0004119873046875, -0.0003757476806640625, -0.000339508056640625, -0.0003032684326171875, -0.00026702880859375, -0.0002307891845703125, -0.000194549560546875, -0.0001583099365234375, -0.0001220703125, -8.58306884765625e-05, -4.9591064453125e-05, -1.33514404296875e-05, 2.288818359375e-05, 5.91278076171875e-05, 9.5367431640625e-05, 0.0001316070556640625, 0.0001678466796875, 0.0002040863037109375, 0.000240325927734375, 0.0002765655517578125, 0.00031280517578125, 0.0003490447998046875, 0.000385284423828125, 0.0004215240478515625, 0.000457763671875, 0.0004940032958984375, 0.000530242919921875, 0.0005664825439453125, 0.00060272216796875, 0.0006389617919921875, 0.000675201416015625, 0.0007114410400390625, 0.0007476806640625, 0.0007839202880859375, 0.000820159912109375, 0.0008563995361328125, 0.00089263916015625, 0.0009288787841796875, 0.000965118408203125, 0.0010013580322265625, 0.00103759765625, 0.0010738372802734375, 0.001110076904296875, 0.0011463165283203125, 0.00118255615234375]}, "gradients/model.layers.28.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 3.0, 0.0, 1.0, 2.0, 3.0, 4.0, 4.0, 13.0, 11.0, 21.0, 27.0, 62.0, 76.0, 196.0, 648.0, 3481.0, 37123.0, 16777216.0, 16777216.0, 48022.0, 3925.0, 665.0, 238.0, 109.0, 43.0, 32.0, 20.0, 11.0, 4.0, 3.0, 2.0, 4.0, 3.0, 2.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.004241943359375, -0.004085063934326172, -0.003928184509277344, -0.0037713050842285156, -0.0036144256591796875, -0.0034575462341308594, -0.0033006668090820312, -0.003143787384033203, -0.002986907958984375, -0.002830028533935547, -0.0026731491088867188, -0.0025162696838378906, -0.0023593902587890625, -0.0022025108337402344, -0.0020456314086914062, -0.0018887519836425781, -0.00173187255859375, -0.0015749931335449219, -0.0014181137084960938, -0.0012612342834472656, -0.0011043548583984375, -0.0009474754333496094, -0.0007905960083007812, -0.0006337165832519531, -0.000476837158203125, -0.0003199577331542969, -0.00016307830810546875, -6.198883056640625e-06, 0.0001506805419921875, 0.0003075599670410156, 0.00046443939208984375, 0.0006213188171386719, 0.0007781982421875, 0.0009350776672363281, 0.0010919570922851562, 0.0012488365173339844, 0.0014057159423828125, 0.0015625953674316406, 0.0017194747924804688, 0.0018763542175292969, 0.002033233642578125, 0.002190113067626953, 0.0023469924926757812, 0.0025038719177246094, 0.0026607513427734375, 0.0028176307678222656, 0.0029745101928710938, 0.003131389617919922, 0.00328826904296875, 0.003445148468017578, 0.0036020278930664062, 0.0037589073181152344, 0.0039157867431640625, 0.004072666168212891, 0.004229545593261719, 0.004386425018310547, 0.004543304443359375, 0.004700183868408203, 0.004857063293457031, 0.005013942718505859, 0.0051708221435546875, 0.005327701568603516, 0.005484580993652344, 0.005641460418701172, 0.00579833984375]}, "gradients/model.layers.28.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 1.0, 1.0, 3.0, 3.0, 7.0, 13.0, 35.0, 119.0, 272.0, 491.0, 758.0, 812.0, 702.0, 451.0, 234.0, 99.0, 48.0, 21.0, 7.0, 3.0, 0.0, 1.0, 2.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.000858306884765625, -0.0008290410041809082, -0.0007997751235961914, -0.0007705092430114746, -0.0007412433624267578, -0.000711977481842041, -0.0006827116012573242, -0.0006534457206726074, -0.0006241798400878906, -0.0005949139595031738, -0.000565648078918457, -0.0005363821983337402, -0.0005071163177490234, -0.00047785043716430664, -0.00044858455657958984, -0.00041931867599487305, -0.00039005279541015625, -0.00036078691482543945, -0.00033152103424072266, -0.00030225515365600586, -0.00027298927307128906, -0.00024372339248657227, -0.00021445751190185547, -0.00018519163131713867, -0.00015592575073242188, -0.00012665987014770508, -9.739398956298828e-05, -6.812810897827148e-05, -3.886222839355469e-05, -9.59634780883789e-06, 1.9669532775878906e-05, 4.89354133605957e-05, 7.82012939453125e-05, 0.0001074671745300293, 0.0001367330551147461, 0.0001659989356994629, 0.0001952648162841797, 0.00022453069686889648, 0.0002537965774536133, 0.0002830624580383301, 0.0003123283386230469, 0.00034159421920776367, 0.00037086009979248047, 0.00040012598037719727, 0.00042939186096191406, 0.00045865774154663086, 0.00048792362213134766, 0.0005171895027160645, 0.0005464553833007812, 0.000575721263885498, 0.0006049871444702148, 0.0006342530250549316, 0.0006635189056396484, 0.0006927847862243652, 0.000722050666809082, 0.0007513165473937988, 0.0007805824279785156, 0.0008098483085632324, 0.0008391141891479492, 0.000868380069732666, 0.0008976459503173828, 0.0009269118309020996, 0.0009561777114868164, 0.0009854435920715332, 0.00101470947265625]}, "gradients/model.layers.28.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 3.0, 0.0, 1.0, 3.0, 2.0, 1.0, 8.0, 7.0, 17.0, 26.0, 55.0, 109.0, 181.0, 298.0, 659.0, 1106.0, 2050.0, 4321.0, 10474.0, 49759.0, 415796.0, 4254273.0, 10586931.0, 1284918.0, 133319.0, 20698.0, 5933.0, 2948.0, 1419.0, 838.0, 520.0, 243.0, 133.0, 73.0, 32.0, 17.0, 12.0, 7.0, 9.0, 4.0, 1.0, 1.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.0013275146484375, -0.0012918710708618164, -0.0012562274932861328, -0.0012205839157104492, -0.0011849403381347656, -0.001149296760559082, -0.0011136531829833984, -0.0010780096054077148, -0.0010423660278320312, -0.0010067224502563477, -0.0009710788726806641, -0.0009354352951049805, -0.0008997917175292969, -0.0008641481399536133, -0.0008285045623779297, -0.0007928609848022461, -0.0007572174072265625, -0.0007215738296508789, -0.0006859302520751953, -0.0006502866744995117, -0.0006146430969238281, -0.0005789995193481445, -0.0005433559417724609, -0.0005077123641967773, -0.00047206878662109375, -0.00043642520904541016, -0.00040078163146972656, -0.00036513805389404297, -0.0003294944763183594, -0.0002938508987426758, -0.0002582073211669922, -0.0002225637435913086, -0.000186920166015625, -0.0001512765884399414, -0.00011563301086425781, -7.998943328857422e-05, -4.4345855712890625e-05, -8.702278137207031e-06, 2.6941299438476562e-05, 6.258487701416016e-05, 9.822845458984375e-05, 0.00013387203216552734, 0.00016951560974121094, 0.00020515918731689453, 0.00024080276489257812, 0.0002764463424682617, 0.0003120899200439453, 0.0003477334976196289, 0.0003833770751953125, 0.0004190206527709961, 0.0004546642303466797, 0.0004903078079223633, 0.0005259513854980469, 0.0005615949630737305, 0.0005972385406494141, 0.0006328821182250977, 0.0006685256958007812, 0.0007041692733764648, 0.0007398128509521484, 0.000775456428527832, 0.0008111000061035156, 0.0008467435836791992, 0.0008823871612548828, 0.0009180307388305664, 0.00095367431640625]}, "gradients/model.layers.28.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 3.0, 3.0, 5.0, 8.0, 6.0, 14.0, 14.0, 16.0, 27.0, 35.0, 46.0, 55.0, 63.0, 130.0, 153.0, 200.0, 399.0, 781.0, 1817.0, 5956.0, 82087.0, 4049504.0, 44549.0, 5103.0, 1564.0, 655.0, 315.0, 223.0, 132.0, 112.0, 88.0, 57.0, 49.0, 41.0, 27.0, 24.0, 9.0, 11.0, 6.0, 5.0, 5.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0042724609375, -0.004122257232666016, -0.003972053527832031, -0.003821849822998047, -0.0036716461181640625, -0.003521442413330078, -0.0033712387084960938, -0.0032210350036621094, -0.003070831298828125, -0.0029206275939941406, -0.0027704238891601562, -0.002620220184326172, -0.0024700164794921875, -0.002319812774658203, -0.0021696090698242188, -0.0020194053649902344, -0.00186920166015625, -0.0017189979553222656, -0.0015687942504882812, -0.0014185905456542969, -0.0012683868408203125, -0.0011181831359863281, -0.0009679794311523438, -0.0008177757263183594, -0.000667572021484375, -0.0005173683166503906, -0.00036716461181640625, -0.00021696090698242188, -6.67572021484375e-05, 8.344650268554688e-05, 0.00023365020751953125, 0.0003838539123535156, 0.0005340576171875, 0.0006842613220214844, 0.0008344650268554688, 0.0009846687316894531, 0.0011348724365234375, 0.0012850761413574219, 0.0014352798461914062, 0.0015854835510253906, 0.001735687255859375, 0.0018858909606933594, 0.0020360946655273438, 0.002186298370361328, 0.0023365020751953125, 0.002486705780029297, 0.0026369094848632812, 0.0027871131896972656, 0.00293731689453125, 0.0030875205993652344, 0.0032377243041992188, 0.003387928009033203, 0.0035381317138671875, 0.003688335418701172, 0.0038385391235351562, 0.003988742828369141, 0.004138946533203125, 0.004289150238037109, 0.004439353942871094, 0.004589557647705078, 0.0047397613525390625, 0.004889965057373047, 0.005040168762207031, 0.005190372467041016, 0.005340576171875]}, "gradients/model.layers.28.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 1.0, 1.0, 1.0, 7.0, 3.0, 5.0, 8.0, 6.0, 8.0, 11.0, 28.0, 34.0, 46.0, 65.0, 118.0, 314.0, 893.0, 3459.0, 28965.0, 4061322.0, 91324.0, 5594.0, 1283.0, 389.0, 160.0, 84.0, 45.0, 32.0, 31.0, 16.0, 8.0, 8.0, 11.0, 1.0, 2.0, 3.0, 6.0, 4.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.002685546875, -0.0026210546493530273, -0.0025565624237060547, -0.002492070198059082, -0.0024275779724121094, -0.0023630857467651367, -0.002298593521118164, -0.0022341012954711914, -0.0021696090698242188, -0.002105116844177246, -0.0020406246185302734, -0.0019761323928833008, -0.0019116401672363281, -0.0018471479415893555, -0.0017826557159423828, -0.0017181634902954102, -0.0016536712646484375, -0.0015891790390014648, -0.0015246868133544922, -0.0014601945877075195, -0.0013957023620605469, -0.0013312101364135742, -0.0012667179107666016, -0.001202225685119629, -0.0011377334594726562, -0.0010732412338256836, -0.001008749008178711, -0.0009442567825317383, -0.0008797645568847656, -0.000815272331237793, -0.0007507801055908203, -0.0006862878799438477, -0.000621795654296875, -0.0005573034286499023, -0.0004928112030029297, -0.00042831897735595703, -0.0003638267517089844, -0.0002993345260620117, -0.00023484230041503906, -0.0001703500747680664, -0.00010585784912109375, -4.1365623474121094e-05, 2.3126602172851562e-05, 8.761882781982422e-05, 0.00015211105346679688, 0.00021660327911376953, 0.0002810955047607422, 0.00034558773040771484, 0.0004100799560546875, 0.00047457218170166016, 0.0005390644073486328, 0.0006035566329956055, 0.0006680488586425781, 0.0007325410842895508, 0.0007970333099365234, 0.0008615255355834961, 0.0009260177612304688, 0.0009905099868774414, 0.001055002212524414, 0.0011194944381713867, 0.0011839866638183594, 0.001248478889465332, 0.0013129711151123047, 0.0013774633407592773, 0.00144195556640625]}, "gradients/model.layers.28.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 5.0, 2.0, 3.0, 10.0, 8.0, 6.0, 4.0, 9.0, 16.0, 20.0, 20.0, 28.0, 30.0, 39.0, 72.0, 95.0, 145.0, 229.0, 514.0, 927.0, 1992.0, 5487.0, 18800.0, 155781.0, 15906340.0, 640907.0, 32448.0, 7872.0, 2692.0, 1220.0, 592.0, 292.0, 189.0, 108.0, 74.0, 53.0, 50.0, 29.0, 16.0, 17.0, 14.0, 9.0, 10.0, 6.0, 5.0, 3.0, 6.0, 2.0, 3.0, 3.0, 5.0, 0.0, 4.0], "bins": [-0.00159454345703125, -0.001549839973449707, -0.001505136489868164, -0.001460433006286621, -0.0014157295227050781, -0.0013710260391235352, -0.0013263225555419922, -0.0012816190719604492, -0.0012369155883789062, -0.0011922121047973633, -0.0011475086212158203, -0.0011028051376342773, -0.0010581016540527344, -0.0010133981704711914, -0.0009686946868896484, -0.0009239912033081055, -0.0008792877197265625, -0.0008345842361450195, -0.0007898807525634766, -0.0007451772689819336, -0.0007004737854003906, -0.0006557703018188477, -0.0006110668182373047, -0.0005663633346557617, -0.0005216598510742188, -0.0004769563674926758, -0.0004322528839111328, -0.00038754940032958984, -0.0003428459167480469, -0.0002981424331665039, -0.00025343894958496094, -0.00020873546600341797, -0.000164031982421875, -0.00011932849884033203, -7.462501525878906e-05, -2.9921531677246094e-05, 1.4781951904296875e-05, 5.9485435485839844e-05, 0.00010418891906738281, 0.00014889240264892578, 0.00019359588623046875, 0.00023829936981201172, 0.0002830028533935547, 0.00032770633697509766, 0.0003724098205566406, 0.0004171133041381836, 0.00046181678771972656, 0.0005065202713012695, 0.0005512237548828125, 0.0005959272384643555, 0.0006406307220458984, 0.0006853342056274414, 0.0007300376892089844, 0.0007747411727905273, 0.0008194446563720703, 0.0008641481399536133, 0.0009088516235351562, 0.0009535551071166992, 0.0009982585906982422, 0.0010429620742797852, 0.0010876655578613281, 0.001132369041442871, 0.001177072525024414, 0.001221776008605957, 0.0012664794921875]}, "gradients/model.layers.28.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 5.0, 8.0, 12.0, 45.0, 652.0, 2305.0, 978.0, 64.0, 13.0, 4.0, 1.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.0027313232421875, -0.0026813149452209473, -0.0026313066482543945, -0.002581298351287842, -0.002531290054321289, -0.0024812817573547363, -0.0024312734603881836, -0.002381265163421631, -0.002331256866455078, -0.0022812485694885254, -0.0022312402725219727, -0.00218123197555542, -0.002131223678588867, -0.0020812153816223145, -0.0020312070846557617, -0.001981198787689209, -0.0019311904907226562, -0.0018811821937561035, -0.0018311738967895508, -0.001781165599822998, -0.0017311573028564453, -0.0016811490058898926, -0.0016311407089233398, -0.0015811324119567871, -0.0015311241149902344, -0.0014811158180236816, -0.001431107521057129, -0.0013810992240905762, -0.0013310909271240234, -0.0012810826301574707, -0.001231074333190918, -0.0011810660362243652, -0.0011310577392578125, -0.0010810494422912598, -0.001031041145324707, -0.0009810328483581543, -0.0009310245513916016, -0.0008810162544250488, -0.0008310079574584961, -0.0007809996604919434, -0.0007309913635253906, -0.0006809830665588379, -0.0006309747695922852, -0.0005809664726257324, -0.0005309581756591797, -0.00048094987869262695, -0.0004309415817260742, -0.0003809332847595215, -0.00033092498779296875, -0.000280916690826416, -0.00023090839385986328, -0.00018090009689331055, -0.0001308917999267578, -8.088350296020508e-05, -3.0875205993652344e-05, 1.913309097290039e-05, 6.914138793945312e-05, 0.00011914968490600586, 0.0001691579818725586, 0.00021916627883911133, 0.00026917457580566406, 0.0003191828727722168, 0.00036919116973876953, 0.00041919946670532227, 0.000469207763671875]}, "gradients/model.layers.27.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 2.0, 1.0, 3.0, 9.0, 12.0, 16.0, 22.0, 36.0, 37.0, 60.0, 100.0, 99.0, 137.0, 209.0, 300.0, 426.0, 859.0, 3786.0, 636974.0, 16777216.0, 34130.0, 1574.0, 575.0, 346.0, 242.0, 163.0, 118.0, 112.0, 68.0, 51.0, 35.0, 21.0, 16.0, 12.0, 4.0, 5.0, 2.0, 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00592041015625, -0.005724906921386719, -0.0055294036865234375, -0.005333900451660156, -0.005138397216796875, -0.004942893981933594, -0.0047473907470703125, -0.004551887512207031, -0.00435638427734375, -0.004160881042480469, -0.0039653778076171875, -0.0037698745727539062, -0.003574371337890625, -0.0033788681030273438, -0.0031833648681640625, -0.0029878616333007812, -0.0027923583984375, -0.0025968551635742188, -0.0024013519287109375, -0.0022058486938476562, -0.002010345458984375, -0.0018148422241210938, -0.0016193389892578125, -0.0014238357543945312, -0.00122833251953125, -0.0010328292846679688, -0.0008373260498046875, -0.0006418228149414062, -0.000446319580078125, -0.00025081634521484375, -5.53131103515625e-05, 0.00014019012451171875, 0.000335693359375, 0.0005311965942382812, 0.0007266998291015625, 0.0009222030639648438, 0.001117706298828125, 0.0013132095336914062, 0.0015087127685546875, 0.0017042160034179688, 0.00189971923828125, 0.0020952224731445312, 0.0022907257080078125, 0.0024862289428710938, 0.002681732177734375, 0.0028772354125976562, 0.0030727386474609375, 0.0032682418823242188, 0.0034637451171875, 0.0036592483520507812, 0.0038547515869140625, 0.004050254821777344, 0.004245758056640625, 0.004441261291503906, 0.0046367645263671875, 0.004832267761230469, 0.00502777099609375, 0.005223274230957031, 0.0054187774658203125, 0.005614280700683594, 0.005809783935546875, 0.006005287170410156, 0.0062007904052734375, 0.006396293640136719, 0.006591796875]}, "gradients/model.layers.27.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 2.0, 0.0, 1.0, 1.0, 4.0, 5.0, 4.0, 11.0, 63.0, 261.0, 1054.0, 14559.0, 16777216.0, 38745.0, 1454.0, 325.0, 76.0, 23.0, 11.0, 3.0, 3.0, 2.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0118408203125, -0.011411666870117188, -0.010982513427734375, -0.010553359985351562, -0.01012420654296875, -0.009695053100585938, -0.009265899658203125, -0.008836746215820312, -0.0084075927734375, -0.007978439331054688, -0.007549285888671875, -0.0071201324462890625, -0.00669097900390625, -0.0062618255615234375, -0.005832672119140625, -0.0054035186767578125, -0.004974365234375, -0.0045452117919921875, -0.004116058349609375, -0.0036869049072265625, -0.00325775146484375, -0.0028285980224609375, -0.002399444580078125, -0.0019702911376953125, -0.0015411376953125, -0.0011119842529296875, -0.000682830810546875, -0.0002536773681640625, 0.00017547607421875, 0.0006046295166015625, 0.001033782958984375, 0.0014629364013671875, 0.00189208984375, 0.0023212432861328125, 0.002750396728515625, 0.0031795501708984375, 0.00360870361328125, 0.0040378570556640625, 0.004467010498046875, 0.0048961639404296875, 0.0053253173828125, 0.0057544708251953125, 0.006183624267578125, 0.0066127777099609375, 0.00704193115234375, 0.0074710845947265625, 0.007900238037109375, 0.008329391479492188, 0.008758544921875, 0.009187698364257812, 0.009616851806640625, 0.010046005249023438, 0.01047515869140625, 0.010904312133789062, 0.011333465576171875, 0.011762619018554688, 0.0121917724609375, 0.012620925903320312, 0.013050079345703125, 0.013479232788085938, 0.01390838623046875, 0.014337539672851562, 0.014766693115234375, 0.015195846557617188, 0.015625]}, "gradients/model.layers.27.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 0.0, 5.0, 7.0, 5.0, 7.0, 13.0, 94.0, 864.0, 2194.0, 838.0, 56.0, 2.0, 2.0], "bins": [-0.005462646484375, -0.005370885133743286, -0.005279123783111572, -0.005187362432479858, -0.0050956010818481445, -0.005003839731216431, -0.004912078380584717, -0.004820317029953003, -0.004728555679321289, -0.004636794328689575, -0.004545032978057861, -0.0044532716274261475, -0.004361510276794434, -0.00426974892616272, -0.004177987575531006, -0.004086226224899292, -0.003994464874267578, -0.0039027035236358643, -0.0038109421730041504, -0.0037191808223724365, -0.0036274194717407227, -0.003535658121109009, -0.003443896770477295, -0.003352135419845581, -0.003260374069213867, -0.0031686127185821533, -0.0030768513679504395, -0.0029850900173187256, -0.0028933286666870117, -0.002801567316055298, -0.002709805965423584, -0.00261804461479187, -0.0025262832641601562, -0.0024345219135284424, -0.0023427605628967285, -0.0022509992122650146, -0.0021592378616333008, -0.002067476511001587, -0.001975715160369873, -0.0018839538097381592, -0.0017921924591064453, -0.0017004311084747314, -0.0016086697578430176, -0.0015169084072113037, -0.0014251470565795898, -0.001333385705947876, -0.0012416243553161621, -0.0011498630046844482, -0.0010581016540527344, -0.0009663403034210205, -0.0008745789527893066, -0.0007828176021575928, -0.0006910562515258789, -0.000599294900894165, -0.0005075335502624512, -0.0004157721996307373, -0.00032401084899902344, -0.00023224949836730957, -0.0001404881477355957, -4.8726797103881836e-05, 4.303455352783203e-05, 0.0001347959041595459, 0.00022655725479125977, 0.00031831860542297363, 0.0004100799560546875]}, "gradients/model.layers.27.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 3.0, 1.0, 2.0, 0.0, 1.0, 2.0, 4.0, 4.0, 8.0, 4.0, 15.0, 7.0, 20.0, 40.0, 55.0, 98.0, 186.0, 309.0, 557.0, 1027.0, 2212.0, 4856.0, 12239.0, 37878.0, 188429.0, 1733465.0, 11963557.0, 2505411.0, 255229.0, 46733.0, 14336.0, 5592.0, 2387.0, 1141.0, 656.0, 274.0, 209.0, 103.0, 60.0, 36.0, 24.0, 12.0, 6.0, 3.0, 4.0, 4.0, 5.0, 3.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.0011444091796875, -0.0011069774627685547, -0.0010695457458496094, -0.001032114028930664, -0.0009946823120117188, -0.0009572505950927734, -0.0009198188781738281, -0.0008823871612548828, -0.0008449554443359375, -0.0008075237274169922, -0.0007700920104980469, -0.0007326602935791016, -0.0006952285766601562, -0.0006577968597412109, -0.0006203651428222656, -0.0005829334259033203, -0.000545501708984375, -0.0005080699920654297, -0.0004706382751464844, -0.00043320655822753906, -0.00039577484130859375, -0.00035834312438964844, -0.0003209114074707031, -0.0002834796905517578, -0.0002460479736328125, -0.0002086162567138672, -0.00017118453979492188, -0.00013375282287597656, -9.632110595703125e-05, -5.888938903808594e-05, -2.1457672119140625e-05, 1.5974044799804688e-05, 5.340576171875e-05, 9.083747863769531e-05, 0.00012826919555664062, 0.00016570091247558594, 0.00020313262939453125, 0.00024056434631347656, 0.0002779960632324219, 0.0003154277801513672, 0.0003528594970703125, 0.0003902912139892578, 0.0004277229309082031, 0.00046515464782714844, 0.0005025863647460938, 0.0005400180816650391, 0.0005774497985839844, 0.0006148815155029297, 0.000652313232421875, 0.0006897449493408203, 0.0007271766662597656, 0.0007646083831787109, 0.0008020401000976562, 0.0008394718170166016, 0.0008769035339355469, 0.0009143352508544922, 0.0009517669677734375, 0.0009891986846923828, 0.0010266304016113281, 0.0010640621185302734, 0.0011014938354492188, 0.001138925552368164, 0.0011763572692871094, 0.0012137889862060547, 0.001251220703125]}, "gradients/model.layers.27.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 4.0, 1.0, 1.0, 8.0, 17.0, 20.0, 29.0, 49.0, 73.0, 107.0, 134.0, 324.0, 612.0, 2006.0, 17267.0, 4119651.0, 49562.0, 2844.0, 759.0, 321.0, 181.0, 110.0, 71.0, 50.0, 45.0, 21.0, 14.0, 3.0, 9.0, 2.0, 1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.007476806640625, -0.007260799407958984, -0.007044792175292969, -0.006828784942626953, -0.0066127777099609375, -0.006396770477294922, -0.006180763244628906, -0.005964756011962891, -0.005748748779296875, -0.005532741546630859, -0.005316734313964844, -0.005100727081298828, -0.0048847198486328125, -0.004668712615966797, -0.004452705383300781, -0.004236698150634766, -0.00402069091796875, -0.0038046836853027344, -0.0035886764526367188, -0.003372669219970703, -0.0031566619873046875, -0.002940654754638672, -0.0027246475219726562, -0.0025086402893066406, -0.002292633056640625, -0.0020766258239746094, -0.0018606185913085938, -0.0016446113586425781, -0.0014286041259765625, -0.0012125968933105469, -0.0009965896606445312, -0.0007805824279785156, -0.0005645751953125, -0.0003485679626464844, -0.00013256072998046875, 8.344650268554688e-05, 0.0002994537353515625, 0.0005154609680175781, 0.0007314682006835938, 0.0009474754333496094, 0.001163482666015625, 0.0013794898986816406, 0.0015954971313476562, 0.0018115043640136719, 0.0020275115966796875, 0.002243518829345703, 0.0024595260620117188, 0.0026755332946777344, 0.00289154052734375, 0.0031075477600097656, 0.0033235549926757812, 0.003539562225341797, 0.0037555694580078125, 0.003971576690673828, 0.004187583923339844, 0.004403591156005859, 0.004619598388671875, 0.004835605621337891, 0.005051612854003906, 0.005267620086669922, 0.0054836273193359375, 0.005699634552001953, 0.005915641784667969, 0.006131649017333984, 0.00634765625]}, "gradients/model.layers.27.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 1.0, 2.0, 0.0, 1.0, 0.0, 1.0, 2.0, 3.0, 1.0, 8.0, 6.0, 9.0, 7.0, 4.0, 14.0, 16.0, 17.0, 32.0, 34.0, 53.0, 88.0, 170.0, 281.0, 685.0, 1866.0, 7223.0, 100788.0, 4053220.0, 23255.0, 4121.0, 1275.0, 505.0, 236.0, 128.0, 69.0, 49.0, 30.0, 26.0, 19.0, 15.0, 12.0, 4.0, 7.0, 3.0, 1.0, 0.0, 3.0, 2.0, 1.0, 1.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0], "bins": [-0.002044677734375, -0.001977205276489258, -0.0019097328186035156, -0.0018422603607177734, -0.0017747879028320312, -0.001707315444946289, -0.0016398429870605469, -0.0015723705291748047, -0.0015048980712890625, -0.0014374256134033203, -0.0013699531555175781, -0.001302480697631836, -0.0012350082397460938, -0.0011675357818603516, -0.0011000633239746094, -0.0010325908660888672, -0.000965118408203125, -0.0008976459503173828, -0.0008301734924316406, -0.0007627010345458984, -0.0006952285766601562, -0.0006277561187744141, -0.0005602836608886719, -0.0004928112030029297, -0.0004253387451171875, -0.0003578662872314453, -0.0002903938293457031, -0.00022292137145996094, -0.00015544891357421875, -8.797645568847656e-05, -2.0503997802734375e-05, 4.696846008300781e-05, 0.00011444091796875, 0.0001819133758544922, 0.0002493858337402344, 0.00031685829162597656, 0.00038433074951171875, 0.00045180320739746094, 0.0005192756652832031, 0.0005867481231689453, 0.0006542205810546875, 0.0007216930389404297, 0.0007891654968261719, 0.0008566379547119141, 0.0009241104125976562, 0.0009915828704833984, 0.0010590553283691406, 0.0011265277862548828, 0.001194000244140625, 0.0012614727020263672, 0.0013289451599121094, 0.0013964176177978516, 0.0014638900756835938, 0.001531362533569336, 0.0015988349914550781, 0.0016663074493408203, 0.0017337799072265625, 0.0018012523651123047, 0.0018687248229980469, 0.001936197280883789, 0.0020036697387695312, 0.0020711421966552734, 0.0021386146545410156, 0.002206087112426758, 0.0022735595703125]}, "gradients/model.layers.27.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 4.0, 2.0, 4.0, 7.0, 3.0, 10.0, 8.0, 7.0, 15.0, 18.0, 40.0, 55.0, 112.0, 166.0, 411.0, 1022.0, 4022.0, 26383.0, 15226619.0, 1497591.0, 16154.0, 2890.0, 854.0, 373.0, 148.0, 97.0, 60.0, 39.0, 28.0, 9.0, 10.0, 7.0, 7.0, 8.0, 6.0, 5.0, 4.0, 2.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 1.0], "bins": [-0.002471923828125, -0.002394437789916992, -0.0023169517517089844, -0.0022394657135009766, -0.0021619796752929688, -0.002084493637084961, -0.002007007598876953, -0.0019295215606689453, -0.0018520355224609375, -0.0017745494842529297, -0.0016970634460449219, -0.001619577407836914, -0.0015420913696289062, -0.0014646053314208984, -0.0013871192932128906, -0.0013096332550048828, -0.001232147216796875, -0.0011546611785888672, -0.0010771751403808594, -0.0009996891021728516, -0.0009222030639648438, -0.0008447170257568359, -0.0007672309875488281, -0.0006897449493408203, -0.0006122589111328125, -0.0005347728729248047, -0.0004572868347167969, -0.00037980079650878906, -0.00030231475830078125, -0.00022482872009277344, -0.00014734268188476562, -6.985664367675781e-05, 7.62939453125e-06, 8.511543273925781e-05, 0.00016260147094726562, 0.00024008750915527344, 0.00031757354736328125, 0.00039505958557128906, 0.0004725456237792969, 0.0005500316619873047, 0.0006275177001953125, 0.0007050037384033203, 0.0007824897766113281, 0.0008599758148193359, 0.0009374618530273438, 0.0010149478912353516, 0.0010924339294433594, 0.0011699199676513672, 0.001247406005859375, 0.0013248920440673828, 0.0014023780822753906, 0.0014798641204833984, 0.0015573501586914062, 0.001634836196899414, 0.0017123222351074219, 0.0017898082733154297, 0.0018672943115234375, 0.0019447803497314453, 0.002022266387939453, 0.002099752426147461, 0.0021772384643554688, 0.0022547245025634766, 0.0023322105407714844, 0.002409696578979492, 0.0024871826171875]}, "gradients/model.layers.27.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 0.0, 2.0, 0.0, 2.0, 3.0, 1.0, 1.0, 3.0, 9.0, 7.0, 14.0, 8.0, 17.0, 148.0, 559.0, 1363.0, 1276.0, 488.0, 123.0, 33.0, 14.0, 8.0, 0.0, 0.0, 3.0, 3.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.000812530517578125, -0.000788569450378418, -0.0007646083831787109, -0.0007406473159790039, -0.0007166862487792969, -0.0006927251815795898, -0.0006687641143798828, -0.0006448030471801758, -0.0006208419799804688, -0.0005968809127807617, -0.0005729198455810547, -0.0005489587783813477, -0.0005249977111816406, -0.0005010366439819336, -0.00047707557678222656, -0.00045311450958251953, -0.0004291534423828125, -0.00040519237518310547, -0.00038123130798339844, -0.0003572702407836914, -0.0003333091735839844, -0.00030934810638427734, -0.0002853870391845703, -0.0002614259719848633, -0.00023746490478515625, -0.00021350383758544922, -0.0001895427703857422, -0.00016558170318603516, -0.00014162063598632812, -0.0001176595687866211, -9.369850158691406e-05, -6.973743438720703e-05, -4.57763671875e-05, -2.181529998779297e-05, 2.1457672119140625e-06, 2.6106834411621094e-05, 5.0067901611328125e-05, 7.402896881103516e-05, 9.799003601074219e-05, 0.00012195110321044922, 0.00014591217041015625, 0.00016987323760986328, 0.0001938343048095703, 0.00021779537200927734, 0.00024175643920898438, 0.0002657175064086914, 0.00028967857360839844, 0.00031363964080810547, 0.0003376007080078125, 0.00036156177520751953, 0.00038552284240722656, 0.0004094839096069336, 0.0004334449768066406, 0.00045740604400634766, 0.0004813671112060547, 0.0005053281784057617, 0.0005292892456054688, 0.0005532503128051758, 0.0005772113800048828, 0.0006011724472045898, 0.0006251335144042969, 0.0006490945816040039, 0.0006730556488037109, 0.000697016716003418, 0.000720977783203125]}, "gradients/model.layers.26.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0, 3.0, 3.0, 1.0, 4.0, 4.0, 2.0, 8.0, 23.0, 25.0, 27.0, 62.0, 96.0, 144.0, 211.0, 499.0, 885.0, 2082.0, 5165.0, 15801.0, 61974.0, 366771.0, 10226740.0, 16777216.0, 2708112.0, 192217.0, 38863.0, 11078.0, 3936.0, 1589.0, 737.0, 363.0, 181.0, 98.0, 76.0, 35.0, 45.0, 18.0, 9.0, 4.0, 6.0, 3.0, 3.0, 4.0, 3.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0015716552734375, -0.00152587890625, -0.0014801025390625, -0.001434326171875, -0.0013885498046875, -0.0013427734375, -0.0012969970703125, -0.001251220703125, -0.0012054443359375, -0.00115966796875, -0.0011138916015625, -0.001068115234375, -0.0010223388671875, -0.0009765625, -0.0009307861328125, -0.000885009765625, -0.0008392333984375, -0.00079345703125, -0.0007476806640625, -0.000701904296875, -0.0006561279296875, -0.0006103515625, -0.0005645751953125, -0.000518798828125, -0.0004730224609375, -0.00042724609375, -0.0003814697265625, -0.000335693359375, -0.0002899169921875, -0.000244140625, -0.0001983642578125, -0.000152587890625, -0.0001068115234375, -6.103515625e-05, -1.52587890625e-05, 3.0517578125e-05, 7.62939453125e-05, 0.0001220703125, 0.0001678466796875, 0.000213623046875, 0.0002593994140625, 0.00030517578125, 0.0003509521484375, 0.000396728515625, 0.0004425048828125, 0.00048828125, 0.0005340576171875, 0.000579833984375, 0.0006256103515625, 0.00067138671875, 0.0007171630859375, 0.000762939453125, 0.0008087158203125, 0.0008544921875, 0.0009002685546875, 0.000946044921875, 0.0009918212890625, 0.00103759765625, 0.0010833740234375, 0.001129150390625, 0.0011749267578125, 0.001220703125, 0.0012664794921875, 0.001312255859375, 0.0013580322265625]}, "gradients/model.layers.26.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 3.0, 1.0, 4.0, 2.0, 4.0, 2.0, 4.0, 4.0, 6.0, 12.0, 19.0, 27.0, 31.0, 56.0, 87.0, 133.0, 373.0, 1052.0, 4090.0, 24128.0, 329566.0, 16777216.0, 1515956.0, 46497.0, 6986.0, 1513.0, 496.0, 207.0, 105.0, 70.0, 34.0, 22.0, 24.0, 11.0, 9.0, 3.0, 5.0, 6.0, 4.0, 1.0, 2.0, 4.0, 0.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.003173828125, -0.0030667781829833984, -0.002959728240966797, -0.0028526782989501953, -0.0027456283569335938, -0.002638578414916992, -0.0025315284729003906, -0.002424478530883789, -0.0023174285888671875, -0.002210378646850586, -0.0021033287048339844, -0.001996278762817383, -0.0018892288208007812, -0.0017821788787841797, -0.0016751289367675781, -0.0015680789947509766, -0.001461029052734375, -0.0013539791107177734, -0.0012469291687011719, -0.0011398792266845703, -0.0010328292846679688, -0.0009257793426513672, -0.0008187294006347656, -0.0007116794586181641, -0.0006046295166015625, -0.0004975795745849609, -0.0003905296325683594, -0.0002834796905517578, -0.00017642974853515625, -6.937980651855469e-05, 3.7670135498046875e-05, 0.00014472007751464844, 0.00025177001953125, 0.00035881996154785156, 0.0004658699035644531, 0.0005729198455810547, 0.0006799697875976562, 0.0007870197296142578, 0.0008940696716308594, 0.001001119613647461, 0.0011081695556640625, 0.001215219497680664, 0.0013222694396972656, 0.0014293193817138672, 0.0015363693237304688, 0.0016434192657470703, 0.0017504692077636719, 0.0018575191497802734, 0.001964569091796875, 0.0020716190338134766, 0.002178668975830078, 0.0022857189178466797, 0.0023927688598632812, 0.002499818801879883, 0.0026068687438964844, 0.002713918685913086, 0.0028209686279296875, 0.002928018569946289, 0.0030350685119628906, 0.003142118453979492, 0.0032491683959960938, 0.0033562183380126953, 0.003463268280029297, 0.0035703182220458984, 0.0036773681640625]}, "gradients/model.layers.26.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 1.0, 5.0, 9.0, 14.0, 64.0, 241.0, 663.0, 1141.0, 1084.0, 588.0, 198.0, 60.0, 12.0, 2.0, 1.0, 2.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00145721435546875, -0.0014115571975708008, -0.0013659000396728516, -0.0013202428817749023, -0.0012745857238769531, -0.001228928565979004, -0.0011832714080810547, -0.0011376142501831055, -0.0010919570922851562, -0.001046299934387207, -0.0010006427764892578, -0.0009549856185913086, -0.0009093284606933594, -0.0008636713027954102, -0.0008180141448974609, -0.0007723569869995117, -0.0007266998291015625, -0.0006810426712036133, -0.0006353855133056641, -0.0005897283554077148, -0.0005440711975097656, -0.0004984140396118164, -0.0004527568817138672, -0.00040709972381591797, -0.00036144256591796875, -0.00031578540802001953, -0.0002701282501220703, -0.0002244710922241211, -0.00017881393432617188, -0.00013315677642822266, -8.749961853027344e-05, -4.184246063232422e-05, 3.814697265625e-06, 4.947185516357422e-05, 9.512901306152344e-05, 0.00014078617095947266, 0.00018644332885742188, 0.0002321004867553711, 0.0002777576446533203, 0.00032341480255126953, 0.00036907196044921875, 0.00041472911834716797, 0.0004603862762451172, 0.0005060434341430664, 0.0005517005920410156, 0.0005973577499389648, 0.0006430149078369141, 0.0006886720657348633, 0.0007343292236328125, 0.0007799863815307617, 0.0008256435394287109, 0.0008713006973266602, 0.0009169578552246094, 0.0009626150131225586, 0.0010082721710205078, 0.001053929328918457, 0.0010995864868164062, 0.0011452436447143555, 0.0011909008026123047, 0.001236557960510254, 0.0012822151184082031, 0.0013278722763061523, 0.0013735294342041016, 0.0014191865921020508, 0.00146484375]}, "gradients/model.layers.26.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 2.0, 4.0, 7.0, 4.0, 3.0, 10.0, 11.0, 28.0, 38.0, 54.0, 106.0, 227.0, 353.0, 565.0, 969.0, 1696.0, 2952.0, 6321.0, 18754.0, 148670.0, 1952703.0, 12754174.0, 1725602.0, 133132.0, 17768.0, 6195.0, 2902.0, 1672.0, 951.0, 528.0, 293.0, 199.0, 120.0, 71.0, 33.0, 37.0, 18.0, 5.0, 8.0, 3.0, 4.0, 5.0, 1.0, 4.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.001434326171875, -0.0013927221298217773, -0.0013511180877685547, -0.001309514045715332, -0.0012679100036621094, -0.0012263059616088867, -0.001184701919555664, -0.0011430978775024414, -0.0011014938354492188, -0.001059889793395996, -0.0010182857513427734, -0.0009766817092895508, -0.0009350776672363281, -0.0008934736251831055, -0.0008518695831298828, -0.0008102655410766602, -0.0007686614990234375, -0.0007270574569702148, -0.0006854534149169922, -0.0006438493728637695, -0.0006022453308105469, -0.0005606412887573242, -0.0005190372467041016, -0.0004774332046508789, -0.00043582916259765625, -0.0003942251205444336, -0.00035262107849121094, -0.0003110170364379883, -0.0002694129943847656, -0.00022780895233154297, -0.0001862049102783203, -0.00014460086822509766, -0.000102996826171875, -6.139278411865234e-05, -1.9788742065429688e-05, 2.181529998779297e-05, 6.341934204101562e-05, 0.00010502338409423828, 0.00014662742614746094, 0.0001882314682006836, 0.00022983551025390625, 0.0002714395523071289, 0.00031304359436035156, 0.0003546476364135742, 0.0003962516784667969, 0.00043785572052001953, 0.0004794597625732422, 0.0005210638046264648, 0.0005626678466796875, 0.0006042718887329102, 0.0006458759307861328, 0.0006874799728393555, 0.0007290840148925781, 0.0007706880569458008, 0.0008122920989990234, 0.0008538961410522461, 0.0008955001831054688, 0.0009371042251586914, 0.000978708267211914, 0.0010203123092651367, 0.0010619163513183594, 0.001103520393371582, 0.0011451244354248047, 0.0011867284774780273, 0.00122833251953125]}, "gradients/model.layers.26.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 2.0, 0.0, 5.0, 2.0, 7.0, 5.0, 2.0, 11.0, 11.0, 16.0, 15.0, 24.0, 26.0, 33.0, 52.0, 79.0, 108.0, 134.0, 164.0, 284.0, 408.0, 920.0, 1730.0, 3695.0, 9186.0, 76797.0, 3707534.0, 369923.0, 13631.0, 4710.0, 2022.0, 1118.0, 541.0, 324.0, 220.0, 140.0, 95.0, 71.0, 52.0, 40.0, 39.0, 28.0, 27.0, 20.0, 9.0, 10.0, 5.0, 4.0, 7.0, 4.0, 3.0, 2.0, 1.0, 0.0, 0.0, 2.0, 2.0], "bins": [-0.0034332275390625, -0.0033311843872070312, -0.0032291412353515625, -0.0031270980834960938, -0.003025054931640625, -0.0029230117797851562, -0.0028209686279296875, -0.0027189254760742188, -0.00261688232421875, -0.0025148391723632812, -0.0024127960205078125, -0.0023107528686523438, -0.002208709716796875, -0.0021066665649414062, -0.0020046234130859375, -0.0019025802612304688, -0.001800537109375, -0.0016984939575195312, -0.0015964508056640625, -0.0014944076538085938, -0.001392364501953125, -0.0012903213500976562, -0.0011882781982421875, -0.0010862350463867188, -0.00098419189453125, -0.0008821487426757812, -0.0007801055908203125, -0.0006780624389648438, -0.000576019287109375, -0.00047397613525390625, -0.0003719329833984375, -0.00026988983154296875, -0.0001678466796875, -6.580352783203125e-05, 3.62396240234375e-05, 0.00013828277587890625, 0.000240325927734375, 0.00034236907958984375, 0.0004444122314453125, 0.0005464553833007812, 0.00064849853515625, 0.0007505416870117188, 0.0008525848388671875, 0.0009546279907226562, 0.001056671142578125, 0.0011587142944335938, 0.0012607574462890625, 0.0013628005981445312, 0.00146484375, 0.0015668869018554688, 0.0016689300537109375, 0.0017709732055664062, 0.001873016357421875, 0.0019750595092773438, 0.0020771026611328125, 0.0021791458129882812, 0.00228118896484375, 0.0023832321166992188, 0.0024852752685546875, 0.0025873184204101562, 0.002689361572265625, 0.0027914047241210938, 0.0028934478759765625, 0.0029954910278320312, 0.0030975341796875]}, "gradients/model.layers.26.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 2.0, 3.0, 3.0, 6.0, 6.0, 8.0, 10.0, 25.0, 19.0, 24.0, 47.0, 59.0, 80.0, 158.0, 282.0, 623.0, 1691.0, 5083.0, 29031.0, 3901860.0, 239605.0, 10755.0, 2946.0, 981.0, 427.0, 161.0, 104.0, 88.0, 46.0, 39.0, 31.0, 15.0, 12.0, 15.0, 14.0, 6.0, 7.0, 7.0, 1.0, 3.0, 2.0, 1.0, 2.0, 0.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0], "bins": [-0.0015716552734375, -0.001518845558166504, -0.0014660358428955078, -0.0014132261276245117, -0.0013604164123535156, -0.0013076066970825195, -0.0012547969818115234, -0.0012019872665405273, -0.0011491775512695312, -0.0010963678359985352, -0.001043558120727539, -0.000990748405456543, -0.0009379386901855469, -0.0008851289749145508, -0.0008323192596435547, -0.0007795095443725586, -0.0007266998291015625, -0.0006738901138305664, -0.0006210803985595703, -0.0005682706832885742, -0.0005154609680175781, -0.00046265125274658203, -0.00040984153747558594, -0.00035703182220458984, -0.00030422210693359375, -0.00025141239166259766, -0.00019860267639160156, -0.00014579296112060547, -9.298324584960938e-05, -4.017353057861328e-05, 1.2636184692382812e-05, 6.54458999633789e-05, 0.000118255615234375, 0.0001710653305053711, 0.0002238750457763672, 0.0002766847610473633, 0.0003294944763183594, 0.00038230419158935547, 0.00043511390686035156, 0.00048792362213134766, 0.0005407333374023438, 0.0005935430526733398, 0.0006463527679443359, 0.000699162483215332, 0.0007519721984863281, 0.0008047819137573242, 0.0008575916290283203, 0.0009104013442993164, 0.0009632110595703125, 0.0010160207748413086, 0.0010688304901123047, 0.0011216402053833008, 0.0011744499206542969, 0.001227259635925293, 0.001280069351196289, 0.0013328790664672852, 0.0013856887817382812, 0.0014384984970092773, 0.0014913082122802734, 0.0015441179275512695, 0.0015969276428222656, 0.0016497373580932617, 0.0017025470733642578, 0.001755356788635254, 0.00180816650390625]}, "gradients/model.layers.26.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 2.0, 0.0, 2.0, 1.0, 0.0, 5.0, 1.0, 4.0, 2.0, 3.0, 7.0, 7.0, 6.0, 11.0, 17.0, 16.0, 42.0, 64.0, 122.0, 170.0, 293.0, 628.0, 1725.0, 5710.0, 40822.0, 16149546.0, 555890.0, 16443.0, 3459.0, 1115.0, 455.0, 229.0, 138.0, 113.0, 40.0, 26.0, 20.0, 23.0, 14.0, 9.0, 9.0, 7.0, 7.0, 2.0, 2.0, 2.0, 0.0, 0.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00238037109375, -0.0023005008697509766, -0.002220630645751953, -0.0021407604217529297, -0.0020608901977539062, -0.001981019973754883, -0.0019011497497558594, -0.001821279525756836, -0.0017414093017578125, -0.001661539077758789, -0.0015816688537597656, -0.0015017986297607422, -0.0014219284057617188, -0.0013420581817626953, -0.0012621879577636719, -0.0011823177337646484, -0.001102447509765625, -0.0010225772857666016, -0.0009427070617675781, -0.0008628368377685547, -0.0007829666137695312, -0.0007030963897705078, -0.0006232261657714844, -0.0005433559417724609, -0.0004634857177734375, -0.00038361549377441406, -0.0003037452697753906, -0.0002238750457763672, -0.00014400482177734375, -6.413459777832031e-05, 1.5735626220703125e-05, 9.560585021972656e-05, 0.00017547607421875, 0.00025534629821777344, 0.0003352165222167969, 0.0004150867462158203, 0.0004949569702148438, 0.0005748271942138672, 0.0006546974182128906, 0.0007345676422119141, 0.0008144378662109375, 0.0008943080902099609, 0.0009741783142089844, 0.0010540485382080078, 0.0011339187622070312, 0.0012137889862060547, 0.0012936592102050781, 0.0013735294342041016, 0.001453399658203125, 0.0015332698822021484, 0.0016131401062011719, 0.0016930103302001953, 0.0017728805541992188, 0.0018527507781982422, 0.0019326210021972656, 0.002012491226196289, 0.0020923614501953125, 0.002172231674194336, 0.0022521018981933594, 0.002331972122192383, 0.0024118423461914062, 0.0024917125701904297, 0.002571582794189453, 0.0026514530181884766, 0.0027313232421875]}, "gradients/model.layers.26.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 2.0, 0.0, 2.0, 1.0, 3.0, 3.0, 6.0, 9.0, 20.0, 113.0, 1124.0, 2062.0, 648.0, 56.0, 20.0, 3.0, 6.0, 5.0, 2.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0012664794921875, -0.0012233257293701172, -0.0011801719665527344, -0.0011370182037353516, -0.0010938644409179688, -0.001050710678100586, -0.0010075569152832031, -0.0009644031524658203, -0.0009212493896484375, -0.0008780956268310547, -0.0008349418640136719, -0.0007917881011962891, -0.0007486343383789062, -0.0007054805755615234, -0.0006623268127441406, -0.0006191730499267578, -0.000576019287109375, -0.0005328655242919922, -0.0004897117614746094, -0.00044655799865722656, -0.00040340423583984375, -0.00036025047302246094, -0.0003170967102050781, -0.0002739429473876953, -0.0002307891845703125, -0.0001876354217529297, -0.00014448165893554688, -0.00010132789611816406, -5.817413330078125e-05, -1.5020370483398438e-05, 2.8133392333984375e-05, 7.128715515136719e-05, 0.00011444091796875, 0.0001575946807861328, 0.00020074844360351562, 0.00024390220642089844, 0.00028705596923828125, 0.00033020973205566406, 0.0003733634948730469, 0.0004165172576904297, 0.0004596710205078125, 0.0005028247833251953, 0.0005459785461425781, 0.0005891323089599609, 0.0006322860717773438, 0.0006754398345947266, 0.0007185935974121094, 0.0007617473602294922, 0.000804901123046875, 0.0008480548858642578, 0.0008912086486816406, 0.0009343624114990234, 0.0009775161743164062, 0.001020669937133789, 0.0010638236999511719, 0.0011069774627685547, 0.0011501312255859375, 0.0011932849884033203, 0.0012364387512207031, 0.001279592514038086, 0.0013227462768554688, 0.0013659000396728516, 0.0014090538024902344, 0.0014522075653076172, 0.001495361328125]}, "gradients/model.layers.25.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 2.0, 3.0, 1.0, 2.0, 9.0, 9.0, 13.0, 6.0, 23.0, 26.0, 35.0, 51.0, 74.0, 86.0, 112.0, 144.0, 221.0, 306.0, 510.0, 1076.0, 2715.0, 9757.0, 57484.0, 927870.0, 16777216.0, 6341760.0, 130620.0, 17602.0, 4077.0, 1409.0, 629.0, 400.0, 220.0, 162.0, 127.0, 82.0, 63.0, 47.0, 40.0, 22.0, 21.0, 21.0, 6.0, 7.0, 4.0, 5.0, 6.0, 3.0, 1.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.00244140625, -0.0023729801177978516, -0.002304553985595703, -0.0022361278533935547, -0.0021677017211914062, -0.002099275588989258, -0.0020308494567871094, -0.001962423324584961, -0.0018939971923828125, -0.001825571060180664, -0.0017571449279785156, -0.0016887187957763672, -0.0016202926635742188, -0.0015518665313720703, -0.0014834403991699219, -0.0014150142669677734, -0.001346588134765625, -0.0012781620025634766, -0.0012097358703613281, -0.0011413097381591797, -0.0010728836059570312, -0.0010044574737548828, -0.0009360313415527344, -0.0008676052093505859, -0.0007991790771484375, -0.0007307529449462891, -0.0006623268127441406, -0.0005939006805419922, -0.0005254745483398438, -0.0004570484161376953, -0.0003886222839355469, -0.00032019615173339844, -0.00025177001953125, -0.00018334388732910156, -0.00011491775512695312, -4.649162292480469e-05, 2.193450927734375e-05, 9.036064147949219e-05, 0.00015878677368164062, 0.00022721290588378906, 0.0002956390380859375, 0.00036406517028808594, 0.0004324913024902344, 0.0005009174346923828, 0.0005693435668945312, 0.0006377696990966797, 0.0007061958312988281, 0.0007746219635009766, 0.000843048095703125, 0.0009114742279052734, 0.0009799003601074219, 0.0010483264923095703, 0.0011167526245117188, 0.0011851787567138672, 0.0012536048889160156, 0.001322031021118164, 0.0013904571533203125, 0.001458883285522461, 0.0015273094177246094, 0.0015957355499267578, 0.0016641616821289062, 0.0017325878143310547, 0.0018010139465332031, 0.0018694400787353516, 0.0019378662109375]}, "gradients/model.layers.25.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 0.0, 2.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 2.0, 1.0, 5.0, 3.0, 1.0, 4.0, 13.0, 5.0, 23.0, 36.0, 48.0, 84.0, 109.0, 236.0, 460.0, 1256.0, 4502.0, 22335.0, 204644.0, 16777216.0, 16777216.0, 127431.0, 16204.0, 3484.0, 1077.0, 421.0, 164.0, 114.0, 59.0, 39.0, 35.0, 11.0, 13.0, 12.0, 5.0, 7.0, 10.0, 6.0, 4.0, 4.0, 3.0, 5.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 3.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.003021240234375, -0.0029234886169433594, -0.0028257369995117188, -0.002727985382080078, -0.0026302337646484375, -0.002532482147216797, -0.0024347305297851562, -0.0023369789123535156, -0.002239227294921875, -0.0021414756774902344, -0.0020437240600585938, -0.0019459724426269531, -0.0018482208251953125, -0.0017504692077636719, -0.0016527175903320312, -0.0015549659729003906, -0.00145721435546875, -0.0013594627380371094, -0.0012617111206054688, -0.0011639595031738281, -0.0010662078857421875, -0.0009684562683105469, -0.0008707046508789062, -0.0007729530334472656, -0.000675201416015625, -0.0005774497985839844, -0.00047969818115234375, -0.0003819465637207031, -0.0002841949462890625, -0.00018644332885742188, -8.869171142578125e-05, 9.059906005859375e-06, 0.0001068115234375, 0.00020456314086914062, 0.00030231475830078125, 0.0004000663757324219, 0.0004978179931640625, 0.0005955696105957031, 0.0006933212280273438, 0.0007910728454589844, 0.000888824462890625, 0.0009865760803222656, 0.0010843276977539062, 0.0011820793151855469, 0.0012798309326171875, 0.0013775825500488281, 0.0014753341674804688, 0.0015730857849121094, 0.00167083740234375, 0.0017685890197753906, 0.0018663406372070312, 0.001964092254638672, 0.0020618438720703125, 0.002159595489501953, 0.0022573471069335938, 0.0023550987243652344, 0.002452850341796875, 0.0025506019592285156, 0.0026483535766601562, 0.002746105194091797, 0.0028438568115234375, 0.002941608428955078, 0.0030393600463867188, 0.0031371116638183594, 0.00323486328125]}, "gradients/model.layers.25.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 2.0, 0.0, 2.0, 2.0, 2.0, 3.0, 1.0, 3.0, 7.0, 12.0, 12.0, 34.0, 55.0, 156.0, 214.0, 327.0, 468.0, 546.0, 554.0, 484.0, 462.0, 305.0, 199.0, 107.0, 58.0, 28.0, 11.0, 13.0, 5.0, 3.0, 2.0, 3.0, 0.0, 4.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.000949859619140625, -0.0009261369705200195, -0.0009024143218994141, -0.0008786916732788086, -0.0008549690246582031, -0.0008312463760375977, -0.0008075237274169922, -0.0007838010787963867, -0.0007600784301757812, -0.0007363557815551758, -0.0007126331329345703, -0.0006889104843139648, -0.0006651878356933594, -0.0006414651870727539, -0.0006177425384521484, -0.000594019889831543, -0.0005702972412109375, -0.000546574592590332, -0.0005228519439697266, -0.0004991292953491211, -0.0004754066467285156, -0.00045168399810791016, -0.0004279613494873047, -0.0004042387008666992, -0.00038051605224609375, -0.0003567934036254883, -0.0003330707550048828, -0.00030934810638427734, -0.0002856254577636719, -0.0002619028091430664, -0.00023818016052246094, -0.00021445751190185547, -0.00019073486328125, -0.00016701221466064453, -0.00014328956604003906, -0.0001195669174194336, -9.584426879882812e-05, -7.212162017822266e-05, -4.839897155761719e-05, -2.467632293701172e-05, -9.5367431640625e-07, 2.276897430419922e-05, 4.649162292480469e-05, 7.021427154541016e-05, 9.393692016601562e-05, 0.0001176595687866211, 0.00014138221740722656, 0.00016510486602783203, 0.0001888275146484375, 0.00021255016326904297, 0.00023627281188964844, 0.0002599954605102539, 0.0002837181091308594, 0.00030744075775146484, 0.0003311634063720703, 0.0003548860549926758, 0.00037860870361328125, 0.0004023313522338867, 0.0004260540008544922, 0.00044977664947509766, 0.0004734992980957031, 0.0004972219467163086, 0.0005209445953369141, 0.0005446672439575195, 0.000568389892578125]}, "gradients/model.layers.25.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 5.0, 2.0, 4.0, 4.0, 7.0, 13.0, 17.0, 23.0, 47.0, 61.0, 108.0, 180.0, 262.0, 429.0, 731.0, 1333.0, 2464.0, 5412.0, 14840.0, 51917.0, 243090.0, 2382470.0, 11739241.0, 2042951.0, 219989.0, 47016.0, 14052.0, 5021.0, 2435.0, 1267.0, 708.0, 426.0, 253.0, 158.0, 90.0, 59.0, 39.0, 27.0, 17.0, 12.0, 13.0, 6.0, 2.0, 1.0, 3.0, 0.0, 1.0, 2.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.00101470947265625, -0.0009790658950805664, -0.0009434223175048828, -0.0009077787399291992, -0.0008721351623535156, -0.000836491584777832, -0.0008008480072021484, -0.0007652044296264648, -0.0007295608520507812, -0.0006939172744750977, -0.0006582736968994141, -0.0006226301193237305, -0.0005869865417480469, -0.0005513429641723633, -0.0005156993865966797, -0.0004800558090209961, -0.0004444122314453125, -0.0004087686538696289, -0.0003731250762939453, -0.0003374814987182617, -0.0003018379211425781, -0.00026619434356689453, -0.00023055076599121094, -0.00019490718841552734, -0.00015926361083984375, -0.00012362003326416016, -8.797645568847656e-05, -5.233287811279297e-05, -1.6689300537109375e-05, 1.895427703857422e-05, 5.459785461425781e-05, 9.02414321899414e-05, 0.000125885009765625, 0.0001615285873413086, 0.0001971721649169922, 0.00023281574249267578, 0.0002684593200683594, 0.00030410289764404297, 0.00033974647521972656, 0.00037539005279541016, 0.00041103363037109375, 0.00044667720794677734, 0.00048232078552246094, 0.0005179643630981445, 0.0005536079406738281, 0.0005892515182495117, 0.0006248950958251953, 0.0006605386734008789, 0.0006961822509765625, 0.0007318258285522461, 0.0007674694061279297, 0.0008031129837036133, 0.0008387565612792969, 0.0008744001388549805, 0.0009100437164306641, 0.0009456872940063477, 0.0009813308715820312, 0.0010169744491577148, 0.0010526180267333984, 0.001088261604309082, 0.0011239051818847656, 0.0011595487594604492, 0.0011951923370361328, 0.0012308359146118164, 0.0012664794921875]}, "gradients/model.layers.25.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 2.0, 2.0, 5.0, 9.0, 3.0, 8.0, 13.0, 23.0, 26.0, 37.0, 57.0, 93.0, 136.0, 167.0, 257.0, 428.0, 896.0, 2304.0, 9149.0, 547215.0, 3604139.0, 22546.0, 3825.0, 1311.0, 615.0, 325.0, 217.0, 155.0, 91.0, 74.0, 50.0, 30.0, 24.0, 15.0, 17.0, 13.0, 7.0, 6.0, 0.0, 2.0, 0.0, 2.0, 2.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.00537109375, -0.005218505859375, -0.00506591796875, -0.004913330078125, -0.0047607421875, -0.004608154296875, -0.00445556640625, -0.004302978515625, -0.004150390625, -0.003997802734375, -0.00384521484375, -0.003692626953125, -0.0035400390625, -0.003387451171875, -0.00323486328125, -0.003082275390625, -0.0029296875, -0.002777099609375, -0.00262451171875, -0.002471923828125, -0.0023193359375, -0.002166748046875, -0.00201416015625, -0.001861572265625, -0.001708984375, -0.001556396484375, -0.00140380859375, -0.001251220703125, -0.0010986328125, -0.000946044921875, -0.00079345703125, -0.000640869140625, -0.00048828125, -0.000335693359375, -0.00018310546875, -3.0517578125e-05, 0.0001220703125, 0.000274658203125, 0.00042724609375, 0.000579833984375, 0.000732421875, 0.000885009765625, 0.00103759765625, 0.001190185546875, 0.0013427734375, 0.001495361328125, 0.00164794921875, 0.001800537109375, 0.001953125, 0.002105712890625, 0.00225830078125, 0.002410888671875, 0.0025634765625, 0.002716064453125, 0.00286865234375, 0.003021240234375, 0.003173828125, 0.003326416015625, 0.00347900390625, 0.003631591796875, 0.0037841796875, 0.003936767578125, 0.00408935546875, 0.004241943359375, 0.00439453125]}, "gradients/model.layers.25.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 0.0, 2.0, 0.0, 1.0, 3.0, 0.0, 5.0, 5.0, 4.0, 3.0, 10.0, 10.0, 7.0, 20.0, 16.0, 26.0, 30.0, 73.0, 103.0, 175.0, 311.0, 707.0, 1603.0, 4567.0, 21600.0, 3789322.0, 358427.0, 11714.0, 3081.0, 1244.0, 547.0, 282.0, 152.0, 75.0, 41.0, 46.0, 18.0, 17.0, 11.0, 4.0, 6.0, 8.0, 3.0, 3.0, 7.0, 4.0, 0.0, 2.0, 0.0, 2.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00183868408203125, -0.001774907112121582, -0.001711130142211914, -0.001647353172302246, -0.0015835762023925781, -0.0015197992324829102, -0.0014560222625732422, -0.0013922452926635742, -0.0013284683227539062, -0.0012646913528442383, -0.0012009143829345703, -0.0011371374130249023, -0.0010733604431152344, -0.0010095834732055664, -0.0009458065032958984, -0.0008820295333862305, -0.0008182525634765625, -0.0007544755935668945, -0.0006906986236572266, -0.0006269216537475586, -0.0005631446838378906, -0.0004993677139282227, -0.0004355907440185547, -0.0003718137741088867, -0.00030803680419921875, -0.0002442598342895508, -0.0001804828643798828, -0.00011670589447021484, -5.2928924560546875e-05, 1.0848045349121094e-05, 7.462501525878906e-05, 0.00013840198516845703, 0.000202178955078125, 0.00026595592498779297, 0.00032973289489746094, 0.0003935098648071289, 0.0004572868347167969, 0.0005210638046264648, 0.0005848407745361328, 0.0006486177444458008, 0.0007123947143554688, 0.0007761716842651367, 0.0008399486541748047, 0.0009037256240844727, 0.0009675025939941406, 0.0010312795639038086, 0.0010950565338134766, 0.0011588335037231445, 0.0012226104736328125, 0.0012863874435424805, 0.0013501644134521484, 0.0014139413833618164, 0.0014777183532714844, 0.0015414953231811523, 0.0016052722930908203, 0.0016690492630004883, 0.0017328262329101562, 0.0017966032028198242, 0.0018603801727294922, 0.0019241571426391602, 0.001987934112548828, 0.002051711082458496, 0.002115488052368164, 0.002179265022277832, 0.0022430419921875]}, "gradients/model.layers.25.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 0.0, 1.0, 2.0, 2.0, 2.0, 0.0, 3.0, 0.0, 6.0, 1.0, 5.0, 5.0, 4.0, 5.0, 7.0, 6.0, 8.0, 11.0, 11.0, 6.0, 23.0, 36.0, 50.0, 84.0, 190.0, 478.0, 1529.0, 5514.0, 8377002.0, 8384069.0, 5677.0, 1517.0, 501.0, 165.0, 90.0, 56.0, 29.0, 22.0, 13.0, 13.0, 10.0, 11.0, 4.0, 5.0, 5.0, 6.0, 6.0, 2.0, 4.0, 2.0, 3.0, 1.0, 1.0, 1.0, 2.0, 3.0, 0.0, 1.0, 0.0, 1.0, 2.0], "bins": [-0.00653076171875, -0.0063266754150390625, -0.006122589111328125, -0.0059185028076171875, -0.00571441650390625, -0.0055103302001953125, -0.005306243896484375, -0.0051021575927734375, -0.0048980712890625, -0.0046939849853515625, -0.004489898681640625, -0.0042858123779296875, -0.00408172607421875, -0.0038776397705078125, -0.003673553466796875, -0.0034694671630859375, -0.003265380859375, -0.0030612945556640625, -0.002857208251953125, -0.0026531219482421875, -0.00244903564453125, -0.0022449493408203125, -0.002040863037109375, -0.0018367767333984375, -0.0016326904296875, -0.0014286041259765625, -0.001224517822265625, -0.0010204315185546875, -0.00081634521484375, -0.0006122589111328125, -0.000408172607421875, -0.0002040863037109375, 0.0, 0.0002040863037109375, 0.000408172607421875, 0.0006122589111328125, 0.00081634521484375, 0.0010204315185546875, 0.001224517822265625, 0.0014286041259765625, 0.0016326904296875, 0.0018367767333984375, 0.002040863037109375, 0.0022449493408203125, 0.00244903564453125, 0.0026531219482421875, 0.002857208251953125, 0.0030612945556640625, 0.003265380859375, 0.0034694671630859375, 0.003673553466796875, 0.0038776397705078125, 0.00408172607421875, 0.0042858123779296875, 0.004489898681640625, 0.0046939849853515625, 0.0048980712890625, 0.0051021575927734375, 0.005306243896484375, 0.0055103302001953125, 0.00571441650390625, 0.0059185028076171875, 0.006122589111328125, 0.0063266754150390625, 0.00653076171875]}, "gradients/model.layers.25.input_layernorm.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 2.0, 1.0, 1.0, 1.0, 1.0, 4.0, 2.0, 3.0, 2.0, 8.0, 12.0, 37.0, 319.0, 1265.0, 1647.0, 622.0, 115.0, 18.0, 8.0, 6.0, 3.0, 4.0, 2.0, 0.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.001495361328125, -0.0014622807502746582, -0.0014292001724243164, -0.0013961195945739746, -0.0013630390167236328, -0.001329958438873291, -0.0012968778610229492, -0.0012637972831726074, -0.0012307167053222656, -0.0011976361274719238, -0.001164555549621582, -0.0011314749717712402, -0.0010983943939208984, -0.0010653138160705566, -0.0010322332382202148, -0.000999152660369873, -0.0009660720825195312, -0.0009329915046691895, -0.0008999109268188477, -0.0008668303489685059, -0.0008337497711181641, -0.0008006691932678223, -0.0007675886154174805, -0.0007345080375671387, -0.0007014274597167969, -0.0006683468818664551, -0.0006352663040161133, -0.0006021857261657715, -0.0005691051483154297, -0.0005360245704650879, -0.0005029439926147461, -0.0004698634147644043, -0.0004367828369140625, -0.0004037022590637207, -0.0003706216812133789, -0.0003375411033630371, -0.0003044605255126953, -0.0002713799476623535, -0.00023829936981201172, -0.00020521879196166992, -0.00017213821411132812, -0.00013905763626098633, -0.00010597705841064453, -7.289648056030273e-05, -3.981590270996094e-05, -6.735324859619141e-06, 2.6345252990722656e-05, 5.942583084106445e-05, 9.250640869140625e-05, 0.00012558698654174805, 0.00015866756439208984, 0.00019174814224243164, 0.00022482872009277344, 0.00025790929794311523, 0.00029098987579345703, 0.00032407045364379883, 0.0003571510314941406, 0.0003902316093444824, 0.0004233121871948242, 0.000456392765045166, 0.0004894733428955078, 0.0005225539207458496, 0.0005556344985961914, 0.0005887150764465332, 0.000621795654296875]}, "gradients/model.layers.24.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 0.0, 2.0, 1.0, 4.0, 6.0, 9.0, 16.0, 30.0, 26.0, 43.0, 85.0, 118.0, 174.0, 327.0, 560.0, 1136.0, 3020.0, 14357.0, 162167.0, 16777216.0, 16777216.0, 235002.0, 18481.0, 3629.0, 1141.0, 639.0, 326.0, 199.0, 128.0, 105.0, 55.0, 37.0, 21.0, 19.0, 8.0, 7.0, 8.0, 3.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0029754638671875, -0.0028905868530273438, -0.0028057098388671875, -0.0027208328247070312, -0.002635955810546875, -0.0025510787963867188, -0.0024662017822265625, -0.0023813247680664062, -0.00229644775390625, -0.0022115707397460938, -0.0021266937255859375, -0.0020418167114257812, -0.001956939697265625, -0.0018720626831054688, -0.0017871856689453125, -0.0017023086547851562, -0.001617431640625, -0.0015325546264648438, -0.0014476776123046875, -0.0013628005981445312, -0.001277923583984375, -0.0011930465698242188, -0.0011081695556640625, -0.0010232925415039062, -0.00093841552734375, -0.0008535385131835938, -0.0007686614990234375, -0.0006837844848632812, -0.000598907470703125, -0.0005140304565429688, -0.0004291534423828125, -0.00034427642822265625, -0.0002593994140625, -0.00017452239990234375, -8.96453857421875e-05, -4.76837158203125e-06, 8.0108642578125e-05, 0.00016498565673828125, 0.0002498626708984375, 0.00033473968505859375, 0.00041961669921875, 0.0005044937133789062, 0.0005893707275390625, 0.0006742477416992188, 0.000759124755859375, 0.0008440017700195312, 0.0009288787841796875, 0.0010137557983398438, 0.0010986328125, 0.0011835098266601562, 0.0012683868408203125, 0.0013532638549804688, 0.001438140869140625, 0.0015230178833007812, 0.0016078948974609375, 0.0016927719116210938, 0.00177764892578125, 0.0018625259399414062, 0.0019474029541015625, 0.0020322799682617188, 0.002117156982421875, 0.0022020339965820312, 0.0022869110107421875, 0.0023717880249023438, 0.0024566650390625]}, "gradients/model.layers.24.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 2.0, 2.0, 1.0, 3.0, 3.0, 1.0, 4.0, 11.0, 8.0, 15.0, 12.0, 20.0, 34.0, 68.0, 88.0, 201.0, 414.0, 1179.0, 4312.0, 25193.0, 429212.0, 16777216.0, 1064172.0, 38413.0, 5737.0, 1491.0, 501.0, 230.0, 100.0, 71.0, 34.0, 31.0, 15.0, 11.0, 6.0, 6.0, 5.0, 2.0, 2.0, 3.0, 1.0, 3.0, 0.0, 2.0, 0.0, 3.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0], "bins": [-0.0040283203125, -0.003908395767211914, -0.003788471221923828, -0.003668546676635742, -0.0035486221313476562, -0.0034286975860595703, -0.0033087730407714844, -0.0031888484954833984, -0.0030689239501953125, -0.0029489994049072266, -0.0028290748596191406, -0.0027091503143310547, -0.0025892257690429688, -0.002469301223754883, -0.002349376678466797, -0.002229452133178711, -0.002109527587890625, -0.001989603042602539, -0.0018696784973144531, -0.0017497539520263672, -0.0016298294067382812, -0.0015099048614501953, -0.0013899803161621094, -0.0012700557708740234, -0.0011501312255859375, -0.0010302066802978516, -0.0009102821350097656, -0.0007903575897216797, -0.0006704330444335938, -0.0005505084991455078, -0.0004305839538574219, -0.00031065940856933594, -0.00019073486328125, -7.081031799316406e-05, 4.9114227294921875e-05, 0.0001690387725830078, 0.00028896331787109375, 0.0004088878631591797, 0.0005288124084472656, 0.0006487369537353516, 0.0007686614990234375, 0.0008885860443115234, 0.0010085105895996094, 0.0011284351348876953, 0.0012483596801757812, 0.0013682842254638672, 0.0014882087707519531, 0.001608133316040039, 0.001728057861328125, 0.001847982406616211, 0.001967906951904297, 0.002087831497192383, 0.0022077560424804688, 0.0023276805877685547, 0.0024476051330566406, 0.0025675296783447266, 0.0026874542236328125, 0.0028073787689208984, 0.0029273033142089844, 0.0030472278594970703, 0.0031671524047851562, 0.003287076950073242, 0.003407001495361328, 0.003526926040649414, 0.0036468505859375]}, "gradients/model.layers.24.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 3.0, 0.0, 1.0, 1.0, 4.0, 5.0, 6.0, 27.0, 91.0, 339.0, 863.0, 1216.0, 949.0, 402.0, 130.0, 38.0, 6.0, 3.0, 4.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.001922607421875, -0.001865386962890625, -0.00180816650390625, -0.001750946044921875, -0.0016937255859375, -0.001636505126953125, -0.00157928466796875, -0.001522064208984375, -0.00146484375, -0.001407623291015625, -0.00135040283203125, -0.001293182373046875, -0.0012359619140625, -0.001178741455078125, -0.00112152099609375, -0.001064300537109375, -0.001007080078125, -0.000949859619140625, -0.00089263916015625, -0.000835418701171875, -0.0007781982421875, -0.000720977783203125, -0.00066375732421875, -0.000606536865234375, -0.00054931640625, -0.000492095947265625, -0.00043487548828125, -0.000377655029296875, -0.0003204345703125, -0.000263214111328125, -0.00020599365234375, -0.000148773193359375, -9.1552734375e-05, -3.4332275390625e-05, 2.288818359375e-05, 8.0108642578125e-05, 0.0001373291015625, 0.000194549560546875, 0.00025177001953125, 0.000308990478515625, 0.0003662109375, 0.000423431396484375, 0.00048065185546875, 0.000537872314453125, 0.0005950927734375, 0.000652313232421875, 0.00070953369140625, 0.000766754150390625, 0.000823974609375, 0.000881195068359375, 0.00093841552734375, 0.000995635986328125, 0.0010528564453125, 0.001110076904296875, 0.00116729736328125, 0.001224517822265625, 0.00128173828125, 0.001338958740234375, 0.00139617919921875, 0.001453399658203125, 0.0015106201171875, 0.001567840576171875, 0.00162506103515625, 0.001682281494140625, 0.001739501953125]}, "gradients/model.layers.24.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 2.0, 1.0, 2.0, 3.0, 6.0, 9.0, 2.0, 18.0, 23.0, 36.0, 60.0, 105.0, 226.0, 373.0, 787.0, 1487.0, 3155.0, 8305.0, 35471.0, 455505.0, 10739287.0, 5275930.0, 221347.0, 23042.0, 6601.0, 2676.0, 1302.0, 650.0, 361.0, 174.0, 105.0, 53.0, 37.0, 22.0, 16.0, 12.0, 4.0, 3.0, 2.0, 4.0, 0.0, 2.0, 1.0, 0.0, 0.0, 1.0, 1.0, 2.0], "bins": [-0.00170135498046875, -0.0016552209854125977, -0.0016090869903564453, -0.001562952995300293, -0.0015168190002441406, -0.0014706850051879883, -0.001424551010131836, -0.0013784170150756836, -0.0013322830200195312, -0.001286149024963379, -0.0012400150299072266, -0.0011938810348510742, -0.0011477470397949219, -0.0011016130447387695, -0.0010554790496826172, -0.0010093450546264648, -0.0009632110595703125, -0.0009170770645141602, -0.0008709430694580078, -0.0008248090744018555, -0.0007786750793457031, -0.0007325410842895508, -0.0006864070892333984, -0.0006402730941772461, -0.0005941390991210938, -0.0005480051040649414, -0.0005018711090087891, -0.0004557371139526367, -0.0004096031188964844, -0.00036346912384033203, -0.0003173351287841797, -0.00027120113372802734, -0.000225067138671875, -0.00017893314361572266, -0.0001327991485595703, -8.666515350341797e-05, -4.0531158447265625e-05, 5.602836608886719e-06, 5.173683166503906e-05, 9.78708267211914e-05, 0.00014400482177734375, 0.0001901388168334961, 0.00023627281188964844, 0.0002824068069458008, 0.0003285408020019531, 0.00037467479705810547, 0.0004208087921142578, 0.00046694278717041016, 0.0005130767822265625, 0.0005592107772827148, 0.0006053447723388672, 0.0006514787673950195, 0.0006976127624511719, 0.0007437467575073242, 0.0007898807525634766, 0.0008360147476196289, 0.0008821487426757812, 0.0009282827377319336, 0.0009744167327880859, 0.0010205507278442383, 0.0010666847229003906, 0.001112818717956543, 0.0011589527130126953, 0.0012050867080688477, 0.001251220703125]}, "gradients/model.layers.24.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 2.0, 3.0, 2.0, 3.0, 2.0, 9.0, 5.0, 9.0, 11.0, 21.0, 26.0, 39.0, 44.0, 80.0, 80.0, 105.0, 159.0, 199.0, 261.0, 401.0, 691.0, 1611.0, 4025.0, 18055.0, 1366012.0, 2767815.0, 25693.0, 4911.0, 1715.0, 815.0, 413.0, 285.0, 197.0, 147.0, 116.0, 90.0, 61.0, 39.0, 29.0, 24.0, 19.0, 22.0, 13.0, 10.0, 4.0, 7.0, 3.0, 5.0, 0.0, 1.0, 2.0, 3.0, 0.0, 1.0, 3.0], "bins": [-0.0042724609375, -0.004147052764892578, -0.004021644592285156, -0.0038962364196777344, -0.0037708282470703125, -0.0036454200744628906, -0.0035200119018554688, -0.003394603729248047, -0.003269195556640625, -0.003143787384033203, -0.0030183792114257812, -0.0028929710388183594, -0.0027675628662109375, -0.0026421546936035156, -0.0025167465209960938, -0.002391338348388672, -0.00226593017578125, -0.002140522003173828, -0.0020151138305664062, -0.0018897056579589844, -0.0017642974853515625, -0.0016388893127441406, -0.0015134811401367188, -0.0013880729675292969, -0.001262664794921875, -0.0011372566223144531, -0.0010118484497070312, -0.0008864402770996094, -0.0007610321044921875, -0.0006356239318847656, -0.0005102157592773438, -0.0003848075866699219, -0.0002593994140625, -0.00013399124145507812, -8.58306884765625e-06, 0.00011682510375976562, 0.0002422332763671875, 0.0003676414489746094, 0.0004930496215820312, 0.0006184577941894531, 0.000743865966796875, 0.0008692741394042969, 0.0009946823120117188, 0.0011200904846191406, 0.0012454986572265625, 0.0013709068298339844, 0.0014963150024414062, 0.0016217231750488281, 0.00174713134765625, 0.0018725395202636719, 0.0019979476928710938, 0.0021233558654785156, 0.0022487640380859375, 0.0023741722106933594, 0.0024995803833007812, 0.002624988555908203, 0.002750396728515625, 0.002875804901123047, 0.0030012130737304688, 0.0031266212463378906, 0.0032520294189453125, 0.0033774375915527344, 0.0035028457641601562, 0.003628253936767578, 0.003753662109375]}, "gradients/model.layers.24.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 0.0, 5.0, 7.0, 6.0, 9.0, 3.0, 9.0, 6.0, 16.0, 35.0, 42.0, 52.0, 69.0, 131.0, 295.0, 613.0, 1513.0, 4668.0, 20155.0, 2068294.0, 2070785.0, 19978.0, 4559.0, 1615.0, 676.0, 301.0, 150.0, 86.0, 60.0, 41.0, 31.0, 17.0, 9.0, 12.0, 7.0, 13.0, 4.0, 5.0, 5.0, 1.0, 0.0, 1.0, 0.0, 2.0, 3.0, 2.0, 0.0, 0.0, 0.0, 1.0, 3.0, 1.0], "bins": [-0.00160980224609375, -0.0015594959259033203, -0.0015091896057128906, -0.001458883285522461, -0.0014085769653320312, -0.0013582706451416016, -0.0013079643249511719, -0.0012576580047607422, -0.0012073516845703125, -0.0011570453643798828, -0.0011067390441894531, -0.0010564327239990234, -0.0010061264038085938, -0.0009558200836181641, -0.0009055137634277344, -0.0008552074432373047, -0.000804901123046875, -0.0007545948028564453, -0.0007042884826660156, -0.0006539821624755859, -0.0006036758422851562, -0.0005533695220947266, -0.0005030632019042969, -0.0004527568817138672, -0.0004024505615234375, -0.0003521442413330078, -0.0003018379211425781, -0.00025153160095214844, -0.00020122528076171875, -0.00015091896057128906, -0.00010061264038085938, -5.030632019042969e-05, 0.0, 5.030632019042969e-05, 0.00010061264038085938, 0.00015091896057128906, 0.00020122528076171875, 0.00025153160095214844, 0.0003018379211425781, 0.0003521442413330078, 0.0004024505615234375, 0.0004527568817138672, 0.0005030632019042969, 0.0005533695220947266, 0.0006036758422851562, 0.0006539821624755859, 0.0007042884826660156, 0.0007545948028564453, 0.000804901123046875, 0.0008552074432373047, 0.0009055137634277344, 0.0009558200836181641, 0.0010061264038085938, 0.0010564327239990234, 0.0011067390441894531, 0.0011570453643798828, 0.0012073516845703125, 0.0012576580047607422, 0.0013079643249511719, 0.0013582706451416016, 0.0014085769653320312, 0.001458883285522461, 0.0015091896057128906, 0.0015594959259033203, 0.00160980224609375]}, "gradients/model.layers.24.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 1.0, 0.0, 2.0, 7.0, 7.0, 3.0, 7.0, 9.0, 10.0, 14.0, 21.0, 32.0, 49.0, 74.0, 108.0, 176.0, 318.0, 588.0, 1432.0, 4253.0, 17037.0, 180550.0, 16329386.0, 217041.0, 18654.0, 4523.0, 1470.0, 643.0, 315.0, 179.0, 90.0, 60.0, 43.0, 26.0, 20.0, 14.0, 9.0, 11.0, 10.0, 6.0, 4.0, 2.0, 2.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.00140380859375, -0.0013508796691894531, -0.0012979507446289062, -0.0012450218200683594, -0.0011920928955078125, -0.0011391639709472656, -0.0010862350463867188, -0.0010333061218261719, -0.000980377197265625, -0.0009274482727050781, -0.0008745193481445312, -0.0008215904235839844, -0.0007686614990234375, -0.0007157325744628906, -0.0006628036499023438, -0.0006098747253417969, -0.00055694580078125, -0.0005040168762207031, -0.00045108795166015625, -0.0003981590270996094, -0.0003452301025390625, -0.0002923011779785156, -0.00023937225341796875, -0.00018644332885742188, -0.000133514404296875, -8.058547973632812e-05, -2.765655517578125e-05, 2.5272369384765625e-05, 7.82012939453125e-05, 0.00013113021850585938, 0.00018405914306640625, 0.00023698806762695312, 0.0002899169921875, 0.0003428459167480469, 0.00039577484130859375, 0.0004487037658691406, 0.0005016326904296875, 0.0005545616149902344, 0.0006074905395507812, 0.0006604194641113281, 0.000713348388671875, 0.0007662773132324219, 0.0008192062377929688, 0.0008721351623535156, 0.0009250640869140625, 0.0009779930114746094, 0.0010309219360351562, 0.0010838508605957031, 0.00113677978515625, 0.0011897087097167969, 0.0012426376342773438, 0.0012955665588378906, 0.0013484954833984375, 0.0014014244079589844, 0.0014543533325195312, 0.0015072822570800781, 0.001560211181640625, 0.0016131401062011719, 0.0016660690307617188, 0.0017189979553222656, 0.0017719268798828125, 0.0018248558044433594, 0.0018777847290039062, 0.0019307136535644531, 0.001983642578125]}, "gradients/model.layers.24.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 3.0, 1.0, 4.0, 3.0, 8.0, 9.0, 18.0, 55.0, 433.0, 1754.0, 1439.0, 297.0, 25.0, 12.0, 5.0, 7.0, 3.0, 1.0, 5.0, 2.0, 1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00165557861328125, -0.0016195178031921387, -0.0015834569931030273, -0.001547396183013916, -0.0015113353729248047, -0.0014752745628356934, -0.001439213752746582, -0.0014031529426574707, -0.0013670921325683594, -0.001331031322479248, -0.0012949705123901367, -0.0012589097023010254, -0.001222848892211914, -0.0011867880821228027, -0.0011507272720336914, -0.00111466646194458, -0.0010786056518554688, -0.0010425448417663574, -0.001006484031677246, -0.0009704232215881348, -0.0009343624114990234, -0.0008983016014099121, -0.0008622407913208008, -0.0008261799812316895, -0.0007901191711425781, -0.0007540583610534668, -0.0007179975509643555, -0.0006819367408752441, -0.0006458759307861328, -0.0006098151206970215, -0.0005737543106079102, -0.0005376935005187988, -0.0005016326904296875, -0.00046557188034057617, -0.00042951107025146484, -0.0003934502601623535, -0.0003573894500732422, -0.00032132863998413086, -0.00028526782989501953, -0.0002492070198059082, -0.00021314620971679688, -0.00017708539962768555, -0.00014102458953857422, -0.00010496377944946289, -6.890296936035156e-05, -3.2842159271240234e-05, 3.2186508178710938e-06, 3.927946090698242e-05, 7.534027099609375e-05, 0.00011140108108520508, 0.0001474618911743164, 0.00018352270126342773, 0.00021958351135253906, 0.0002556443214416504, 0.0002917051315307617, 0.00032776594161987305, 0.0003638267517089844, 0.0003998875617980957, 0.00043594837188720703, 0.00047200918197631836, 0.0005080699920654297, 0.000544130802154541, 0.0005801916122436523, 0.0006162524223327637, 0.000652313232421875]}, "gradients/model.layers.23.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 4.0, 0.0, 2.0, 1.0, 5.0, 1.0, 8.0, 11.0, 24.0, 46.0, 55.0, 86.0, 156.0, 249.0, 477.0, 831.0, 1638.0, 4644.0, 20052.0, 163122.0, 6821672.0, 16777216.0, 1610891.0, 81619.0, 12522.0, 3266.0, 1267.0, 688.0, 370.0, 223.0, 155.0, 104.0, 76.0, 37.0, 29.0, 13.0, 5.0, 5.0, 1.0, 1.0, 2.0, 4.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.0026092529296875, -0.002539396286010742, -0.0024695396423339844, -0.0023996829986572266, -0.0023298263549804688, -0.002259969711303711, -0.002190113067626953, -0.0021202564239501953, -0.0020503997802734375, -0.0019805431365966797, -0.0019106864929199219, -0.001840829849243164, -0.0017709732055664062, -0.0017011165618896484, -0.0016312599182128906, -0.0015614032745361328, -0.001491546630859375, -0.0014216899871826172, -0.0013518333435058594, -0.0012819766998291016, -0.0012121200561523438, -0.001142263412475586, -0.0010724067687988281, -0.0010025501251220703, -0.0009326934814453125, -0.0008628368377685547, -0.0007929801940917969, -0.0007231235504150391, -0.0006532669067382812, -0.0005834102630615234, -0.0005135536193847656, -0.0004436969757080078, -0.00037384033203125, -0.0003039836883544922, -0.00023412704467773438, -0.00016427040100097656, -9.441375732421875e-05, -2.4557113647460938e-05, 4.5299530029296875e-05, 0.00011515617370605469, 0.0001850128173828125, 0.0002548694610595703, 0.0003247261047363281, 0.00039458274841308594, 0.00046443939208984375, 0.0005342960357666016, 0.0006041526794433594, 0.0006740093231201172, 0.000743865966796875, 0.0008137226104736328, 0.0008835792541503906, 0.0009534358978271484, 0.0010232925415039062, 0.001093149185180664, 0.0011630058288574219, 0.0012328624725341797, 0.0013027191162109375, 0.0013725757598876953, 0.0014424324035644531, 0.001512289047241211, 0.0015821456909179688, 0.0016520023345947266, 0.0017218589782714844, 0.0017917156219482422, 0.001861572265625]}, "gradients/model.layers.23.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 2.0, 1.0, 0.0, 2.0, 6.0, 9.0, 19.0, 19.0, 39.0, 57.0, 107.0, 250.0, 869.0, 4254.0, 37848.0, 7196554.0, 16777216.0, 102903.0, 7892.0, 1381.0, 361.0, 133.0, 78.0, 46.0, 25.0, 14.0, 9.0, 6.0, 2.0, 6.0, 5.0, 5.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.004547119140625, -0.004391193389892578, -0.004235267639160156, -0.004079341888427734, -0.0039234161376953125, -0.0037674903869628906, -0.0036115646362304688, -0.003455638885498047, -0.003299713134765625, -0.003143787384033203, -0.0029878616333007812, -0.0028319358825683594, -0.0026760101318359375, -0.0025200843811035156, -0.0023641586303710938, -0.002208232879638672, -0.00205230712890625, -0.0018963813781738281, -0.0017404556274414062, -0.0015845298767089844, -0.0014286041259765625, -0.0012726783752441406, -0.0011167526245117188, -0.0009608268737792969, -0.000804901123046875, -0.0006489753723144531, -0.0004930496215820312, -0.0003371238708496094, -0.0001811981201171875, -2.5272369384765625e-05, 0.00013065338134765625, 0.0002865791320800781, 0.0004425048828125, 0.0005984306335449219, 0.0007543563842773438, 0.0009102821350097656, 0.0010662078857421875, 0.0012221336364746094, 0.0013780593872070312, 0.0015339851379394531, 0.001689910888671875, 0.0018458366394042969, 0.0020017623901367188, 0.0021576881408691406, 0.0023136138916015625, 0.0024695396423339844, 0.0026254653930664062, 0.002781391143798828, 0.00293731689453125, 0.003093242645263672, 0.0032491683959960938, 0.0034050941467285156, 0.0035610198974609375, 0.0037169456481933594, 0.0038728713989257812, 0.004028797149658203, 0.004184722900390625, 0.004340648651123047, 0.004496574401855469, 0.004652500152587891, 0.0048084259033203125, 0.004964351654052734, 0.005120277404785156, 0.005276203155517578, 0.00543212890625]}, "gradients/model.layers.23.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 5.0, 2.0, 1.0, 4.0, 20.0, 66.0, 224.0, 575.0, 905.0, 1052.0, 721.0, 338.0, 111.0, 35.0, 17.0, 5.0, 3.0, 3.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.001251220703125, -0.0011973381042480469, -0.0011434555053710938, -0.0010895729064941406, -0.0010356903076171875, -0.0009818077087402344, -0.0009279251098632812, -0.0008740425109863281, -0.000820159912109375, -0.0007662773132324219, -0.0007123947143554688, -0.0006585121154785156, -0.0006046295166015625, -0.0005507469177246094, -0.0004968643188476562, -0.0004429817199707031, -0.00038909912109375, -0.0003352165222167969, -0.00028133392333984375, -0.00022745132446289062, -0.0001735687255859375, -0.00011968612670898438, -6.580352783203125e-05, -1.1920928955078125e-05, 4.1961669921875e-05, 9.584426879882812e-05, 0.00014972686767578125, 0.00020360946655273438, 0.0002574920654296875, 0.0003113746643066406, 0.00036525726318359375, 0.0004191398620605469, 0.0004730224609375, 0.0005269050598144531, 0.0005807876586914062, 0.0006346702575683594, 0.0006885528564453125, 0.0007424354553222656, 0.0007963180541992188, 0.0008502006530761719, 0.000904083251953125, 0.0009579658508300781, 0.0010118484497070312, 0.0010657310485839844, 0.0011196136474609375, 0.0011734962463378906, 0.0012273788452148438, 0.0012812614440917969, 0.00133514404296875, 0.0013890266418457031, 0.0014429092407226562, 0.0014967918395996094, 0.0015506744384765625, 0.0016045570373535156, 0.0016584396362304688, 0.0017123222351074219, 0.001766204833984375, 0.0018200874328613281, 0.0018739700317382812, 0.0019278526306152344, 0.0019817352294921875, 0.0020356178283691406, 0.0020895004272460938, 0.002143383026123047, 0.002197265625]}, "gradients/model.layers.23.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 3.0, 5.0, 3.0, 4.0, 6.0, 16.0, 22.0, 35.0, 57.0, 105.0, 167.0, 299.0, 519.0, 1104.0, 2894.0, 31018.0, 2948959.0, 13601809.0, 181154.0, 5974.0, 1509.0, 722.0, 348.0, 187.0, 103.0, 87.0, 39.0, 27.0, 14.0, 8.0, 4.0, 4.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0023040771484375, -0.0022346973419189453, -0.0021653175354003906, -0.002095937728881836, -0.0020265579223632812, -0.0019571781158447266, -0.0018877983093261719, -0.0018184185028076172, -0.0017490386962890625, -0.0016796588897705078, -0.0016102790832519531, -0.0015408992767333984, -0.0014715194702148438, -0.001402139663696289, -0.0013327598571777344, -0.0012633800506591797, -0.001194000244140625, -0.0011246204376220703, -0.0010552406311035156, -0.000985860824584961, -0.0009164810180664062, -0.0008471012115478516, -0.0007777214050292969, -0.0007083415985107422, -0.0006389617919921875, -0.0005695819854736328, -0.0005002021789550781, -0.00043082237243652344, -0.00036144256591796875, -0.00029206275939941406, -0.00022268295288085938, -0.0001533031463623047, -8.392333984375e-05, -1.4543533325195312e-05, 5.4836273193359375e-05, 0.00012421607971191406, 0.00019359588623046875, 0.00026297569274902344, 0.0003323554992675781, 0.0004017353057861328, 0.0004711151123046875, 0.0005404949188232422, 0.0006098747253417969, 0.0006792545318603516, 0.0007486343383789062, 0.0008180141448974609, 0.0008873939514160156, 0.0009567737579345703, 0.001026153564453125, 0.0010955333709716797, 0.0011649131774902344, 0.001234292984008789, 0.0013036727905273438, 0.0013730525970458984, 0.0014424324035644531, 0.0015118122100830078, 0.0015811920166015625, 0.0016505718231201172, 0.0017199516296386719, 0.0017893314361572266, 0.0018587112426757812, 0.001928091049194336, 0.0019974708557128906, 0.0020668506622314453, 0.00213623046875]}, "gradients/model.layers.23.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 5.0, 3.0, 6.0, 10.0, 13.0, 18.0, 19.0, 13.0, 26.0, 43.0, 45.0, 57.0, 88.0, 101.0, 137.0, 193.0, 279.0, 398.0, 725.0, 1433.0, 3755.0, 17180.0, 771896.0, 3338878.0, 48601.0, 5817.0, 2079.0, 838.0, 537.0, 269.0, 193.0, 131.0, 107.0, 86.0, 69.0, 53.0, 55.0, 43.0, 19.0, 25.0, 21.0, 10.0, 5.0, 3.0, 4.0, 1.0, 2.0, 2.0, 3.0, 2.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.004302978515625, -0.00417327880859375, -0.0040435791015625, -0.00391387939453125, -0.0037841796875, -0.00365447998046875, -0.0035247802734375, -0.00339508056640625, -0.003265380859375, -0.00313568115234375, -0.0030059814453125, -0.00287628173828125, -0.00274658203125, -0.00261688232421875, -0.0024871826171875, -0.00235748291015625, -0.002227783203125, -0.00209808349609375, -0.0019683837890625, -0.00183868408203125, -0.001708984375, -0.00157928466796875, -0.0014495849609375, -0.00131988525390625, -0.001190185546875, -0.00106048583984375, -0.0009307861328125, -0.00080108642578125, -0.00067138671875, -0.00054168701171875, -0.0004119873046875, -0.00028228759765625, -0.000152587890625, -2.288818359375e-05, 0.0001068115234375, 0.00023651123046875, 0.0003662109375, 0.00049591064453125, 0.0006256103515625, 0.00075531005859375, 0.000885009765625, 0.00101470947265625, 0.0011444091796875, 0.00127410888671875, 0.00140380859375, 0.00153350830078125, 0.0016632080078125, 0.00179290771484375, 0.001922607421875, 0.00205230712890625, 0.0021820068359375, 0.00231170654296875, 0.00244140625, 0.00257110595703125, 0.0027008056640625, 0.00283050537109375, 0.002960205078125, 0.00308990478515625, 0.0032196044921875, 0.00334930419921875, 0.00347900390625, 0.00360870361328125, 0.0037384033203125, 0.00386810302734375, 0.003997802734375]}, "gradients/model.layers.23.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 2.0, 0.0, 0.0, 3.0, 3.0, 4.0, 5.0, 6.0, 11.0, 11.0, 11.0, 14.0, 32.0, 63.0, 147.0, 322.0, 834.0, 3021.0, 32364.0, 4114441.0, 38303.0, 3192.0, 890.0, 306.0, 126.0, 68.0, 35.0, 17.0, 12.0, 11.0, 10.0, 5.0, 4.0, 4.0, 2.0, 2.0, 2.0, 2.0, 0.0, 3.0, 2.0, 2.0, 0.0, 1.0, 2.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.0028228759765625, -0.002741098403930664, -0.002659320831298828, -0.002577543258666992, -0.0024957656860351562, -0.0024139881134033203, -0.0023322105407714844, -0.0022504329681396484, -0.0021686553955078125, -0.0020868778228759766, -0.0020051002502441406, -0.0019233226776123047, -0.0018415451049804688, -0.0017597675323486328, -0.0016779899597167969, -0.001596212387084961, -0.001514434814453125, -0.001432657241821289, -0.0013508796691894531, -0.0012691020965576172, -0.0011873245239257812, -0.0011055469512939453, -0.0010237693786621094, -0.0009419918060302734, -0.0008602142333984375, -0.0007784366607666016, -0.0006966590881347656, -0.0006148815155029297, -0.0005331039428710938, -0.0004513263702392578, -0.0003695487976074219, -0.00028777122497558594, -0.00020599365234375, -0.00012421607971191406, -4.2438507080078125e-05, 3.933906555175781e-05, 0.00012111663818359375, 0.0002028942108154297, 0.0002846717834472656, 0.00036644935607910156, 0.0004482269287109375, 0.0005300045013427734, 0.0006117820739746094, 0.0006935596466064453, 0.0007753372192382812, 0.0008571147918701172, 0.0009388923645019531, 0.001020669937133789, 0.001102447509765625, 0.001184225082397461, 0.0012660026550292969, 0.0013477802276611328, 0.0014295578002929688, 0.0015113353729248047, 0.0015931129455566406, 0.0016748905181884766, 0.0017566680908203125, 0.0018384456634521484, 0.0019202232360839844, 0.0020020008087158203, 0.0020837783813476562, 0.002165555953979492, 0.002247333526611328, 0.002329111099243164, 0.002410888671875]}, "gradients/model.layers.23.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 4.0, 2.0, 3.0, 5.0, 2.0, 4.0, 6.0, 8.0, 10.0, 11.0, 12.0, 20.0, 36.0, 43.0, 65.0, 179.0, 401.0, 1166.0, 4237.0, 31428.0, 16544819.0, 183763.0, 8094.0, 1763.0, 605.0, 225.0, 102.0, 55.0, 42.0, 21.0, 18.0, 10.0, 14.0, 8.0, 4.0, 6.0, 5.0, 2.0, 1.0, 2.0, 3.0, 0.0, 0.0, 1.0, 1.0, 1.0, 3.0, 2.0], "bins": [-0.003631591796875, -0.0035326480865478516, -0.003433704376220703, -0.0033347606658935547, -0.0032358169555664062, -0.003136873245239258, -0.0030379295349121094, -0.002938985824584961, -0.0028400421142578125, -0.002741098403930664, -0.0026421546936035156, -0.002543210983276367, -0.0024442672729492188, -0.0023453235626220703, -0.002246379852294922, -0.0021474361419677734, -0.002048492431640625, -0.0019495487213134766, -0.0018506050109863281, -0.0017516613006591797, -0.0016527175903320312, -0.0015537738800048828, -0.0014548301696777344, -0.001355886459350586, -0.0012569427490234375, -0.001157999038696289, -0.0010590553283691406, -0.0009601116180419922, -0.0008611679077148438, -0.0007622241973876953, -0.0006632804870605469, -0.0005643367767333984, -0.00046539306640625, -0.00036644935607910156, -0.0002675056457519531, -0.0001685619354248047, -6.961822509765625e-05, 2.9325485229492188e-05, 0.00012826919555664062, 0.00022721290588378906, 0.0003261566162109375, 0.00042510032653808594, 0.0005240440368652344, 0.0006229877471923828, 0.0007219314575195312, 0.0008208751678466797, 0.0009198188781738281, 0.0010187625885009766, 0.001117706298828125, 0.0012166500091552734, 0.0013155937194824219, 0.0014145374298095703, 0.0015134811401367188, 0.0016124248504638672, 0.0017113685607910156, 0.001810312271118164, 0.0019092559814453125, 0.002008199691772461, 0.0021071434020996094, 0.002206087112426758, 0.0023050308227539062, 0.0024039745330810547, 0.002502918243408203, 0.0026018619537353516, 0.0027008056640625]}, "gradients/model.layers.23.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 5.0, 7.0, 8.0, 36.0, 978.0, 2795.0, 225.0, 21.0, 4.0, 3.0, 0.0, 0.0, 2.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.0028076171875, -0.0027322769165039062, -0.0026569366455078125, -0.0025815963745117188, -0.002506256103515625, -0.0024309158325195312, -0.0023555755615234375, -0.0022802352905273438, -0.00220489501953125, -0.0021295547485351562, -0.0020542144775390625, -0.0019788742065429688, -0.001903533935546875, -0.0018281936645507812, -0.0017528533935546875, -0.0016775131225585938, -0.0016021728515625, -0.0015268325805664062, -0.0014514923095703125, -0.0013761520385742188, -0.001300811767578125, -0.0012254714965820312, -0.0011501312255859375, -0.0010747909545898438, -0.00099945068359375, -0.0009241104125976562, -0.0008487701416015625, -0.0007734298706054688, -0.000698089599609375, -0.0006227493286132812, -0.0005474090576171875, -0.00047206878662109375, -0.000396728515625, -0.00032138824462890625, -0.0002460479736328125, -0.00017070770263671875, -9.5367431640625e-05, -2.002716064453125e-05, 5.53131103515625e-05, 0.00013065338134765625, 0.00020599365234375, 0.00028133392333984375, 0.0003566741943359375, 0.00043201446533203125, 0.000507354736328125, 0.0005826950073242188, 0.0006580352783203125, 0.0007333755493164062, 0.0008087158203125, 0.0008840560913085938, 0.0009593963623046875, 0.0010347366333007812, 0.001110076904296875, 0.0011854171752929688, 0.0012607574462890625, 0.0013360977172851562, 0.00141143798828125, 0.0014867782592773438, 0.0015621185302734375, 0.0016374588012695312, 0.001712799072265625, 0.0017881393432617188, 0.0018634796142578125, 0.0019388198852539062, 0.00201416015625]}, "gradients/model.layers.22.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 1.0, 1.0, 0.0, 2.0, 0.0, 2.0, 0.0, 7.0, 3.0, 2.0, 13.0, 19.0, 22.0, 34.0, 56.0, 87.0, 119.0, 149.0, 270.0, 440.0, 795.0, 1450.0, 3235.0, 7840.0, 24893.0, 96924.0, 638227.0, 13623589.0, 16777216.0, 3984198.0, 277498.0, 56149.0, 15815.0, 5319.0, 2276.0, 1206.0, 613.0, 338.0, 222.0, 121.0, 77.0, 64.0, 38.0, 27.0, 13.0, 19.0, 10.0, 9.0, 6.0, 3.0, 3.0, 2.0, 2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00154876708984375, -0.0014992952346801758, -0.0014498233795166016, -0.0014003515243530273, -0.0013508796691894531, -0.001301407814025879, -0.0012519359588623047, -0.0012024641036987305, -0.0011529922485351562, -0.001103520393371582, -0.0010540485382080078, -0.0010045766830444336, -0.0009551048278808594, -0.0009056329727172852, -0.0008561611175537109, -0.0008066892623901367, -0.0007572174072265625, -0.0007077455520629883, -0.0006582736968994141, -0.0006088018417358398, -0.0005593299865722656, -0.0005098581314086914, -0.0004603862762451172, -0.00041091442108154297, -0.00036144256591796875, -0.00031197071075439453, -0.0002624988555908203, -0.0002130270004272461, -0.00016355514526367188, -0.00011408329010009766, -6.461143493652344e-05, -1.5139579772949219e-05, 3.4332275390625e-05, 8.380413055419922e-05, 0.00013327598571777344, 0.00018274784088134766, 0.00023221969604492188, 0.0002816915512084961, 0.0003311634063720703, 0.00038063526153564453, 0.00043010711669921875, 0.00047957897186279297, 0.0005290508270263672, 0.0005785226821899414, 0.0006279945373535156, 0.0006774663925170898, 0.0007269382476806641, 0.0007764101028442383, 0.0008258819580078125, 0.0008753538131713867, 0.0009248256683349609, 0.0009742975234985352, 0.0010237693786621094, 0.0010732412338256836, 0.0011227130889892578, 0.001172184944152832, 0.0012216567993164062, 0.0012711286544799805, 0.0013206005096435547, 0.001370072364807129, 0.0014195442199707031, 0.0014690160751342773, 0.0015184879302978516, 0.0015679597854614258, 0.001617431640625]}, "gradients/model.layers.22.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 2.0, 0.0, 1.0, 1.0, 1.0, 3.0, 3.0, 5.0, 8.0, 8.0, 29.0, 34.0, 74.0, 192.0, 562.0, 2324.0, 14882.0, 375369.0, 16777216.0, 311597.0, 14074.0, 2081.0, 585.0, 193.0, 74.0, 41.0, 17.0, 21.0, 5.0, 6.0, 8.0, 4.0, 2.0, 2.0, 1.0, 0.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0042724609375, -0.004097938537597656, -0.0039234161376953125, -0.0037488937377929688, -0.003574371337890625, -0.0033998489379882812, -0.0032253265380859375, -0.0030508041381835938, -0.00287628173828125, -0.0027017593383789062, -0.0025272369384765625, -0.0023527145385742188, -0.002178192138671875, -0.0020036697387695312, -0.0018291473388671875, -0.0016546249389648438, -0.0014801025390625, -0.0013055801391601562, -0.0011310577392578125, -0.0009565353393554688, -0.000782012939453125, -0.0006074905395507812, -0.0004329681396484375, -0.00025844573974609375, -8.392333984375e-05, 9.059906005859375e-05, 0.0002651214599609375, 0.00043964385986328125, 0.000614166259765625, 0.0007886886596679688, 0.0009632110595703125, 0.0011377334594726562, 0.001312255859375, 0.0014867782592773438, 0.0016613006591796875, 0.0018358230590820312, 0.002010345458984375, 0.0021848678588867188, 0.0023593902587890625, 0.0025339126586914062, 0.00270843505859375, 0.0028829574584960938, 0.0030574798583984375, 0.0032320022583007812, 0.003406524658203125, 0.0035810470581054688, 0.0037555694580078125, 0.003930091857910156, 0.0041046142578125, 0.004279136657714844, 0.0044536590576171875, 0.004628181457519531, 0.004802703857421875, 0.004977226257324219, 0.0051517486572265625, 0.005326271057128906, 0.00550079345703125, 0.005675315856933594, 0.0058498382568359375, 0.006024360656738281, 0.006198883056640625, 0.006373405456542969, 0.0065479278564453125, 0.006722450256347656, 0.00689697265625]}, "gradients/model.layers.22.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 2.0, 0.0, 0.0, 2.0, 2.0, 1.0, 9.0, 10.0, 14.0, 31.0, 82.0, 143.0, 256.0, 414.0, 529.0, 624.0, 602.0, 531.0, 359.0, 210.0, 134.0, 61.0, 29.0, 17.0, 10.0, 3.0, 4.0, 5.0, 0.0, 0.0, 1.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0010528564453125, -0.0010186433792114258, -0.0009844303131103516, -0.0009502172470092773, -0.0009160041809082031, -0.0008817911148071289, -0.0008475780487060547, -0.0008133649826049805, -0.0007791519165039062, -0.000744938850402832, -0.0007107257843017578, -0.0006765127182006836, -0.0006422996520996094, -0.0006080865859985352, -0.0005738735198974609, -0.0005396604537963867, -0.0005054473876953125, -0.0004712343215942383, -0.00043702125549316406, -0.00040280818939208984, -0.0003685951232910156, -0.0003343820571899414, -0.0003001689910888672, -0.00026595592498779297, -0.00023174285888671875, -0.00019752979278564453, -0.0001633167266845703, -0.0001291036605834961, -9.489059448242188e-05, -6.0677528381347656e-05, -2.6464462280273438e-05, 7.748603820800781e-06, 4.1961669921875e-05, 7.617473602294922e-05, 0.00011038780212402344, 0.00014460086822509766, 0.00017881393432617188, 0.0002130270004272461, 0.0002472400665283203, 0.00028145313262939453, 0.00031566619873046875, 0.00034987926483154297, 0.0003840923309326172, 0.0004183053970336914, 0.0004525184631347656, 0.00048673152923583984, 0.0005209445953369141, 0.0005551576614379883, 0.0005893707275390625, 0.0006235837936401367, 0.0006577968597412109, 0.0006920099258422852, 0.0007262229919433594, 0.0007604360580444336, 0.0007946491241455078, 0.000828862190246582, 0.0008630752563476562, 0.0008972883224487305, 0.0009315013885498047, 0.0009657144546508789, 0.0009999275207519531, 0.0010341405868530273, 0.0010683536529541016, 0.0011025667190551758, 0.00113677978515625]}, "gradients/model.layers.22.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 2.0, 0.0, 2.0, 2.0, 3.0, 8.0, 9.0, 16.0, 47.0, 53.0, 118.0, 237.0, 465.0, 916.0, 2017.0, 4839.0, 18659.0, 244538.0, 6887067.0, 9244974.0, 341373.0, 22181.0, 5421.0, 2177.0, 1034.0, 489.0, 233.0, 144.0, 85.0, 36.0, 21.0, 13.0, 9.0, 7.0, 6.0, 2.0, 2.0, 2.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.001708984375, -0.001645803451538086, -0.0015826225280761719, -0.0015194416046142578, -0.0014562606811523438, -0.0013930797576904297, -0.0013298988342285156, -0.0012667179107666016, -0.0012035369873046875, -0.0011403560638427734, -0.0010771751403808594, -0.0010139942169189453, -0.0009508132934570312, -0.0008876323699951172, -0.0008244514465332031, -0.0007612705230712891, -0.000698089599609375, -0.0006349086761474609, -0.0005717277526855469, -0.0005085468292236328, -0.00044536590576171875, -0.0003821849822998047, -0.0003190040588378906, -0.00025582313537597656, -0.0001926422119140625, -0.00012946128845214844, -6.628036499023438e-05, -3.0994415283203125e-06, 6.008148193359375e-05, 0.0001232624053955078, 0.00018644332885742188, 0.00024962425231933594, 0.00031280517578125, 0.00037598609924316406, 0.0004391670227050781, 0.0005023479461669922, 0.0005655288696289062, 0.0006287097930908203, 0.0006918907165527344, 0.0007550716400146484, 0.0008182525634765625, 0.0008814334869384766, 0.0009446144104003906, 0.0010077953338623047, 0.0010709762573242188, 0.0011341571807861328, 0.0011973381042480469, 0.001260519027709961, 0.001323699951171875, 0.001386880874633789, 0.0014500617980957031, 0.0015132427215576172, 0.0015764236450195312, 0.0016396045684814453, 0.0017027854919433594, 0.0017659664154052734, 0.0018291473388671875, 0.0018923282623291016, 0.0019555091857910156, 0.0020186901092529297, 0.0020818710327148438, 0.002145051956176758, 0.002208232879638672, 0.002271413803100586, 0.0023345947265625]}, "gradients/model.layers.22.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 5.0, 0.0, 7.0, 3.0, 3.0, 9.0, 11.0, 17.0, 22.0, 41.0, 37.0, 62.0, 115.0, 162.0, 271.0, 461.0, 876.0, 2127.0, 7287.0, 123288.0, 3974405.0, 74725.0, 6366.0, 1978.0, 823.0, 456.0, 264.0, 147.0, 103.0, 67.0, 57.0, 29.0, 24.0, 13.0, 18.0, 6.0, 4.0, 4.0, 3.0, 1.0, 0.0, 2.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.006439208984375, -0.006252288818359375, -0.00606536865234375, -0.005878448486328125, -0.0056915283203125, -0.005504608154296875, -0.00531768798828125, -0.005130767822265625, -0.00494384765625, -0.004756927490234375, -0.00457000732421875, -0.004383087158203125, -0.0041961669921875, -0.004009246826171875, -0.00382232666015625, -0.003635406494140625, -0.003448486328125, -0.003261566162109375, -0.00307464599609375, -0.002887725830078125, -0.0027008056640625, -0.002513885498046875, -0.00232696533203125, -0.002140045166015625, -0.001953125, -0.001766204833984375, -0.00157928466796875, -0.001392364501953125, -0.0012054443359375, -0.001018524169921875, -0.00083160400390625, -0.000644683837890625, -0.000457763671875, -0.000270843505859375, -8.392333984375e-05, 0.000102996826171875, 0.0002899169921875, 0.000476837158203125, 0.00066375732421875, 0.000850677490234375, 0.00103759765625, 0.001224517822265625, 0.00141143798828125, 0.001598358154296875, 0.0017852783203125, 0.001972198486328125, 0.00215911865234375, 0.002346038818359375, 0.002532958984375, 0.002719879150390625, 0.00290679931640625, 0.003093719482421875, 0.0032806396484375, 0.003467559814453125, 0.00365447998046875, 0.003841400146484375, 0.0040283203125, 0.004215240478515625, 0.00440216064453125, 0.004589080810546875, 0.0047760009765625, 0.004962921142578125, 0.00514984130859375, 0.005336761474609375, 0.005523681640625]}, "gradients/model.layers.22.self_attn.k_proj.weight": {"_type": "histogram", "values": [3.0, 1.0, 0.0, 0.0, 2.0, 0.0, 1.0, 1.0, 2.0, 3.0, 1.0, 2.0, 2.0, 7.0, 6.0, 10.0, 14.0, 20.0, 21.0, 34.0, 46.0, 83.0, 124.0, 213.0, 337.0, 865.0, 2243.0, 8058.0, 45942.0, 3729735.0, 381962.0, 17446.0, 4298.0, 1512.0, 560.0, 277.0, 146.0, 81.0, 62.0, 38.0, 33.0, 24.0, 14.0, 20.0, 9.0, 7.0, 10.0, 8.0, 5.0, 2.0, 0.0, 5.0, 1.0, 2.0, 1.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00147247314453125, -0.0014230012893676758, -0.0013735294342041016, -0.0013240575790405273, -0.0012745857238769531, -0.001225113868713379, -0.0011756420135498047, -0.0011261701583862305, -0.0010766983032226562, -0.001027226448059082, -0.0009777545928955078, -0.0009282827377319336, -0.0008788108825683594, -0.0008293390274047852, -0.0007798671722412109, -0.0007303953170776367, -0.0006809234619140625, -0.0006314516067504883, -0.0005819797515869141, -0.0005325078964233398, -0.0004830360412597656, -0.0004335641860961914, -0.0003840923309326172, -0.00033462047576904297, -0.00028514862060546875, -0.00023567676544189453, -0.0001862049102783203, -0.0001367330551147461, -8.726119995117188e-05, -3.7789344787597656e-05, 1.1682510375976562e-05, 6.115436553955078e-05, 0.000110626220703125, 0.00016009807586669922, 0.00020956993103027344, 0.00025904178619384766, 0.0003085136413574219, 0.0003579854965209961, 0.0004074573516845703, 0.00045692920684814453, 0.0005064010620117188, 0.000555872917175293, 0.0006053447723388672, 0.0006548166275024414, 0.0007042884826660156, 0.0007537603378295898, 0.0008032321929931641, 0.0008527040481567383, 0.0009021759033203125, 0.0009516477584838867, 0.001001119613647461, 0.0010505914688110352, 0.0011000633239746094, 0.0011495351791381836, 0.0011990070343017578, 0.001248478889465332, 0.0012979507446289062, 0.0013474225997924805, 0.0013968944549560547, 0.001446366310119629, 0.0014958381652832031, 0.0015453100204467773, 0.0015947818756103516, 0.0016442537307739258, 0.0016937255859375]}, "gradients/model.layers.22.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 0.0, 1.0, 1.0, 3.0, 2.0, 3.0, 3.0, 2.0, 7.0, 10.0, 8.0, 12.0, 17.0, 13.0, 29.0, 32.0, 53.0, 81.0, 131.0, 212.0, 472.0, 1170.0, 3169.0, 14389.0, 220697.0, 16421038.0, 100949.0, 10208.0, 2602.0, 903.0, 431.0, 213.0, 99.0, 61.0, 57.0, 29.0, 21.0, 24.0, 12.0, 9.0, 4.0, 10.0, 6.0, 2.0, 6.0, 3.0, 2.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.002288818359375, -0.002208232879638672, -0.0021276473999023438, -0.0020470619201660156, -0.0019664764404296875, -0.0018858909606933594, -0.0018053054809570312, -0.0017247200012207031, -0.001644134521484375, -0.0015635490417480469, -0.0014829635620117188, -0.0014023780822753906, -0.0013217926025390625, -0.0012412071228027344, -0.0011606216430664062, -0.0010800361633300781, -0.00099945068359375, -0.0009188652038574219, -0.0008382797241210938, -0.0007576942443847656, -0.0006771087646484375, -0.0005965232849121094, -0.0005159378051757812, -0.0004353523254394531, -0.000354766845703125, -0.0002741813659667969, -0.00019359588623046875, -0.00011301040649414062, -3.24249267578125e-05, 4.8160552978515625e-05, 0.00012874603271484375, 0.00020933151245117188, 0.0002899169921875, 0.0003705024719238281, 0.00045108795166015625, 0.0005316734313964844, 0.0006122589111328125, 0.0006928443908691406, 0.0007734298706054688, 0.0008540153503417969, 0.000934600830078125, 0.0010151863098144531, 0.0010957717895507812, 0.0011763572692871094, 0.0012569427490234375, 0.0013375282287597656, 0.0014181137084960938, 0.0014986991882324219, 0.00157928466796875, 0.0016598701477050781, 0.0017404556274414062, 0.0018210411071777344, 0.0019016265869140625, 0.0019822120666503906, 0.0020627975463867188, 0.002143383026123047, 0.002223968505859375, 0.002304553985595703, 0.0023851394653320312, 0.0024657249450683594, 0.0025463104248046875, 0.0026268959045410156, 0.0027074813842773438, 0.002788066864013672, 0.00286865234375]}, "gradients/model.layers.22.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0, 1.0, 4.0, 5.0, 10.0, 13.0, 38.0, 106.0, 307.0, 682.0, 1018.0, 971.0, 528.0, 266.0, 66.0, 28.0, 11.0, 7.0, 6.0, 5.0, 1.0, 1.0, 1.0, 2.0, 0.0, 2.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0007781982421875, -0.0007544755935668945, -0.0007307529449462891, -0.0007070302963256836, -0.0006833076477050781, -0.0006595849990844727, -0.0006358623504638672, -0.0006121397018432617, -0.0005884170532226562, -0.0005646944046020508, -0.0005409717559814453, -0.0005172491073608398, -0.0004935264587402344, -0.0004698038101196289, -0.00044608116149902344, -0.00042235851287841797, -0.0003986358642578125, -0.00037491321563720703, -0.00035119056701660156, -0.0003274679183959961, -0.0003037452697753906, -0.00028002262115478516, -0.0002562999725341797, -0.00023257732391357422, -0.00020885467529296875, -0.00018513202667236328, -0.0001614093780517578, -0.00013768672943115234, -0.00011396408081054688, -9.02414321899414e-05, -6.651878356933594e-05, -4.279613494873047e-05, -1.9073486328125e-05, 4.649162292480469e-06, 2.8371810913085938e-05, 5.2094459533691406e-05, 7.581710815429688e-05, 9.953975677490234e-05, 0.0001232624053955078, 0.00014698505401611328, 0.00017070770263671875, 0.00019443035125732422, 0.0002181529998779297, 0.00024187564849853516, 0.0002655982971191406, 0.0002893209457397461, 0.00031304359436035156, 0.00033676624298095703, 0.0003604888916015625, 0.00038421154022216797, 0.00040793418884277344, 0.0004316568374633789, 0.0004553794860839844, 0.00047910213470458984, 0.0005028247833251953, 0.0005265474319458008, 0.0005502700805664062, 0.0005739927291870117, 0.0005977153778076172, 0.0006214380264282227, 0.0006451606750488281, 0.0006688833236694336, 0.0006926059722900391, 0.0007163286209106445, 0.00074005126953125]}, "gradients/model.layers.21.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 3.0, 3.0, 3.0, 5.0, 4.0, 8.0, 8.0, 16.0, 20.0, 36.0, 44.0, 78.0, 140.0, 280.0, 532.0, 1052.0, 1947.0, 4117.0, 9339.0, 23072.0, 65732.0, 270452.0, 2424426.0, 16777216.0, 16777216.0, 2323317.0, 263857.0, 66630.0, 22384.0, 9238.0, 4355.0, 2078.0, 939.0, 541.0, 291.0, 156.0, 83.0, 40.0, 24.0, 34.0, 12.0, 4.0, 8.0, 5.0, 3.0, 4.0, 2.0, 5.0, 1.0, 0.0, 2.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.00140380859375, -0.0013585090637207031, -0.0013132095336914062, -0.0012679100036621094, -0.0012226104736328125, -0.0011773109436035156, -0.0011320114135742188, -0.0010867118835449219, -0.001041412353515625, -0.0009961128234863281, -0.0009508132934570312, -0.0009055137634277344, -0.0008602142333984375, -0.0008149147033691406, -0.0007696151733398438, -0.0007243156433105469, -0.00067901611328125, -0.0006337165832519531, -0.0005884170532226562, -0.0005431175231933594, -0.0004978179931640625, -0.0004525184631347656, -0.00040721893310546875, -0.0003619194030761719, -0.000316619873046875, -0.0002713203430175781, -0.00022602081298828125, -0.00018072128295898438, -0.0001354217529296875, -9.012222290039062e-05, -4.482269287109375e-05, 4.76837158203125e-07, 4.57763671875e-05, 9.107589721679688e-05, 0.00013637542724609375, 0.00018167495727539062, 0.0002269744873046875, 0.0002722740173339844, 0.00031757354736328125, 0.0003628730773925781, 0.000408172607421875, 0.0004534721374511719, 0.0004987716674804688, 0.0005440711975097656, 0.0005893707275390625, 0.0006346702575683594, 0.0006799697875976562, 0.0007252693176269531, 0.00077056884765625, 0.0008158683776855469, 0.0008611679077148438, 0.0009064674377441406, 0.0009517669677734375, 0.0009970664978027344, 0.0010423660278320312, 0.0010876655578613281, 0.001132965087890625, 0.0011782646179199219, 0.0012235641479492188, 0.0012688636779785156, 0.0013141632080078125, 0.0013594627380371094, 0.0014047622680664062, 0.0014500617980957031, 0.001495361328125]}, "gradients/model.layers.21.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 2.0, 2.0, 1.0, 2.0, 2.0, 5.0, 4.0, 7.0, 6.0, 11.0, 13.0, 22.0, 39.0, 56.0, 129.0, 316.0, 869.0, 3369.0, 16862.0, 270672.0, 16777216.0, 2187587.0, 38959.0, 5792.0, 1408.0, 434.0, 175.0, 71.0, 41.0, 25.0, 14.0, 10.0, 12.0, 3.0, 5.0, 3.0, 4.0, 2.0, 3.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.005615234375, -0.005457878112792969, -0.0053005218505859375, -0.005143165588378906, -0.004985809326171875, -0.004828453063964844, -0.0046710968017578125, -0.004513740539550781, -0.00435638427734375, -0.004199028015136719, -0.0040416717529296875, -0.0038843154907226562, -0.003726959228515625, -0.0035696029663085938, -0.0034122467041015625, -0.0032548904418945312, -0.0030975341796875, -0.0029401779174804688, -0.0027828216552734375, -0.0026254653930664062, -0.002468109130859375, -0.0023107528686523438, -0.0021533966064453125, -0.0019960403442382812, -0.00183868408203125, -0.0016813278198242188, -0.0015239715576171875, -0.0013666152954101562, -0.001209259033203125, -0.0010519027709960938, -0.0008945465087890625, -0.0007371902465820312, -0.000579833984375, -0.00042247772216796875, -0.0002651214599609375, -0.00010776519775390625, 4.9591064453125e-05, 0.00020694732666015625, 0.0003643035888671875, 0.0005216598510742188, 0.00067901611328125, 0.0008363723754882812, 0.0009937286376953125, 0.0011510848999023438, 0.001308441162109375, 0.0014657974243164062, 0.0016231536865234375, 0.0017805099487304688, 0.0019378662109375, 0.0020952224731445312, 0.0022525787353515625, 0.0024099349975585938, 0.002567291259765625, 0.0027246475219726562, 0.0028820037841796875, 0.0030393600463867188, 0.00319671630859375, 0.0033540725708007812, 0.0035114288330078125, 0.0036687850952148438, 0.003826141357421875, 0.003983497619628906, 0.0041408538818359375, 0.004298210144042969, 0.00445556640625]}, "gradients/model.layers.21.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 1.0, 2.0, 1.0, 4.0, 2.0, 2.0, 2.0, 1.0, 2.0, 13.0, 14.0, 23.0, 45.0, 46.0, 107.0, 160.0, 227.0, 263.0, 316.0, 405.0, 402.0, 406.0, 357.0, 338.0, 277.0, 211.0, 170.0, 100.0, 75.0, 40.0, 25.0, 18.0, 7.0, 8.0, 4.0, 4.0, 2.0, 2.0, 1.0, 1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.00072479248046875, -0.0006990432739257812, -0.0006732940673828125, -0.0006475448608398438, -0.000621795654296875, -0.0005960464477539062, -0.0005702972412109375, -0.0005445480346679688, -0.000518798828125, -0.0004930496215820312, -0.0004673004150390625, -0.00044155120849609375, -0.000415802001953125, -0.00039005279541015625, -0.0003643035888671875, -0.00033855438232421875, -0.00031280517578125, -0.00028705596923828125, -0.0002613067626953125, -0.00023555755615234375, -0.000209808349609375, -0.00018405914306640625, -0.0001583099365234375, -0.00013256072998046875, -0.0001068115234375, -8.106231689453125e-05, -5.53131103515625e-05, -2.956390380859375e-05, -3.814697265625e-06, 2.193450927734375e-05, 4.76837158203125e-05, 7.343292236328125e-05, 9.918212890625e-05, 0.00012493133544921875, 0.0001506805419921875, 0.00017642974853515625, 0.000202178955078125, 0.00022792816162109375, 0.0002536773681640625, 0.00027942657470703125, 0.00030517578125, 0.00033092498779296875, 0.0003566741943359375, 0.00038242340087890625, 0.000408172607421875, 0.00043392181396484375, 0.0004596710205078125, 0.00048542022705078125, 0.00051116943359375, 0.0005369186401367188, 0.0005626678466796875, 0.0005884170532226562, 0.000614166259765625, 0.0006399154663085938, 0.0006656646728515625, 0.0006914138793945312, 0.0007171630859375, 0.0007429122924804688, 0.0007686614990234375, 0.0007944107055664062, 0.000820159912109375, 0.0008459091186523438, 0.0008716583251953125, 0.0008974075317382812, 0.00092315673828125]}, "gradients/model.layers.21.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 3.0, 7.0, 5.0, 19.0, 15.0, 30.0, 25.0, 33.0, 55.0, 82.0, 126.0, 197.0, 336.0, 534.0, 1091.0, 2027.0, 4692.0, 15787.0, 93223.0, 1449704.0, 13473020.0, 1610327.0, 100113.0, 15964.0, 4975.0, 2151.0, 1059.0, 571.0, 356.0, 218.0, 146.0, 96.0, 65.0, 45.0, 32.0, 31.0, 12.0, 9.0, 10.0, 9.0, 5.0, 1.0, 0.0, 2.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.002777099609375, -0.0027010440826416016, -0.002624988555908203, -0.0025489330291748047, -0.0024728775024414062, -0.002396821975708008, -0.0023207664489746094, -0.002244710922241211, -0.0021686553955078125, -0.002092599868774414, -0.0020165443420410156, -0.0019404888153076172, -0.0018644332885742188, -0.0017883777618408203, -0.0017123222351074219, -0.0016362667083740234, -0.001560211181640625, -0.0014841556549072266, -0.0014081001281738281, -0.0013320446014404297, -0.0012559890747070312, -0.0011799335479736328, -0.0011038780212402344, -0.001027822494506836, -0.0009517669677734375, -0.0008757114410400391, -0.0007996559143066406, -0.0007236003875732422, -0.0006475448608398438, -0.0005714893341064453, -0.0004954338073730469, -0.00041937828063964844, -0.00034332275390625, -0.00026726722717285156, -0.00019121170043945312, -0.00011515617370605469, -3.910064697265625e-05, 3.695487976074219e-05, 0.00011301040649414062, 0.00018906593322753906, 0.0002651214599609375, 0.00034117698669433594, 0.0004172325134277344, 0.0004932880401611328, 0.0005693435668945312, 0.0006453990936279297, 0.0007214546203613281, 0.0007975101470947266, 0.000873565673828125, 0.0009496212005615234, 0.0010256767272949219, 0.0011017322540283203, 0.0011777877807617188, 0.0012538433074951172, 0.0013298988342285156, 0.001405954360961914, 0.0014820098876953125, 0.001558065414428711, 0.0016341209411621094, 0.0017101764678955078, 0.0017862319946289062, 0.0018622875213623047, 0.0019383430480957031, 0.0020143985748291016, 0.0020904541015625]}, "gradients/model.layers.21.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 1.0, 0.0, 4.0, 3.0, 4.0, 5.0, 6.0, 9.0, 11.0, 13.0, 18.0, 27.0, 30.0, 40.0, 66.0, 81.0, 118.0, 141.0, 249.0, 326.0, 662.0, 1131.0, 2187.0, 4469.0, 12970.0, 184377.0, 3326915.0, 625398.0, 22864.0, 6080.0, 2650.0, 1344.0, 788.0, 424.0, 269.0, 179.0, 115.0, 89.0, 60.0, 39.0, 30.0, 22.0, 19.0, 16.0, 13.0, 6.0, 10.0, 6.0, 4.0, 1.0, 3.0, 1.0, 0.0, 2.0, 1.0, 0.0, 2.0, 1.0], "bins": [-0.004547119140625, -0.0044078826904296875, -0.004268646240234375, -0.0041294097900390625, -0.00399017333984375, -0.0038509368896484375, -0.003711700439453125, -0.0035724639892578125, -0.0034332275390625, -0.0032939910888671875, -0.003154754638671875, -0.0030155181884765625, -0.00287628173828125, -0.0027370452880859375, -0.002597808837890625, -0.0024585723876953125, -0.0023193359375, -0.0021800994873046875, -0.002040863037109375, -0.0019016265869140625, -0.00176239013671875, -0.0016231536865234375, -0.001483917236328125, -0.0013446807861328125, -0.0012054443359375, -0.0010662078857421875, -0.000926971435546875, -0.0007877349853515625, -0.00064849853515625, -0.0005092620849609375, -0.000370025634765625, -0.0002307891845703125, -9.1552734375e-05, 4.76837158203125e-05, 0.000186920166015625, 0.0003261566162109375, 0.00046539306640625, 0.0006046295166015625, 0.000743865966796875, 0.0008831024169921875, 0.0010223388671875, 0.0011615753173828125, 0.001300811767578125, 0.0014400482177734375, 0.00157928466796875, 0.0017185211181640625, 0.001857757568359375, 0.0019969940185546875, 0.00213623046875, 0.0022754669189453125, 0.002414703369140625, 0.0025539398193359375, 0.00269317626953125, 0.0028324127197265625, 0.002971649169921875, 0.0031108856201171875, 0.0032501220703125, 0.0033893585205078125, 0.003528594970703125, 0.0036678314208984375, 0.00380706787109375, 0.0039463043212890625, 0.004085540771484375, 0.0042247772216796875, 0.004364013671875]}, "gradients/model.layers.21.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 2.0, 4.0, 0.0, 0.0, 2.0, 4.0, 9.0, 3.0, 9.0, 18.0, 30.0, 54.0, 146.0, 323.0, 911.0, 4236.0, 58542.0, 4082683.0, 42330.0, 3534.0, 797.0, 334.0, 149.0, 73.0, 30.0, 21.0, 11.0, 16.0, 3.0, 5.0, 4.0, 3.0, 3.0, 1.0, 2.0, 3.0, 1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0023040771484375, -0.00222015380859375, -0.00213623046875, -0.00205230712890625, -0.0019683837890625, -0.00188446044921875, -0.001800537109375, -0.00171661376953125, -0.0016326904296875, -0.00154876708984375, -0.00146484375, -0.00138092041015625, -0.0012969970703125, -0.00121307373046875, -0.001129150390625, -0.00104522705078125, -0.0009613037109375, -0.00087738037109375, -0.00079345703125, -0.00070953369140625, -0.0006256103515625, -0.00054168701171875, -0.000457763671875, -0.00037384033203125, -0.0002899169921875, -0.00020599365234375, -0.0001220703125, -3.814697265625e-05, 4.57763671875e-05, 0.00012969970703125, 0.000213623046875, 0.00029754638671875, 0.0003814697265625, 0.00046539306640625, 0.00054931640625, 0.00063323974609375, 0.0007171630859375, 0.00080108642578125, 0.000885009765625, 0.00096893310546875, 0.0010528564453125, 0.00113677978515625, 0.001220703125, 0.00130462646484375, 0.0013885498046875, 0.00147247314453125, 0.001556396484375, 0.00164031982421875, 0.0017242431640625, 0.00180816650390625, 0.00189208984375, 0.00197601318359375, 0.0020599365234375, 0.00214385986328125, 0.002227783203125, 0.00231170654296875, 0.0023956298828125, 0.00247955322265625, 0.0025634765625, 0.00264739990234375, 0.0027313232421875, 0.00281524658203125, 0.002899169921875, 0.00298309326171875, 0.0030670166015625]}, "gradients/model.layers.21.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 1.0, 2.0, 1.0, 1.0, 4.0, 2.0, 1.0, 1.0, 3.0, 6.0, 6.0, 3.0, 11.0, 10.0, 18.0, 19.0, 27.0, 34.0, 69.0, 72.0, 147.0, 247.0, 637.0, 1681.0, 6045.0, 58064.0, 16560402.0, 137269.0, 8676.0, 2098.0, 816.0, 301.0, 165.0, 104.0, 80.0, 48.0, 23.0, 36.0, 23.0, 12.0, 6.0, 6.0, 4.0, 4.0, 4.0, 6.0, 4.0, 2.0, 0.0, 2.0, 0.0, 2.0, 2.0, 2.0], "bins": [-0.004486083984375, -0.004363536834716797, -0.004240989685058594, -0.004118442535400391, -0.0039958953857421875, -0.0038733482360839844, -0.0037508010864257812, -0.003628253936767578, -0.003505706787109375, -0.003383159637451172, -0.0032606124877929688, -0.0031380653381347656, -0.0030155181884765625, -0.0028929710388183594, -0.0027704238891601562, -0.002647876739501953, -0.00252532958984375, -0.002402782440185547, -0.0022802352905273438, -0.0021576881408691406, -0.0020351409912109375, -0.0019125938415527344, -0.0017900466918945312, -0.0016674995422363281, -0.001544952392578125, -0.0014224052429199219, -0.0012998580932617188, -0.0011773109436035156, -0.0010547637939453125, -0.0009322166442871094, -0.0008096694946289062, -0.0006871223449707031, -0.0005645751953125, -0.0004420280456542969, -0.00031948089599609375, -0.00019693374633789062, -7.43865966796875e-05, 4.8160552978515625e-05, 0.00017070770263671875, 0.0002932548522949219, 0.000415802001953125, 0.0005383491516113281, 0.0006608963012695312, 0.0007834434509277344, 0.0009059906005859375, 0.0010285377502441406, 0.0011510848999023438, 0.0012736320495605469, 0.00139617919921875, 0.0015187263488769531, 0.0016412734985351562, 0.0017638206481933594, 0.0018863677978515625, 0.0020089149475097656, 0.0021314620971679688, 0.002254009246826172, 0.002376556396484375, 0.002499103546142578, 0.0026216506958007812, 0.0027441978454589844, 0.0028667449951171875, 0.0029892921447753906, 0.0031118392944335938, 0.003234386444091797, 0.00335693359375]}, "gradients/model.layers.21.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 3.0, 3.0, 4.0, 4.0, 6.0, 11.0, 19.0, 80.0, 578.0, 1498.0, 1367.0, 437.0, 49.0, 12.0, 5.0, 3.0, 2.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.002197265625, -0.0021504759788513184, -0.0021036863327026367, -0.002056896686553955, -0.0020101070404052734, -0.001963317394256592, -0.0019165277481079102, -0.0018697381019592285, -0.0018229484558105469, -0.0017761588096618652, -0.0017293691635131836, -0.001682579517364502, -0.0016357898712158203, -0.0015890002250671387, -0.001542210578918457, -0.0014954209327697754, -0.0014486312866210938, -0.0014018416404724121, -0.0013550519943237305, -0.0013082623481750488, -0.0012614727020263672, -0.0012146830558776855, -0.001167893409729004, -0.0011211037635803223, -0.0010743141174316406, -0.001027524471282959, -0.0009807348251342773, -0.0009339451789855957, -0.0008871555328369141, -0.0008403658866882324, -0.0007935762405395508, -0.0007467865943908691, -0.0006999969482421875, -0.0006532073020935059, -0.0006064176559448242, -0.0005596280097961426, -0.0005128383636474609, -0.0004660487174987793, -0.00041925907135009766, -0.000372469425201416, -0.0003256797790527344, -0.00027889013290405273, -0.0002321004867553711, -0.00018531084060668945, -0.0001385211944580078, -9.173154830932617e-05, -4.494190216064453e-05, 1.8477439880371094e-06, 4.863739013671875e-05, 9.542703628540039e-05, 0.00014221668243408203, 0.00018900632858276367, 0.0002357959747314453, 0.00028258562088012695, 0.0003293752670288086, 0.00037616491317749023, 0.0004229545593261719, 0.0004697442054748535, 0.0005165338516235352, 0.0005633234977722168, 0.0006101131439208984, 0.0006569027900695801, 0.0007036924362182617, 0.0007504820823669434, 0.000797271728515625]}, "gradients/model.layers.20.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 3.0, 2.0, 8.0, 4.0, 11.0, 17.0, 30.0, 29.0, 47.0, 73.0, 74.0, 123.0, 230.0, 345.0, 717.0, 1794.0, 5287.0, 22187.0, 155981.0, 3943767.0, 16777216.0, 6271063.0, 202629.0, 27054.0, 5812.0, 1939.0, 805.0, 401.0, 254.0, 144.0, 102.0, 80.0, 53.0, 33.0, 37.0, 16.0, 12.0, 15.0, 5.0, 2.0, 3.0, 2.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 2.0, 0.0, 1.0], "bins": [-0.002471923828125, -0.002398252487182617, -0.0023245811462402344, -0.0022509098052978516, -0.0021772384643554688, -0.002103567123413086, -0.002029895782470703, -0.0019562244415283203, -0.0018825531005859375, -0.0018088817596435547, -0.0017352104187011719, -0.001661539077758789, -0.0015878677368164062, -0.0015141963958740234, -0.0014405250549316406, -0.0013668537139892578, -0.001293182373046875, -0.0012195110321044922, -0.0011458396911621094, -0.0010721683502197266, -0.0009984970092773438, -0.0009248256683349609, -0.0008511543273925781, -0.0007774829864501953, -0.0007038116455078125, -0.0006301403045654297, -0.0005564689636230469, -0.00048279762268066406, -0.00040912628173828125, -0.00033545494079589844, -0.0002617835998535156, -0.0001881122589111328, -0.00011444091796875, -4.076957702636719e-05, 3.2901763916015625e-05, 0.00010657310485839844, 0.00018024444580078125, 0.00025391578674316406, 0.0003275871276855469, 0.0004012584686279297, 0.0004749298095703125, 0.0005486011505126953, 0.0006222724914550781, 0.0006959438323974609, 0.0007696151733398438, 0.0008432865142822266, 0.0009169578552246094, 0.0009906291961669922, 0.001064300537109375, 0.0011379718780517578, 0.0012116432189941406, 0.0012853145599365234, 0.0013589859008789062, 0.001432657241821289, 0.0015063285827636719, 0.0015799999237060547, 0.0016536712646484375, 0.0017273426055908203, 0.0018010139465332031, 0.001874685287475586, 0.0019483566284179688, 0.0020220279693603516, 0.0020956993103027344, 0.002169370651245117, 0.0022430419921875]}, "gradients/model.layers.20.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 1.0, 1.0, 3.0, 3.0, 5.0, 8.0, 7.0, 14.0, 26.0, 52.0, 79.0, 164.0, 396.0, 1392.0, 6941.0, 69927.0, 16777216.0, 16777216.0, 130552.0, 10112.0, 1819.0, 509.0, 179.0, 95.0, 47.0, 24.0, 17.0, 9.0, 9.0, 8.0, 6.0, 3.0, 4.0, 0.0, 2.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.0050048828125, -0.004848957061767578, -0.004693031311035156, -0.004537105560302734, -0.0043811798095703125, -0.004225254058837891, -0.004069328308105469, -0.003913402557373047, -0.003757476806640625, -0.003601551055908203, -0.0034456253051757812, -0.0032896995544433594, -0.0031337738037109375, -0.0029778480529785156, -0.0028219223022460938, -0.002665996551513672, -0.00251007080078125, -0.002354145050048828, -0.0021982192993164062, -0.0020422935485839844, -0.0018863677978515625, -0.0017304420471191406, -0.0015745162963867188, -0.0014185905456542969, -0.001262664794921875, -0.0011067390441894531, -0.0009508132934570312, -0.0007948875427246094, -0.0006389617919921875, -0.0004830360412597656, -0.00032711029052734375, -0.00017118453979492188, -1.52587890625e-05, 0.00014066696166992188, 0.00029659271240234375, 0.0004525184631347656, 0.0006084442138671875, 0.0007643699645996094, 0.0009202957153320312, 0.0010762214660644531, 0.001232147216796875, 0.0013880729675292969, 0.0015439987182617188, 0.0016999244689941406, 0.0018558502197265625, 0.0020117759704589844, 0.0021677017211914062, 0.002323627471923828, 0.00247955322265625, 0.002635478973388672, 0.0027914047241210938, 0.0029473304748535156, 0.0031032562255859375, 0.0032591819763183594, 0.0034151077270507812, 0.003571033477783203, 0.003726959228515625, 0.003882884979248047, 0.004038810729980469, 0.004194736480712891, 0.0043506622314453125, 0.004506587982177734, 0.004662513732910156, 0.004818439483642578, 0.004974365234375]}, "gradients/model.layers.20.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 2.0, 2.0, 8.0, 16.0, 34.0, 73.0, 181.0, 369.0, 600.0, 738.0, 738.0, 595.0, 375.0, 202.0, 88.0, 37.0, 12.0, 5.0, 3.0, 5.0, 0.0, 1.0, 2.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00170135498046875, -0.0016481876373291016, -0.0015950202941894531, -0.0015418529510498047, -0.0014886856079101562, -0.0014355182647705078, -0.0013823509216308594, -0.001329183578491211, -0.0012760162353515625, -0.001222848892211914, -0.0011696815490722656, -0.0011165142059326172, -0.0010633468627929688, -0.0010101795196533203, -0.0009570121765136719, -0.0009038448333740234, -0.000850677490234375, -0.0007975101470947266, -0.0007443428039550781, -0.0006911754608154297, -0.0006380081176757812, -0.0005848407745361328, -0.0005316734313964844, -0.00047850608825683594, -0.0004253387451171875, -0.00037217140197753906, -0.0003190040588378906, -0.0002658367156982422, -0.00021266937255859375, -0.0001595020294189453, -0.00010633468627929688, -5.316734313964844e-05, 0.0, 5.316734313964844e-05, 0.00010633468627929688, 0.0001595020294189453, 0.00021266937255859375, 0.0002658367156982422, 0.0003190040588378906, 0.00037217140197753906, 0.0004253387451171875, 0.00047850608825683594, 0.0005316734313964844, 0.0005848407745361328, 0.0006380081176757812, 0.0006911754608154297, 0.0007443428039550781, 0.0007975101470947266, 0.000850677490234375, 0.0009038448333740234, 0.0009570121765136719, 0.0010101795196533203, 0.0010633468627929688, 0.0011165142059326172, 0.0011696815490722656, 0.001222848892211914, 0.0012760162353515625, 0.001329183578491211, 0.0013823509216308594, 0.0014355182647705078, 0.0014886856079101562, 0.0015418529510498047, 0.0015950202941894531, 0.0016481876373291016, 0.00170135498046875]}, "gradients/model.layers.20.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 3.0, 6.0, 3.0, 4.0, 14.0, 21.0, 50.0, 48.0, 108.0, 137.0, 251.0, 402.0, 774.0, 1422.0, 2644.0, 5600.0, 16353.0, 70080.0, 372475.0, 2167917.0, 9963882.0, 3460809.0, 572345.0, 104231.0, 23290.0, 7360.0, 3221.0, 1527.0, 943.0, 539.0, 300.0, 166.0, 116.0, 70.0, 39.0, 18.0, 9.0, 10.0, 4.0, 8.0, 6.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.00125885009765625, -0.0012224912643432617, -0.0011861324310302734, -0.0011497735977172852, -0.0011134147644042969, -0.0010770559310913086, -0.0010406970977783203, -0.001004338264465332, -0.0009679794311523438, -0.0009316205978393555, -0.0008952617645263672, -0.0008589029312133789, -0.0008225440979003906, -0.0007861852645874023, -0.0007498264312744141, -0.0007134675979614258, -0.0006771087646484375, -0.0006407499313354492, -0.0006043910980224609, -0.0005680322647094727, -0.0005316734313964844, -0.0004953145980834961, -0.0004589557647705078, -0.00042259693145751953, -0.00038623809814453125, -0.00034987926483154297, -0.0003135204315185547, -0.0002771615982055664, -0.00024080276489257812, -0.00020444393157958984, -0.00016808509826660156, -0.00013172626495361328, -9.5367431640625e-05, -5.900859832763672e-05, -2.2649765014648438e-05, 1.3709068298339844e-05, 5.0067901611328125e-05, 8.64267349243164e-05, 0.0001227855682373047, 0.00015914440155029297, 0.00019550323486328125, 0.00023186206817626953, 0.0002682209014892578, 0.0003045797348022461, 0.0003409385681152344, 0.00037729740142822266, 0.00041365623474121094, 0.0004500150680541992, 0.0004863739013671875, 0.0005227327346801758, 0.0005590915679931641, 0.0005954504013061523, 0.0006318092346191406, 0.0006681680679321289, 0.0007045269012451172, 0.0007408857345581055, 0.0007772445678710938, 0.000813603401184082, 0.0008499622344970703, 0.0008863210678100586, 0.0009226799011230469, 0.0009590387344360352, 0.0009953975677490234, 0.0010317564010620117, 0.001068115234375]}, "gradients/model.layers.20.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 1.0, 2.0, 4.0, 4.0, 6.0, 6.0, 7.0, 3.0, 9.0, 20.0, 21.0, 26.0, 38.0, 41.0, 48.0, 79.0, 117.0, 147.0, 213.0, 367.0, 642.0, 1325.0, 2875.0, 7535.0, 45374.0, 2193449.0, 1888183.0, 40542.0, 7271.0, 2761.0, 1297.0, 690.0, 413.0, 236.0, 135.0, 104.0, 74.0, 55.0, 49.0, 26.0, 31.0, 17.0, 18.0, 9.0, 11.0, 5.0, 3.0, 2.0, 1.0, 1.0, 1.0, 1.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.004974365234375, -0.0048084259033203125, -0.004642486572265625, -0.0044765472412109375, -0.00431060791015625, -0.0041446685791015625, -0.003978729248046875, -0.0038127899169921875, -0.0036468505859375, -0.0034809112548828125, -0.003314971923828125, -0.0031490325927734375, -0.00298309326171875, -0.0028171539306640625, -0.002651214599609375, -0.0024852752685546875, -0.0023193359375, -0.0021533966064453125, -0.001987457275390625, -0.0018215179443359375, -0.00165557861328125, -0.0014896392822265625, -0.001323699951171875, -0.0011577606201171875, -0.0009918212890625, -0.0008258819580078125, -0.000659942626953125, -0.0004940032958984375, -0.00032806396484375, -0.0001621246337890625, 3.814697265625e-06, 0.0001697540283203125, 0.000335693359375, 0.0005016326904296875, 0.000667572021484375, 0.0008335113525390625, 0.00099945068359375, 0.0011653900146484375, 0.001331329345703125, 0.0014972686767578125, 0.0016632080078125, 0.0018291473388671875, 0.001995086669921875, 0.0021610260009765625, 0.00232696533203125, 0.0024929046630859375, 0.002658843994140625, 0.0028247833251953125, 0.00299072265625, 0.0031566619873046875, 0.003322601318359375, 0.0034885406494140625, 0.00365447998046875, 0.0038204193115234375, 0.003986358642578125, 0.0041522979736328125, 0.0043182373046875, 0.0044841766357421875, 0.004650115966796875, 0.0048160552978515625, 0.00498199462890625, 0.0051479339599609375, 0.005313873291015625, 0.0054798126220703125, 0.005645751953125]}, "gradients/model.layers.20.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 2.0, 0.0, 2.0, 2.0, 1.0, 0.0, 2.0, 0.0, 3.0, 5.0, 2.0, 5.0, 15.0, 12.0, 7.0, 23.0, 24.0, 67.0, 104.0, 210.0, 392.0, 1020.0, 2799.0, 12971.0, 308182.0, 3819109.0, 40603.0, 5800.0, 1632.0, 674.0, 276.0, 136.0, 70.0, 43.0, 25.0, 25.0, 13.0, 6.0, 10.0, 7.0, 3.0, 5.0, 2.0, 1.0, 0.0, 2.0, 3.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.001953125, -0.0018925666809082031, -0.0018320083618164062, -0.0017714500427246094, -0.0017108917236328125, -0.0016503334045410156, -0.0015897750854492188, -0.0015292167663574219, -0.001468658447265625, -0.0014081001281738281, -0.0013475418090820312, -0.0012869834899902344, -0.0012264251708984375, -0.0011658668518066406, -0.0011053085327148438, -0.0010447502136230469, -0.00098419189453125, -0.0009236335754394531, -0.0008630752563476562, -0.0008025169372558594, -0.0007419586181640625, -0.0006814002990722656, -0.0006208419799804688, -0.0005602836608886719, -0.000499725341796875, -0.0004391670227050781, -0.00037860870361328125, -0.0003180503845214844, -0.0002574920654296875, -0.00019693374633789062, -0.00013637542724609375, -7.581710815429688e-05, -1.52587890625e-05, 4.5299530029296875e-05, 0.00010585784912109375, 0.00016641616821289062, 0.0002269744873046875, 0.0002875328063964844, 0.00034809112548828125, 0.0004086494445800781, 0.000469207763671875, 0.0005297660827636719, 0.0005903244018554688, 0.0006508827209472656, 0.0007114410400390625, 0.0007719993591308594, 0.0008325576782226562, 0.0008931159973144531, 0.00095367431640625, 0.0010142326354980469, 0.0010747909545898438, 0.0011353492736816406, 0.0011959075927734375, 0.0012564659118652344, 0.0013170242309570312, 0.0013775825500488281, 0.001438140869140625, 0.0014986991882324219, 0.0015592575073242188, 0.0016198158264160156, 0.0016803741455078125, 0.0017409324645996094, 0.0018014907836914062, 0.0018620491027832031, 0.001922607421875]}, "gradients/model.layers.20.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 0.0, 1.0, 0.0, 2.0, 4.0, 1.0, 2.0, 2.0, 2.0, 5.0, 5.0, 2.0, 7.0, 10.0, 19.0, 11.0, 19.0, 25.0, 45.0, 51.0, 82.0, 123.0, 206.0, 361.0, 781.0, 1720.0, 4268.0, 14090.0, 79665.0, 7109421.0, 9457597.0, 85868.0, 14630.0, 4519.0, 1815.0, 759.0, 424.0, 226.0, 136.0, 88.0, 56.0, 38.0, 23.0, 22.0, 21.0, 12.0, 13.0, 9.0, 7.0, 1.0, 3.0, 2.0, 1.0, 3.0, 0.0, 1.0, 3.0, 0.0, 2.0, 0.0, 4.0], "bins": [-0.001861572265625, -0.0018051862716674805, -0.001748800277709961, -0.0016924142837524414, -0.0016360282897949219, -0.0015796422958374023, -0.0015232563018798828, -0.0014668703079223633, -0.0014104843139648438, -0.0013540983200073242, -0.0012977123260498047, -0.0012413263320922852, -0.0011849403381347656, -0.001128554344177246, -0.0010721683502197266, -0.001015782356262207, -0.0009593963623046875, -0.000903010368347168, -0.0008466243743896484, -0.0007902383804321289, -0.0007338523864746094, -0.0006774663925170898, -0.0006210803985595703, -0.0005646944046020508, -0.0005083084106445312, -0.0004519224166870117, -0.0003955364227294922, -0.00033915042877197266, -0.0002827644348144531, -0.0002263784408569336, -0.00016999244689941406, -0.00011360645294189453, -5.7220458984375e-05, -8.344650268554688e-07, 5.555152893066406e-05, 0.0001119375228881836, 0.00016832351684570312, 0.00022470951080322266, 0.0002810955047607422, 0.0003374814987182617, 0.00039386749267578125, 0.0004502534866333008, 0.0005066394805908203, 0.0005630254745483398, 0.0006194114685058594, 0.0006757974624633789, 0.0007321834564208984, 0.000788569450378418, 0.0008449554443359375, 0.000901341438293457, 0.0009577274322509766, 0.001014113426208496, 0.0010704994201660156, 0.0011268854141235352, 0.0011832714080810547, 0.0012396574020385742, 0.0012960433959960938, 0.0013524293899536133, 0.0014088153839111328, 0.0014652013778686523, 0.0015215873718261719, 0.0015779733657836914, 0.001634359359741211, 0.0016907453536987305, 0.00174713134765625]}, "gradients/model.layers.20.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 1.0, 3.0, 0.0, 10.0, 10.0, 22.0, 109.0, 649.0, 1615.0, 1232.0, 367.0, 41.0, 17.0, 5.0, 4.0, 0.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00057220458984375, -0.000523686408996582, -0.00047516822814941406, -0.0004266500473022461, -0.0003781318664550781, -0.00032961368560791016, -0.0002810955047607422, -0.00023257732391357422, -0.00018405914306640625, -0.00013554096221923828, -8.702278137207031e-05, -3.8504600524902344e-05, 1.0013580322265625e-05, 5.8531761169433594e-05, 0.00010704994201660156, 0.00015556812286376953, 0.0002040863037109375, 0.00025260448455810547, 0.00030112266540527344, 0.0003496408462524414, 0.0003981590270996094, 0.00044667720794677734, 0.0004951953887939453, 0.0005437135696411133, 0.0005922317504882812, 0.0006407499313354492, 0.0006892681121826172, 0.0007377862930297852, 0.0007863044738769531, 0.0008348226547241211, 0.0008833408355712891, 0.000931859016418457, 0.000980377197265625, 0.001028895378112793, 0.001077413558959961, 0.001125931739807129, 0.0011744499206542969, 0.0012229681015014648, 0.0012714862823486328, 0.0013200044631958008, 0.0013685226440429688, 0.0014170408248901367, 0.0014655590057373047, 0.0015140771865844727, 0.0015625953674316406, 0.0016111135482788086, 0.0016596317291259766, 0.0017081499099731445, 0.0017566680908203125, 0.0018051862716674805, 0.0018537044525146484, 0.0019022226333618164, 0.0019507408142089844, 0.0019992589950561523, 0.0020477771759033203, 0.0020962953567504883, 0.0021448135375976562, 0.0021933317184448242, 0.002241849899291992, 0.00229036808013916, 0.002338886260986328, 0.002387404441833496, 0.002435922622680664, 0.002484440803527832, 0.002532958984375]}, "gradients/model.layers.19.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [3.0, 1.0, 1.0, 3.0, 3.0, 5.0, 7.0, 8.0, 12.0, 20.0, 28.0, 44.0, 47.0, 109.0, 138.0, 227.0, 368.0, 540.0, 1009.0, 1766.0, 3510.0, 8199.0, 18607.0, 50943.0, 191848.0, 1224195.0, 13241384.0, 16777216.0, 8846191.0, 813204.0, 145820.0, 41507.0, 15812.0, 6476.0, 3196.0, 1637.0, 848.0, 565.0, 328.0, 212.0, 146.0, 77.0, 58.0, 37.0, 18.0, 7.0, 6.0, 5.0, 7.0, 4.0, 2.0, 1.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.00127410888671875, -0.0012276172637939453, -0.0011811256408691406, -0.001134634017944336, -0.0010881423950195312, -0.0010416507720947266, -0.0009951591491699219, -0.0009486675262451172, -0.0009021759033203125, -0.0008556842803955078, -0.0008091926574707031, -0.0007627010345458984, -0.0007162094116210938, -0.0006697177886962891, -0.0006232261657714844, -0.0005767345428466797, -0.000530242919921875, -0.0004837512969970703, -0.0004372596740722656, -0.00039076805114746094, -0.00034427642822265625, -0.00029778480529785156, -0.0002512931823730469, -0.0002048015594482422, -0.0001583099365234375, -0.00011181831359863281, -6.532669067382812e-05, -1.8835067749023438e-05, 2.765655517578125e-05, 7.414817810058594e-05, 0.00012063980102539062, 0.0001671314239501953, 0.000213623046875, 0.0002601146697998047, 0.0003066062927246094, 0.00035309791564941406, 0.00039958953857421875, 0.00044608116149902344, 0.0004925727844238281, 0.0005390644073486328, 0.0005855560302734375, 0.0006320476531982422, 0.0006785392761230469, 0.0007250308990478516, 0.0007715225219726562, 0.0008180141448974609, 0.0008645057678222656, 0.0009109973907470703, 0.000957489013671875, 0.0010039806365966797, 0.0010504722595214844, 0.001096963882446289, 0.0011434555053710938, 0.0011899471282958984, 0.0012364387512207031, 0.0012829303741455078, 0.0013294219970703125, 0.0013759136199951172, 0.0014224052429199219, 0.0014688968658447266, 0.0015153884887695312, 0.001561880111694336, 0.0016083717346191406, 0.0016548633575439453, 0.00170135498046875]}, "gradients/model.layers.19.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 0.0, 3.0, 2.0, 1.0, 1.0, 7.0, 8.0, 6.0, 5.0, 14.0, 13.0, 17.0, 28.0, 64.0, 103.0, 179.0, 398.0, 923.0, 2512.0, 10191.0, 84569.0, 11858050.0, 16777216.0, 344185.0, 22475.0, 4349.0, 1366.0, 562.0, 260.0, 128.0, 65.0, 42.0, 18.0, 11.0, 17.0, 4.0, 10.0, 5.0, 4.0, 1.0, 6.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.004974365234375, -0.004844188690185547, -0.004714012145996094, -0.004583835601806641, -0.0044536590576171875, -0.004323482513427734, -0.004193305969238281, -0.004063129425048828, -0.003932952880859375, -0.003802776336669922, -0.0036725997924804688, -0.0035424232482910156, -0.0034122467041015625, -0.0032820701599121094, -0.0031518936157226562, -0.003021717071533203, -0.00289154052734375, -0.002761363983154297, -0.0026311874389648438, -0.0025010108947753906, -0.0023708343505859375, -0.0022406578063964844, -0.0021104812622070312, -0.001980304718017578, -0.001850128173828125, -0.0017199516296386719, -0.0015897750854492188, -0.0014595985412597656, -0.0013294219970703125, -0.0011992454528808594, -0.0010690689086914062, -0.0009388923645019531, -0.0008087158203125, -0.0006785392761230469, -0.0005483627319335938, -0.0004181861877441406, -0.0002880096435546875, -0.00015783309936523438, -2.765655517578125e-05, 0.00010251998901367188, 0.000232696533203125, 0.0003628730773925781, 0.0004930496215820312, 0.0006232261657714844, 0.0007534027099609375, 0.0008835792541503906, 0.0010137557983398438, 0.0011439323425292969, 0.00127410888671875, 0.0014042854309082031, 0.0015344619750976562, 0.0016646385192871094, 0.0017948150634765625, 0.0019249916076660156, 0.0020551681518554688, 0.002185344696044922, 0.002315521240234375, 0.002445697784423828, 0.0025758743286132812, 0.0027060508728027344, 0.0028362274169921875, 0.0029664039611816406, 0.0030965805053710938, 0.003226757049560547, 0.00335693359375]}, "gradients/model.layers.19.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 2.0, 5.0, 3.0, 2.0, 8.0, 7.0, 8.0, 12.0, 18.0, 72.0, 191.0, 434.0, 750.0, 866.0, 796.0, 508.0, 243.0, 123.0, 30.0, 5.0, 3.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.0032196044921875, -0.003152012825012207, -0.003084421157836914, -0.003016829490661621, -0.002949237823486328, -0.002881646156311035, -0.002814054489135742, -0.0027464628219604492, -0.0026788711547851562, -0.0026112794876098633, -0.0025436878204345703, -0.0024760961532592773, -0.0024085044860839844, -0.0023409128189086914, -0.0022733211517333984, -0.0022057294845581055, -0.0021381378173828125, -0.0020705461502075195, -0.0020029544830322266, -0.0019353628158569336, -0.0018677711486816406, -0.0018001794815063477, -0.0017325878143310547, -0.0016649961471557617, -0.0015974044799804688, -0.0015298128128051758, -0.0014622211456298828, -0.0013946294784545898, -0.0013270378112792969, -0.001259446144104004, -0.001191854476928711, -0.001124262809753418, -0.001056671142578125, -0.000989079475402832, -0.0009214878082275391, -0.0008538961410522461, -0.0007863044738769531, -0.0007187128067016602, -0.0006511211395263672, -0.0005835294723510742, -0.0005159378051757812, -0.0004483461380004883, -0.0003807544708251953, -0.00031316280364990234, -0.0002455711364746094, -0.0001779794692993164, -0.00011038780212402344, -4.279613494873047e-05, 2.47955322265625e-05, 9.238719940185547e-05, 0.00015997886657714844, 0.0002275705337524414, 0.0002951622009277344, 0.00036275386810302734, 0.0004303455352783203, 0.0004979372024536133, 0.0005655288696289062, 0.0006331205368041992, 0.0007007122039794922, 0.0007683038711547852, 0.0008358955383300781, 0.0009034872055053711, 0.0009710788726806641, 0.001038670539855957, 0.00110626220703125]}, "gradients/model.layers.19.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 4.0, 2.0, 7.0, 7.0, 9.0, 8.0, 17.0, 24.0, 42.0, 68.0, 106.0, 124.0, 206.0, 316.0, 436.0, 812.0, 1397.0, 2572.0, 5422.0, 12492.0, 35864.0, 108787.0, 356902.0, 1726541.0, 9790643.0, 3862886.0, 612113.0, 169770.0, 55489.0, 18805.0, 7362.0, 3490.0, 1771.0, 1032.0, 605.0, 371.0, 227.0, 168.0, 103.0, 63.0, 61.0, 30.0, 18.0, 15.0, 8.0, 6.0, 2.0, 1.0, 0.0, 1.0, 2.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00119781494140625, -0.0011600255966186523, -0.0011222362518310547, -0.001084446907043457, -0.0010466575622558594, -0.0010088682174682617, -0.0009710788726806641, -0.0009332895278930664, -0.0008955001831054688, -0.0008577108383178711, -0.0008199214935302734, -0.0007821321487426758, -0.0007443428039550781, -0.0007065534591674805, -0.0006687641143798828, -0.0006309747695922852, -0.0005931854248046875, -0.0005553960800170898, -0.0005176067352294922, -0.00047981739044189453, -0.0004420280456542969, -0.0004042387008666992, -0.00036644935607910156, -0.0003286600112915039, -0.00029087066650390625, -0.0002530813217163086, -0.00021529197692871094, -0.00017750263214111328, -0.00013971328735351562, -0.00010192394256591797, -6.413459777832031e-05, -2.6345252990722656e-05, 1.1444091796875e-05, 4.9233436584472656e-05, 8.702278137207031e-05, 0.00012481212615966797, 0.00016260147094726562, 0.00020039081573486328, 0.00023818016052246094, 0.0002759695053100586, 0.00031375885009765625, 0.0003515481948852539, 0.00038933753967285156, 0.0004271268844604492, 0.0004649162292480469, 0.0005027055740356445, 0.0005404949188232422, 0.0005782842636108398, 0.0006160736083984375, 0.0006538629531860352, 0.0006916522979736328, 0.0007294416427612305, 0.0007672309875488281, 0.0008050203323364258, 0.0008428096771240234, 0.0008805990219116211, 0.0009183883666992188, 0.0009561777114868164, 0.000993967056274414, 0.0010317564010620117, 0.0010695457458496094, 0.001107335090637207, 0.0011451244354248047, 0.0011829137802124023, 0.001220703125]}, "gradients/model.layers.19.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 2.0, 2.0, 4.0, 12.0, 5.0, 13.0, 12.0, 16.0, 16.0, 15.0, 21.0, 31.0, 44.0, 30.0, 53.0, 65.0, 117.0, 133.0, 203.0, 287.0, 496.0, 1016.0, 2730.0, 11517.0, 986103.0, 3162732.0, 21817.0, 3778.0, 1315.0, 548.0, 299.0, 207.0, 142.0, 120.0, 76.0, 53.0, 60.0, 41.0, 25.0, 35.0, 24.0, 19.0, 13.0, 14.0, 10.0, 7.0, 4.0, 4.0, 5.0, 1.0, 2.0, 0.0, 2.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.0078125, -0.007569313049316406, -0.0073261260986328125, -0.007082939147949219, -0.006839752197265625, -0.006596565246582031, -0.0063533782958984375, -0.006110191345214844, -0.00586700439453125, -0.005623817443847656, -0.0053806304931640625, -0.005137443542480469, -0.004894256591796875, -0.004651069641113281, -0.0044078826904296875, -0.004164695739746094, -0.0039215087890625, -0.0036783218383789062, -0.0034351348876953125, -0.0031919479370117188, -0.002948760986328125, -0.0027055740356445312, -0.0024623870849609375, -0.0022192001342773438, -0.00197601318359375, -0.0017328262329101562, -0.0014896392822265625, -0.0012464523315429688, -0.001003265380859375, -0.0007600784301757812, -0.0005168914794921875, -0.00027370452880859375, -3.0517578125e-05, 0.00021266937255859375, 0.0004558563232421875, 0.0006990432739257812, 0.000942230224609375, 0.0011854171752929688, 0.0014286041259765625, 0.0016717910766601562, 0.00191497802734375, 0.0021581649780273438, 0.0024013519287109375, 0.0026445388793945312, 0.002887725830078125, 0.0031309127807617188, 0.0033740997314453125, 0.0036172866821289062, 0.0038604736328125, 0.004103660583496094, 0.0043468475341796875, 0.004590034484863281, 0.004833221435546875, 0.005076408386230469, 0.0053195953369140625, 0.005562782287597656, 0.00580596923828125, 0.006049156188964844, 0.0062923431396484375, 0.006535530090332031, 0.006778717041015625, 0.007021903991699219, 0.0072650909423828125, 0.007508277893066406, 0.00775146484375]}, "gradients/model.layers.19.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 1.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 4.0, 1.0, 2.0, 2.0, 5.0, 7.0, 6.0, 12.0, 23.0, 58.0, 78.0, 168.0, 307.0, 672.0, 2205.0, 10475.0, 224362.0, 3915290.0, 33732.0, 4694.0, 1230.0, 458.0, 222.0, 106.0, 51.0, 44.0, 22.0, 16.0, 13.0, 9.0, 4.0, 7.0, 3.0, 2.0, 2.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0022125244140625, -0.002146005630493164, -0.002079486846923828, -0.002012968063354492, -0.0019464492797851562, -0.0018799304962158203, -0.0018134117126464844, -0.0017468929290771484, -0.0016803741455078125, -0.0016138553619384766, -0.0015473365783691406, -0.0014808177947998047, -0.0014142990112304688, -0.0013477802276611328, -0.0012812614440917969, -0.001214742660522461, -0.001148223876953125, -0.001081705093383789, -0.0010151863098144531, -0.0009486675262451172, -0.0008821487426757812, -0.0008156299591064453, -0.0007491111755371094, -0.0006825923919677734, -0.0006160736083984375, -0.0005495548248291016, -0.0004830360412597656, -0.0004165172576904297, -0.00034999847412109375, -0.0002834796905517578, -0.00021696090698242188, -0.00015044212341308594, -8.392333984375e-05, -1.7404556274414062e-05, 4.9114227294921875e-05, 0.00011563301086425781, 0.00018215179443359375, 0.0002486705780029297, 0.0003151893615722656, 0.00038170814514160156, 0.0004482269287109375, 0.0005147457122802734, 0.0005812644958496094, 0.0006477832794189453, 0.0007143020629882812, 0.0007808208465576172, 0.0008473396301269531, 0.0009138584136962891, 0.000980377197265625, 0.001046895980834961, 0.0011134147644042969, 0.0011799335479736328, 0.0012464523315429688, 0.0013129711151123047, 0.0013794898986816406, 0.0014460086822509766, 0.0015125274658203125, 0.0015790462493896484, 0.0016455650329589844, 0.0017120838165283203, 0.0017786026000976562, 0.0018451213836669922, 0.0019116401672363281, 0.001978158950805664, 0.002044677734375]}, "gradients/model.layers.19.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 0.0, 0.0, 1.0, 2.0, 2.0, 2.0, 6.0, 7.0, 15.0, 14.0, 28.0, 42.0, 105.0, 192.0, 525.0, 1609.0, 6963.0, 82215.0, 16415928.0, 255132.0, 10972.0, 2215.0, 758.0, 241.0, 104.0, 55.0, 17.0, 20.0, 14.0, 4.0, 4.0, 4.0, 2.0, 2.0, 2.0, 1.0, 1.0, 0.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.003021240234375, -0.002933979034423828, -0.0028467178344726562, -0.0027594566345214844, -0.0026721954345703125, -0.0025849342346191406, -0.0024976730346679688, -0.002410411834716797, -0.002323150634765625, -0.002235889434814453, -0.0021486282348632812, -0.0020613670349121094, -0.0019741058349609375, -0.0018868446350097656, -0.0017995834350585938, -0.0017123222351074219, -0.00162506103515625, -0.0015377998352050781, -0.0014505386352539062, -0.0013632774353027344, -0.0012760162353515625, -0.0011887550354003906, -0.0011014938354492188, -0.0010142326354980469, -0.000926971435546875, -0.0008397102355957031, -0.0007524490356445312, -0.0006651878356933594, -0.0005779266357421875, -0.0004906654357910156, -0.00040340423583984375, -0.0003161430358886719, -0.0002288818359375, -0.00014162063598632812, -5.435943603515625e-05, 3.2901763916015625e-05, 0.0001201629638671875, 0.00020742416381835938, 0.00029468536376953125, 0.0003819465637207031, 0.000469207763671875, 0.0005564689636230469, 0.0006437301635742188, 0.0007309913635253906, 0.0008182525634765625, 0.0009055137634277344, 0.0009927749633789062, 0.0010800361633300781, 0.00116729736328125, 0.0012545585632324219, 0.0013418197631835938, 0.0014290809631347656, 0.0015163421630859375, 0.0016036033630371094, 0.0016908645629882812, 0.0017781257629394531, 0.001865386962890625, 0.0019526481628417969, 0.0020399093627929688, 0.0021271705627441406, 0.0022144317626953125, 0.0023016929626464844, 0.0023889541625976562, 0.002476215362548828, 0.0025634765625]}, "gradients/model.layers.19.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 3.0, 3.0, 18.0, 77.0, 749.0, 2076.0, 1000.0, 124.0, 21.0, 8.0, 5.0, 0.0, 2.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.0032196044921875, -0.003154575824737549, -0.0030895471572875977, -0.0030245184898376465, -0.0029594898223876953, -0.002894461154937744, -0.002829432487487793, -0.002764403820037842, -0.0026993751525878906, -0.0026343464851379395, -0.0025693178176879883, -0.002504289150238037, -0.002439260482788086, -0.0023742318153381348, -0.0023092031478881836, -0.0022441744804382324, -0.0021791458129882812, -0.00211411714553833, -0.002049088478088379, -0.0019840598106384277, -0.0019190311431884766, -0.0018540024757385254, -0.0017889738082885742, -0.001723945140838623, -0.0016589164733886719, -0.0015938878059387207, -0.0015288591384887695, -0.0014638304710388184, -0.0013988018035888672, -0.001333773136138916, -0.0012687444686889648, -0.0012037158012390137, -0.0011386871337890625, -0.0010736584663391113, -0.0010086297988891602, -0.000943601131439209, -0.0008785724639892578, -0.0008135437965393066, -0.0007485151290893555, -0.0006834864616394043, -0.0006184577941894531, -0.000553429126739502, -0.0004884004592895508, -0.0004233717918395996, -0.00035834312438964844, -0.00029331445693969727, -0.0002282857894897461, -0.00016325712203979492, -9.822845458984375e-05, -3.319978713989258e-05, 3.1828880310058594e-05, 9.685754776000977e-05, 0.00016188621520996094, 0.0002269148826599121, 0.0002919435501098633, 0.00035697221755981445, 0.0004220008850097656, 0.0004870295524597168, 0.000552058219909668, 0.0006170868873596191, 0.0006821155548095703, 0.0007471442222595215, 0.0008121728897094727, 0.0008772015571594238, 0.000942230224609375]}, "gradients/model.layers.18.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 3.0, 3.0, 6.0, 5.0, 10.0, 19.0, 17.0, 46.0, 83.0, 95.0, 147.0, 218.0, 357.0, 734.0, 1643.0, 4959.0, 19985.0, 138317.0, 3413196.0, 16777216.0, 7851581.0, 224653.0, 27843.0, 6490.0, 1990.0, 851.0, 421.0, 264.0, 143.0, 103.0, 68.0, 53.0, 37.0, 22.0, 21.0, 9.0, 14.0, 8.0, 3.0, 3.0, 3.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0029449462890625, -0.002859830856323242, -0.0027747154235839844, -0.0026895999908447266, -0.0026044845581054688, -0.002519369125366211, -0.002434253692626953, -0.0023491382598876953, -0.0022640228271484375, -0.0021789073944091797, -0.002093791961669922, -0.002008676528930664, -0.0019235610961914062, -0.0018384456634521484, -0.0017533302307128906, -0.0016682147979736328, -0.001583099365234375, -0.0014979839324951172, -0.0014128684997558594, -0.0013277530670166016, -0.0012426376342773438, -0.001157522201538086, -0.0010724067687988281, -0.0009872913360595703, -0.0009021759033203125, -0.0008170604705810547, -0.0007319450378417969, -0.0006468296051025391, -0.0005617141723632812, -0.00047659873962402344, -0.0003914833068847656, -0.0003063678741455078, -0.00022125244140625, -0.0001361370086669922, -5.1021575927734375e-05, 3.409385681152344e-05, 0.00011920928955078125, 0.00020432472229003906, 0.0002894401550292969, 0.0003745555877685547, 0.0004596710205078125, 0.0005447864532470703, 0.0006299018859863281, 0.0007150173187255859, 0.0008001327514648438, 0.0008852481842041016, 0.0009703636169433594, 0.0010554790496826172, 0.001140594482421875, 0.0012257099151611328, 0.0013108253479003906, 0.0013959407806396484, 0.0014810562133789062, 0.001566171646118164, 0.0016512870788574219, 0.0017364025115966797, 0.0018215179443359375, 0.0019066333770751953, 0.001991748809814453, 0.002076864242553711, 0.0021619796752929688, 0.0022470951080322266, 0.0023322105407714844, 0.002417325973510742, 0.00250244140625]}, "gradients/model.layers.18.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 2.0, 4.0, 2.0, 4.0, 5.0, 11.0, 28.0, 46.0, 129.0, 351.0, 1047.0, 4300.0, 45836.0, 16777216.0, 16777216.0, 38324.0, 3828.0, 1001.0, 313.0, 124.0, 54.0, 21.0, 12.0, 12.0, 3.0, 4.0, 6.0, 1.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0059814453125, -0.005774974822998047, -0.005568504333496094, -0.005362033843994141, -0.0051555633544921875, -0.004949092864990234, -0.004742622375488281, -0.004536151885986328, -0.004329681396484375, -0.004123210906982422, -0.003916740417480469, -0.0037102699279785156, -0.0035037994384765625, -0.0032973289489746094, -0.0030908584594726562, -0.002884387969970703, -0.00267791748046875, -0.002471446990966797, -0.0022649765014648438, -0.0020585060119628906, -0.0018520355224609375, -0.0016455650329589844, -0.0014390945434570312, -0.0012326240539550781, -0.001026153564453125, -0.0008196830749511719, -0.0006132125854492188, -0.0004067420959472656, -0.0002002716064453125, 6.198883056640625e-06, 0.00021266937255859375, 0.0004191398620605469, 0.0006256103515625, 0.0008320808410644531, 0.0010385513305664062, 0.0012450218200683594, 0.0014514923095703125, 0.0016579627990722656, 0.0018644332885742188, 0.002070903778076172, 0.002277374267578125, 0.002483844757080078, 0.0026903152465820312, 0.0028967857360839844, 0.0031032562255859375, 0.0033097267150878906, 0.0035161972045898438, 0.003722667694091797, 0.00392913818359375, 0.004135608673095703, 0.004342079162597656, 0.004548549652099609, 0.0047550201416015625, 0.004961490631103516, 0.005167961120605469, 0.005374431610107422, 0.005580902099609375, 0.005787372589111328, 0.005993843078613281, 0.006200313568115234, 0.0064067840576171875, 0.006613254547119141, 0.006819725036621094, 0.007026195526123047, 0.007232666015625]}, "gradients/model.layers.18.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 3.0, 2.0, 4.0, 6.0, 20.0, 33.0, 71.0, 119.0, 194.0, 316.0, 404.0, 467.0, 549.0, 540.0, 461.0, 302.0, 258.0, 129.0, 90.0, 45.0, 32.0, 12.0, 8.0, 3.0, 6.0, 5.0, 1.0, 1.0, 2.0, 2.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.00164031982421875, -0.001596689224243164, -0.0015530586242675781, -0.0015094280242919922, -0.0014657974243164062, -0.0014221668243408203, -0.0013785362243652344, -0.0013349056243896484, -0.0012912750244140625, -0.0012476444244384766, -0.0012040138244628906, -0.0011603832244873047, -0.0011167526245117188, -0.0010731220245361328, -0.0010294914245605469, -0.000985860824584961, -0.000942230224609375, -0.0008985996246337891, -0.0008549690246582031, -0.0008113384246826172, -0.0007677078247070312, -0.0007240772247314453, -0.0006804466247558594, -0.0006368160247802734, -0.0005931854248046875, -0.0005495548248291016, -0.0005059242248535156, -0.0004622936248779297, -0.00041866302490234375, -0.0003750324249267578, -0.0003314018249511719, -0.00028777122497558594, -0.000244140625, -0.00020051002502441406, -0.00015687942504882812, -0.00011324882507324219, -6.961822509765625e-05, -2.5987625122070312e-05, 1.7642974853515625e-05, 6.127357482910156e-05, 0.0001049041748046875, 0.00014853477478027344, 0.00019216537475585938, 0.0002357959747314453, 0.00027942657470703125, 0.0003230571746826172, 0.0003666877746582031, 0.00041031837463378906, 0.000453948974609375, 0.0004975795745849609, 0.0005412101745605469, 0.0005848407745361328, 0.0006284713745117188, 0.0006721019744873047, 0.0007157325744628906, 0.0007593631744384766, 0.0008029937744140625, 0.0008466243743896484, 0.0008902549743652344, 0.0009338855743408203, 0.0009775161743164062, 0.0010211467742919922, 0.0010647773742675781, 0.001108407974243164, 0.00115203857421875]}, "gradients/model.layers.18.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 4.0, 0.0, 1.0, 0.0, 0.0, 3.0, 3.0, 6.0, 5.0, 9.0, 18.0, 28.0, 25.0, 39.0, 74.0, 81.0, 137.0, 240.0, 457.0, 801.0, 1550.0, 3731.0, 11270.0, 66470.0, 913032.0, 12308234.0, 3242017.0, 195855.0, 22566.0, 5847.0, 2235.0, 1151.0, 554.0, 274.0, 164.0, 110.0, 65.0, 45.0, 36.0, 23.0, 14.0, 12.0, 8.0, 4.0, 2.0, 7.0, 3.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.0026092529296875, -0.002526998519897461, -0.002444744110107422, -0.002362489700317383, -0.0022802352905273438, -0.0021979808807373047, -0.0021157264709472656, -0.0020334720611572266, -0.0019512176513671875, -0.0018689632415771484, -0.0017867088317871094, -0.0017044544219970703, -0.0016222000122070312, -0.0015399456024169922, -0.0014576911926269531, -0.001375436782836914, -0.001293182373046875, -0.001210927963256836, -0.0011286735534667969, -0.0010464191436767578, -0.0009641647338867188, -0.0008819103240966797, -0.0007996559143066406, -0.0007174015045166016, -0.0006351470947265625, -0.0005528926849365234, -0.0004706382751464844, -0.0003883838653564453, -0.00030612945556640625, -0.0002238750457763672, -0.00014162063598632812, -5.936622619628906e-05, 2.288818359375e-05, 0.00010514259338378906, 0.00018739700317382812, 0.0002696514129638672, 0.00035190582275390625, 0.0004341602325439453, 0.0005164146423339844, 0.0005986690521240234, 0.0006809234619140625, 0.0007631778717041016, 0.0008454322814941406, 0.0009276866912841797, 0.0010099411010742188, 0.0010921955108642578, 0.0011744499206542969, 0.001256704330444336, 0.001338958740234375, 0.001421213150024414, 0.0015034675598144531, 0.0015857219696044922, 0.0016679763793945312, 0.0017502307891845703, 0.0018324851989746094, 0.0019147396087646484, 0.0019969940185546875, 0.0020792484283447266, 0.0021615028381347656, 0.0022437572479248047, 0.0023260116577148438, 0.002408266067504883, 0.002490520477294922, 0.002572774887084961, 0.002655029296875]}, "gradients/model.layers.18.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 3.0, 4.0, 4.0, 6.0, 5.0, 3.0, 6.0, 11.0, 16.0, 12.0, 18.0, 30.0, 34.0, 46.0, 59.0, 70.0, 122.0, 149.0, 226.0, 342.0, 526.0, 824.0, 1491.0, 2801.0, 6576.0, 29366.0, 809442.0, 3139011.0, 179428.0, 13485.0, 4500.0, 2196.0, 1208.0, 696.0, 473.0, 320.0, 187.0, 160.0, 116.0, 76.0, 61.0, 54.0, 31.0, 22.0, 21.0, 19.0, 8.0, 11.0, 9.0, 3.0, 7.0, 1.0, 2.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.005157470703125, -0.004987239837646484, -0.004817008972167969, -0.004646778106689453, -0.0044765472412109375, -0.004306316375732422, -0.004136085510253906, -0.003965854644775391, -0.003795623779296875, -0.0036253929138183594, -0.0034551620483398438, -0.003284931182861328, -0.0031147003173828125, -0.002944469451904297, -0.0027742385864257812, -0.0026040077209472656, -0.00243377685546875, -0.0022635459899902344, -0.0020933151245117188, -0.0019230842590332031, -0.0017528533935546875, -0.0015826225280761719, -0.0014123916625976562, -0.0012421607971191406, -0.001071929931640625, -0.0009016990661621094, -0.0007314682006835938, -0.0005612373352050781, -0.0003910064697265625, -0.00022077560424804688, -5.054473876953125e-05, 0.00011968612670898438, 0.0002899169921875, 0.0004601478576660156, 0.0006303787231445312, 0.0008006095886230469, 0.0009708404541015625, 0.0011410713195800781, 0.0013113021850585938, 0.0014815330505371094, 0.001651763916015625, 0.0018219947814941406, 0.0019922256469726562, 0.002162456512451172, 0.0023326873779296875, 0.002502918243408203, 0.0026731491088867188, 0.0028433799743652344, 0.00301361083984375, 0.0031838417053222656, 0.0033540725708007812, 0.003524303436279297, 0.0036945343017578125, 0.003864765167236328, 0.004034996032714844, 0.004205226898193359, 0.004375457763671875, 0.004545688629150391, 0.004715919494628906, 0.004886150360107422, 0.0050563812255859375, 0.005226612091064453, 0.005396842956542969, 0.005567073822021484, 0.0057373046875]}, "gradients/model.layers.18.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 2.0, 3.0, 5.0, 10.0, 5.0, 10.0, 12.0, 24.0, 40.0, 69.0, 120.0, 213.0, 499.0, 1247.0, 4243.0, 24887.0, 2441976.0, 1692738.0, 22006.0, 4004.0, 1185.0, 502.0, 203.0, 109.0, 68.0, 43.0, 25.0, 7.0, 10.0, 9.0, 3.0, 2.0, 2.0, 2.0, 5.0, 3.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00164031982421875, -0.001583695411682129, -0.0015270709991455078, -0.0014704465866088867, -0.0014138221740722656, -0.0013571977615356445, -0.0013005733489990234, -0.0012439489364624023, -0.0011873245239257812, -0.0011307001113891602, -0.001074075698852539, -0.001017451286315918, -0.0009608268737792969, -0.0009042024612426758, -0.0008475780487060547, -0.0007909536361694336, -0.0007343292236328125, -0.0006777048110961914, -0.0006210803985595703, -0.0005644559860229492, -0.0005078315734863281, -0.00045120716094970703, -0.00039458274841308594, -0.00033795833587646484, -0.00028133392333984375, -0.00022470951080322266, -0.00016808509826660156, -0.00011146068572998047, -5.4836273193359375e-05, 1.7881393432617188e-06, 5.841255187988281e-05, 0.0001150369644165039, 0.000171661376953125, 0.0002282857894897461, 0.0002849102020263672, 0.0003415346145629883, 0.0003981590270996094, 0.00045478343963623047, 0.0005114078521728516, 0.0005680322647094727, 0.0006246566772460938, 0.0006812810897827148, 0.0007379055023193359, 0.000794529914855957, 0.0008511543273925781, 0.0009077787399291992, 0.0009644031524658203, 0.0010210275650024414, 0.0010776519775390625, 0.0011342763900756836, 0.0011909008026123047, 0.0012475252151489258, 0.0013041496276855469, 0.001360774040222168, 0.001417398452758789, 0.0014740228652954102, 0.0015306472778320312, 0.0015872716903686523, 0.0016438961029052734, 0.0017005205154418945, 0.0017571449279785156, 0.0018137693405151367, 0.0018703937530517578, 0.001927018165588379, 0.001983642578125]}, "gradients/model.layers.18.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 1.0, 1.0, 1.0, 2.0, 0.0, 1.0, 2.0, 1.0, 0.0, 2.0, 3.0, 4.0, 2.0, 8.0, 5.0, 14.0, 10.0, 23.0, 17.0, 27.0, 67.0, 69.0, 128.0, 182.0, 319.0, 778.0, 1758.0, 4544.0, 14586.0, 84099.0, 4523878.0, 11998873.0, 120901.0, 17582.0, 5306.0, 2097.0, 909.0, 417.0, 230.0, 124.0, 80.0, 41.0, 37.0, 27.0, 10.0, 9.0, 11.0, 6.0, 5.0, 3.0, 5.0, 1.0, 0.0, 1.0, 1.0, 0.0, 2.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.0017242431640625, -0.0016704797744750977, -0.0016167163848876953, -0.001562952995300293, -0.0015091896057128906, -0.0014554262161254883, -0.001401662826538086, -0.0013478994369506836, -0.0012941360473632812, -0.001240372657775879, -0.0011866092681884766, -0.0011328458786010742, -0.0010790824890136719, -0.0010253190994262695, -0.0009715557098388672, -0.0009177923202514648, -0.0008640289306640625, -0.0008102655410766602, -0.0007565021514892578, -0.0007027387619018555, -0.0006489753723144531, -0.0005952119827270508, -0.0005414485931396484, -0.0004876852035522461, -0.00043392181396484375, -0.0003801584243774414, -0.00032639503479003906, -0.0002726316452026367, -0.00021886825561523438, -0.00016510486602783203, -0.00011134147644042969, -5.7578086853027344e-05, -3.814697265625e-06, 4.9948692321777344e-05, 0.00010371208190917969, 0.00015747547149658203, 0.00021123886108398438, 0.0002650022506713867, 0.00031876564025878906, 0.0003725290298461914, 0.00042629241943359375, 0.0004800558090209961, 0.0005338191986083984, 0.0005875825881958008, 0.0006413459777832031, 0.0006951093673706055, 0.0007488727569580078, 0.0008026361465454102, 0.0008563995361328125, 0.0009101629257202148, 0.0009639263153076172, 0.0010176897048950195, 0.0010714530944824219, 0.0011252164840698242, 0.0011789798736572266, 0.001232743263244629, 0.0012865066528320312, 0.0013402700424194336, 0.001394033432006836, 0.0014477968215942383, 0.0015015602111816406, 0.001555323600769043, 0.0016090869903564453, 0.0016628503799438477, 0.00171661376953125]}, "gradients/model.layers.18.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 3.0, 6.0, 10.0, 15.0, 37.0, 91.0, 206.0, 451.0, 701.0, 882.0, 737.0, 527.0, 230.0, 116.0, 26.0, 11.0, 10.0, 10.0, 2.0, 2.0, 3.0, 2.0, 1.0, 0.0, 2.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.001068115234375, -0.001039743423461914, -0.0010113716125488281, -0.0009829998016357422, -0.0009546279907226562, -0.0009262561798095703, -0.0008978843688964844, -0.0008695125579833984, -0.0008411407470703125, -0.0008127689361572266, -0.0007843971252441406, -0.0007560253143310547, -0.0007276535034179688, -0.0006992816925048828, -0.0006709098815917969, -0.0006425380706787109, -0.000614166259765625, -0.0005857944488525391, -0.0005574226379394531, -0.0005290508270263672, -0.0005006790161132812, -0.0004723072052001953, -0.0004439353942871094, -0.00041556358337402344, -0.0003871917724609375, -0.00035881996154785156, -0.0003304481506347656, -0.0003020763397216797, -0.00027370452880859375, -0.0002453327178955078, -0.00021696090698242188, -0.00018858909606933594, -0.00016021728515625, -0.00013184547424316406, -0.00010347366333007812, -7.510185241699219e-05, -4.673004150390625e-05, -1.8358230590820312e-05, 1.0013580322265625e-05, 3.838539123535156e-05, 6.67572021484375e-05, 9.512901306152344e-05, 0.00012350082397460938, 0.0001518726348876953, 0.00018024444580078125, 0.0002086162567138672, 0.00023698806762695312, 0.00026535987854003906, 0.000293731689453125, 0.00032210350036621094, 0.0003504753112792969, 0.0003788471221923828, 0.00040721893310546875, 0.0004355907440185547, 0.0004639625549316406, 0.0004923343658447266, 0.0005207061767578125, 0.0005490779876708984, 0.0005774497985839844, 0.0006058216094970703, 0.0006341934204101562, 0.0006625652313232422, 0.0006909370422363281, 0.0007193088531494141, 0.0007476806640625]}, "gradients/model.layers.17.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 3.0, 2.0, 1.0, 3.0, 5.0, 7.0, 18.0, 43.0, 57.0, 147.0, 320.0, 745.0, 1942.0, 5367.0, 21261.0, 138562.0, 3010060.0, 16777216.0, 10952971.0, 315864.0, 37422.0, 8532.0, 2693.0, 1120.0, 441.0, 202.0, 84.0, 53.0, 13.0, 16.0, 6.0, 5.0, 1.0, 2.0, 1.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.0037994384765625, -0.0037103891372680664, -0.003621339797973633, -0.0035322904586791992, -0.0034432411193847656, -0.003354191780090332, -0.0032651424407958984, -0.003176093101501465, -0.0030870437622070312, -0.0029979944229125977, -0.002908945083618164, -0.0028198957443237305, -0.002730846405029297, -0.0026417970657348633, -0.0025527477264404297, -0.002463698387145996, -0.0023746490478515625, -0.002285599708557129, -0.0021965503692626953, -0.0021075010299682617, -0.002018451690673828, -0.0019294023513793945, -0.001840353012084961, -0.0017513036727905273, -0.0016622543334960938, -0.0015732049942016602, -0.0014841556549072266, -0.001395106315612793, -0.0013060569763183594, -0.0012170076370239258, -0.0011279582977294922, -0.0010389089584350586, -0.000949859619140625, -0.0008608102798461914, -0.0007717609405517578, -0.0006827116012573242, -0.0005936622619628906, -0.000504612922668457, -0.00041556358337402344, -0.00032651424407958984, -0.00023746490478515625, -0.00014841556549072266, -5.936622619628906e-05, 2.968311309814453e-05, 0.00011873245239257812, 0.00020778179168701172, 0.0002968311309814453, 0.0003858804702758789, 0.0004749298095703125, 0.0005639791488647461, 0.0006530284881591797, 0.0007420778274536133, 0.0008311271667480469, 0.0009201765060424805, 0.001009225845336914, 0.0010982751846313477, 0.0011873245239257812, 0.0012763738632202148, 0.0013654232025146484, 0.001454472541809082, 0.0015435218811035156, 0.0016325712203979492, 0.0017216205596923828, 0.0018106698989868164, 0.00189971923828125]}, "gradients/model.layers.17.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 2.0, 3.0, 3.0, 5.0, 4.0, 7.0, 18.0, 30.0, 41.0, 105.0, 217.0, 552.0, 1503.0, 4738.0, 44075.0, 16777216.0, 16777216.0, 45083.0, 4709.0, 1447.0, 594.0, 228.0, 105.0, 40.0, 26.0, 13.0, 9.0, 12.0, 1.0, 3.0, 4.0, 2.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0064697265625, -0.006267547607421875, -0.00606536865234375, -0.005863189697265625, -0.0056610107421875, -0.005458831787109375, -0.00525665283203125, -0.005054473876953125, -0.004852294921875, -0.004650115966796875, -0.00444793701171875, -0.004245758056640625, -0.0040435791015625, -0.003841400146484375, -0.00363922119140625, -0.003437042236328125, -0.00323486328125, -0.003032684326171875, -0.00283050537109375, -0.002628326416015625, -0.0024261474609375, -0.002223968505859375, -0.00202178955078125, -0.001819610595703125, -0.001617431640625, -0.001415252685546875, -0.00121307373046875, -0.001010894775390625, -0.0008087158203125, -0.000606536865234375, -0.00040435791015625, -0.000202178955078125, 0.0, 0.000202178955078125, 0.00040435791015625, 0.000606536865234375, 0.0008087158203125, 0.001010894775390625, 0.00121307373046875, 0.001415252685546875, 0.001617431640625, 0.001819610595703125, 0.00202178955078125, 0.002223968505859375, 0.0024261474609375, 0.002628326416015625, 0.00283050537109375, 0.003032684326171875, 0.00323486328125, 0.003437042236328125, 0.00363922119140625, 0.003841400146484375, 0.0040435791015625, 0.004245758056640625, 0.00444793701171875, 0.004650115966796875, 0.004852294921875, 0.005054473876953125, 0.00525665283203125, 0.005458831787109375, 0.0056610107421875, 0.005863189697265625, 0.00606536865234375, 0.006267547607421875, 0.0064697265625]}, "gradients/model.layers.17.post_attention_layernorm.weight": {"_type": "histogram", "values": [2.0, 2.0, 1.0, 1.0, 1.0, 0.0, 8.0, 9.0, 20.0, 50.0, 121.0, 312.0, 527.0, 787.0, 780.0, 651.0, 419.0, 231.0, 107.0, 31.0, 16.0, 9.0, 5.0, 2.0, 0.0, 1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00103759765625, -0.0009658336639404297, -0.0008940696716308594, -0.0008223056793212891, -0.0007505416870117188, -0.0006787776947021484, -0.0006070137023925781, -0.0005352497100830078, -0.0004634857177734375, -0.0003917217254638672, -0.0003199577331542969, -0.00024819374084472656, -0.00017642974853515625, -0.00010466575622558594, -3.2901763916015625e-05, 3.886222839355469e-05, 0.000110626220703125, 0.0001823902130126953, 0.0002541542053222656, 0.00032591819763183594, 0.00039768218994140625, 0.00046944618225097656, 0.0005412101745605469, 0.0006129741668701172, 0.0006847381591796875, 0.0007565021514892578, 0.0008282661437988281, 0.0009000301361083984, 0.0009717941284179688, 0.001043558120727539, 0.0011153221130371094, 0.0011870861053466797, 0.00125885009765625, 0.0013306140899658203, 0.0014023780822753906, 0.001474142074584961, 0.0015459060668945312, 0.0016176700592041016, 0.0016894340515136719, 0.0017611980438232422, 0.0018329620361328125, 0.0019047260284423828, 0.001976490020751953, 0.0020482540130615234, 0.0021200180053710938, 0.002191781997680664, 0.0022635459899902344, 0.0023353099822998047, 0.002407073974609375, 0.0024788379669189453, 0.0025506019592285156, 0.002622365951538086, 0.0026941299438476562, 0.0027658939361572266, 0.002837657928466797, 0.002909421920776367, 0.0029811859130859375, 0.003052949905395508, 0.003124713897705078, 0.0031964778900146484, 0.0032682418823242188, 0.003340005874633789, 0.0034117698669433594, 0.0034835338592529297, 0.0035552978515625]}, "gradients/model.layers.17.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 2.0, 1.0, 10.0, 9.0, 17.0, 22.0, 26.0, 63.0, 83.0, 156.0, 263.0, 469.0, 878.0, 1613.0, 3216.0, 7026.0, 18377.0, 58610.0, 253939.0, 1525907.0, 8389542.0, 5427505.0, 868066.0, 155872.0, 41005.0, 13443.0, 5532.0, 2594.0, 1350.0, 659.0, 381.0, 232.0, 133.0, 65.0, 35.0, 36.0, 24.0, 14.0, 12.0, 9.0, 2.0, 0.0, 1.0, 1.0, 1.0, 0.0, 2.0, 1.0, 2.0, 1.0, 1.0, 1.0], "bins": [-0.0020751953125, -0.002012014389038086, -0.0019488334655761719, -0.0018856525421142578, -0.0018224716186523438, -0.0017592906951904297, -0.0016961097717285156, -0.0016329288482666016, -0.0015697479248046875, -0.0015065670013427734, -0.0014433860778808594, -0.0013802051544189453, -0.0013170242309570312, -0.0012538433074951172, -0.0011906623840332031, -0.001127481460571289, -0.001064300537109375, -0.001001119613647461, -0.0009379386901855469, -0.0008747577667236328, -0.0008115768432617188, -0.0007483959197998047, -0.0006852149963378906, -0.0006220340728759766, -0.0005588531494140625, -0.0004956722259521484, -0.0004324913024902344, -0.0003693103790283203, -0.00030612945556640625, -0.0002429485321044922, -0.00017976760864257812, -0.00011658668518066406, -5.340576171875e-05, 9.775161743164062e-06, 7.295608520507812e-05, 0.0001361370086669922, 0.00019931793212890625, 0.0002624988555908203, 0.0003256797790527344, 0.00038886070251464844, 0.0004520416259765625, 0.0005152225494384766, 0.0005784034729003906, 0.0006415843963623047, 0.0007047653198242188, 0.0007679462432861328, 0.0008311271667480469, 0.0008943080902099609, 0.000957489013671875, 0.001020669937133789, 0.0010838508605957031, 0.0011470317840576172, 0.0012102127075195312, 0.0012733936309814453, 0.0013365745544433594, 0.0013997554779052734, 0.0014629364013671875, 0.0015261173248291016, 0.0015892982482910156, 0.0016524791717529297, 0.0017156600952148438, 0.0017788410186767578, 0.0018420219421386719, 0.001905202865600586, 0.0019683837890625]}, "gradients/model.layers.17.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 2.0, 1.0, 2.0, 2.0, 3.0, 3.0, 1.0, 2.0, 1.0, 1.0, 2.0, 6.0, 4.0, 10.0, 11.0, 13.0, 8.0, 17.0, 27.0, 59.0, 72.0, 93.0, 168.0, 288.0, 393.0, 865.0, 1710.0, 3848.0, 11108.0, 452392.0, 3644982.0, 65245.0, 7041.0, 2773.0, 1438.0, 642.0, 412.0, 221.0, 137.0, 89.0, 61.0, 35.0, 26.0, 18.0, 22.0, 8.0, 9.0, 9.0, 5.0, 2.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 1.0, 4.0, 0.0, 2.0, 2.0], "bins": [-0.00927734375, -0.008990287780761719, -0.008703231811523438, -0.008416175842285156, -0.008129119873046875, -0.007842063903808594, -0.0075550079345703125, -0.007267951965332031, -0.00698089599609375, -0.006693840026855469, -0.0064067840576171875, -0.006119728088378906, -0.005832672119140625, -0.005545616149902344, -0.0052585601806640625, -0.004971504211425781, -0.0046844482421875, -0.004397392272949219, -0.0041103363037109375, -0.0038232803344726562, -0.003536224365234375, -0.0032491683959960938, -0.0029621124267578125, -0.0026750564575195312, -0.00238800048828125, -0.0021009445190429688, -0.0018138885498046875, -0.0015268325805664062, -0.001239776611328125, -0.0009527206420898438, -0.0006656646728515625, -0.00037860870361328125, -9.1552734375e-05, 0.00019550323486328125, 0.0004825592041015625, 0.0007696151733398438, 0.001056671142578125, 0.0013437271118164062, 0.0016307830810546875, 0.0019178390502929688, 0.00220489501953125, 0.0024919509887695312, 0.0027790069580078125, 0.0030660629272460938, 0.003353118896484375, 0.0036401748657226562, 0.0039272308349609375, 0.004214286804199219, 0.0045013427734375, 0.004788398742675781, 0.0050754547119140625, 0.005362510681152344, 0.005649566650390625, 0.005936622619628906, 0.0062236785888671875, 0.006510734558105469, 0.00679779052734375, 0.007084846496582031, 0.0073719024658203125, 0.007658958435058594, 0.007946014404296875, 0.008233070373535156, 0.008520126342773438, 0.008807182312011719, 0.00909423828125]}, "gradients/model.layers.17.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 6.0, 2.0, 7.0, 8.0, 17.0, 23.0, 36.0, 47.0, 87.0, 123.0, 292.0, 726.0, 2416.0, 9536.0, 67361.0, 3917193.0, 177222.0, 14115.0, 3289.0, 1017.0, 355.0, 161.0, 69.0, 56.0, 38.0, 21.0, 28.0, 10.0, 8.0, 9.0, 4.0, 5.0, 0.0, 0.0, 1.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00194549560546875, -0.0018819570541381836, -0.0018184185028076172, -0.0017548799514770508, -0.0016913414001464844, -0.001627802848815918, -0.0015642642974853516, -0.0015007257461547852, -0.0014371871948242188, -0.0013736486434936523, -0.001310110092163086, -0.0012465715408325195, -0.0011830329895019531, -0.0011194944381713867, -0.0010559558868408203, -0.000992417335510254, -0.0009288787841796875, -0.0008653402328491211, -0.0008018016815185547, -0.0007382631301879883, -0.0006747245788574219, -0.0006111860275268555, -0.0005476474761962891, -0.00048410892486572266, -0.00042057037353515625, -0.00035703182220458984, -0.00029349327087402344, -0.00022995471954345703, -0.00016641616821289062, -0.00010287761688232422, -3.933906555175781e-05, 2.4199485778808594e-05, 8.7738037109375e-05, 0.0001512765884399414, 0.0002148151397705078, 0.0002783536911010742, 0.0003418922424316406, 0.00040543079376220703, 0.00046896934509277344, 0.0005325078964233398, 0.0005960464477539062, 0.0006595849990844727, 0.0007231235504150391, 0.0007866621017456055, 0.0008502006530761719, 0.0009137392044067383, 0.0009772777557373047, 0.001040816307067871, 0.0011043548583984375, 0.001167893409729004, 0.0012314319610595703, 0.0012949705123901367, 0.0013585090637207031, 0.0014220476150512695, 0.001485586166381836, 0.0015491247177124023, 0.0016126632690429688, 0.0016762018203735352, 0.0017397403717041016, 0.001803278923034668, 0.0018668174743652344, 0.0019303560256958008, 0.001993894577026367, 0.0020574331283569336, 0.0021209716796875]}, "gradients/model.layers.17.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 0.0, 3.0, 0.0, 2.0, 4.0, 2.0, 6.0, 4.0, 1.0, 4.0, 5.0, 7.0, 9.0, 9.0, 13.0, 26.0, 30.0, 46.0, 71.0, 139.0, 172.0, 356.0, 685.0, 1475.0, 3846.0, 11802.0, 61473.0, 3409870.0, 13153101.0, 109647.0, 15731.0, 4879.0, 1872.0, 903.0, 399.0, 199.0, 133.0, 90.0, 50.0, 36.0, 22.0, 23.0, 12.0, 10.0, 8.0, 5.0, 10.0, 4.0, 4.0, 4.0, 3.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0], "bins": [-0.0021514892578125, -0.0020902156829833984, -0.002028942108154297, -0.0019676685333251953, -0.0019063949584960938, -0.0018451213836669922, -0.0017838478088378906, -0.001722574234008789, -0.0016613006591796875, -0.001600027084350586, -0.0015387535095214844, -0.0014774799346923828, -0.0014162063598632812, -0.0013549327850341797, -0.0012936592102050781, -0.0012323856353759766, -0.001171112060546875, -0.0011098384857177734, -0.0010485649108886719, -0.0009872913360595703, -0.0009260177612304688, -0.0008647441864013672, -0.0008034706115722656, -0.0007421970367431641, -0.0006809234619140625, -0.0006196498870849609, -0.0005583763122558594, -0.0004971027374267578, -0.00043582916259765625, -0.0003745555877685547, -0.0003132820129394531, -0.00025200843811035156, -0.00019073486328125, -0.00012946128845214844, -6.818771362304688e-05, -6.9141387939453125e-06, 5.435943603515625e-05, 0.00011563301086425781, 0.00017690658569335938, 0.00023818016052246094, 0.0002994537353515625, 0.00036072731018066406, 0.0004220008850097656, 0.0004832744598388672, 0.0005445480346679688, 0.0006058216094970703, 0.0006670951843261719, 0.0007283687591552734, 0.000789642333984375, 0.0008509159088134766, 0.0009121894836425781, 0.0009734630584716797, 0.0010347366333007812, 0.0010960102081298828, 0.0011572837829589844, 0.001218557357788086, 0.0012798309326171875, 0.001341104507446289, 0.0014023780822753906, 0.0014636516571044922, 0.0015249252319335938, 0.0015861988067626953, 0.0016474723815917969, 0.0017087459564208984, 0.00177001953125]}, "gradients/model.layers.17.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 3.0, 1.0, 1.0, 3.0, 3.0, 3.0, 0.0, 3.0, 6.0, 6.0, 19.0, 31.0, 63.0, 127.0, 254.0, 493.0, 609.0, 676.0, 685.0, 525.0, 271.0, 157.0, 89.0, 25.0, 13.0, 6.0, 8.0, 3.0, 4.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.00098419189453125, -0.0009557604789733887, -0.0009273290634155273, -0.000898897647857666, -0.0008704662322998047, -0.0008420348167419434, -0.000813603401184082, -0.0007851719856262207, -0.0007567405700683594, -0.000728309154510498, -0.0006998777389526367, -0.0006714463233947754, -0.0006430149078369141, -0.0006145834922790527, -0.0005861520767211914, -0.0005577206611633301, -0.0005292892456054688, -0.0005008578300476074, -0.0004724264144897461, -0.00044399499893188477, -0.00041556358337402344, -0.0003871321678161621, -0.0003587007522583008, -0.00033026933670043945, -0.0003018379211425781, -0.0002734065055847168, -0.00024497509002685547, -0.00021654367446899414, -0.0001881122589111328, -0.00015968084335327148, -0.00013124942779541016, -0.00010281801223754883, -7.43865966796875e-05, -4.595518112182617e-05, -1.7523765563964844e-05, 1.0907649993896484e-05, 3.933906555175781e-05, 6.777048110961914e-05, 9.620189666748047e-05, 0.0001246333122253418, 0.00015306472778320312, 0.00018149614334106445, 0.00020992755889892578, 0.0002383589744567871, 0.00026679039001464844, 0.00029522180557250977, 0.0003236532211303711, 0.0003520846366882324, 0.00038051605224609375, 0.0004089474678039551, 0.0004373788833618164, 0.00046581029891967773, 0.0004942417144775391, 0.0005226731300354004, 0.0005511045455932617, 0.000579535961151123, 0.0006079673767089844, 0.0006363987922668457, 0.000664830207824707, 0.0006932616233825684, 0.0007216930389404297, 0.000750124454498291, 0.0007785558700561523, 0.0008069872856140137, 0.000835418701171875]}, "gradients/model.layers.16.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 3.0, 2.0, 5.0, 6.0, 5.0, 12.0, 29.0, 45.0, 78.0, 112.0, 237.0, 333.0, 616.0, 1252.0, 3766.0, 19595.0, 276273.0, 16777216.0, 16777216.0, 162744.0, 14717.0, 3192.0, 1063.0, 538.0, 351.0, 234.0, 149.0, 71.0, 49.0, 24.0, 11.0, 9.0, 4.0, 2.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.005126953125, -0.004991769790649414, -0.004856586456298828, -0.004721403121948242, -0.004586219787597656, -0.00445103645324707, -0.004315853118896484, -0.0041806697845458984, -0.0040454864501953125, -0.0039103031158447266, -0.0037751197814941406, -0.0036399364471435547, -0.0035047531127929688, -0.003369569778442383, -0.003234386444091797, -0.003099203109741211, -0.002964019775390625, -0.002828836441040039, -0.002693653106689453, -0.002558469772338867, -0.0024232864379882812, -0.0022881031036376953, -0.0021529197692871094, -0.0020177364349365234, -0.0018825531005859375, -0.0017473697662353516, -0.0016121864318847656, -0.0014770030975341797, -0.0013418197631835938, -0.0012066364288330078, -0.0010714530944824219, -0.0009362697601318359, -0.00080108642578125, -0.0006659030914306641, -0.0005307197570800781, -0.0003955364227294922, -0.00026035308837890625, -0.0001251697540283203, 1.0013580322265625e-05, 0.00014519691467285156, 0.0002803802490234375, 0.00041556358337402344, 0.0005507469177246094, 0.0006859302520751953, 0.0008211135864257812, 0.0009562969207763672, 0.0010914802551269531, 0.001226663589477539, 0.001361846923828125, 0.001497030258178711, 0.0016322135925292969, 0.0017673969268798828, 0.0019025802612304688, 0.0020377635955810547, 0.0021729469299316406, 0.0023081302642822266, 0.0024433135986328125, 0.0025784969329833984, 0.0027136802673339844, 0.0028488636016845703, 0.0029840469360351562, 0.003119230270385742, 0.003254413604736328, 0.003389596939086914, 0.0035247802734375]}, "gradients/model.layers.16.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 4.0, 4.0, 6.0, 8.0, 16.0, 31.0, 37.0, 87.0, 143.0, 271.0, 569.0, 1152.0, 2775.0, 11121.0, 172677.0, 16777216.0, 11179187.0, 46105.0, 5868.0, 1844.0, 829.0, 414.0, 192.0, 115.0, 75.0, 35.0, 33.0, 12.0, 9.0, 7.0, 5.0, 4.0, 3.0, 2.0, 3.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.006195068359375, -0.006026744842529297, -0.005858421325683594, -0.005690097808837891, -0.0055217742919921875, -0.005353450775146484, -0.005185127258300781, -0.005016803741455078, -0.004848480224609375, -0.004680156707763672, -0.004511833190917969, -0.004343509674072266, -0.0041751861572265625, -0.004006862640380859, -0.0038385391235351562, -0.003670215606689453, -0.00350189208984375, -0.003333568572998047, -0.0031652450561523438, -0.0029969215393066406, -0.0028285980224609375, -0.0026602745056152344, -0.0024919509887695312, -0.002323627471923828, -0.002155303955078125, -0.001986980438232422, -0.0018186569213867188, -0.0016503334045410156, -0.0014820098876953125, -0.0013136863708496094, -0.0011453628540039062, -0.0009770393371582031, -0.0008087158203125, -0.0006403923034667969, -0.00047206878662109375, -0.0003037452697753906, -0.0001354217529296875, 3.2901763916015625e-05, 0.00020122528076171875, 0.0003695487976074219, 0.000537872314453125, 0.0007061958312988281, 0.0008745193481445312, 0.0010428428649902344, 0.0012111663818359375, 0.0013794898986816406, 0.0015478134155273438, 0.0017161369323730469, 0.00188446044921875, 0.002052783966064453, 0.0022211074829101562, 0.0023894309997558594, 0.0025577545166015625, 0.0027260780334472656, 0.0028944015502929688, 0.003062725067138672, 0.003231048583984375, 0.003399372100830078, 0.0035676956176757812, 0.0037360191345214844, 0.0039043426513671875, 0.004072666168212891, 0.004240989685058594, 0.004409313201904297, 0.00457763671875]}, "gradients/model.layers.16.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 0.0, 2.0, 8.0, 20.0, 19.0, 55.0, 91.0, 189.0, 337.0, 464.0, 570.0, 613.0, 564.0, 455.0, 307.0, 190.0, 88.0, 54.0, 22.0, 11.0, 7.0, 8.0, 6.0, 2.0, 2.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0019989013671875, -0.0019439458847045898, -0.0018889904022216797, -0.0018340349197387695, -0.0017790794372558594, -0.0017241239547729492, -0.001669168472290039, -0.001614212989807129, -0.0015592575073242188, -0.0015043020248413086, -0.0014493465423583984, -0.0013943910598754883, -0.0013394355773925781, -0.001284480094909668, -0.0012295246124267578, -0.0011745691299438477, -0.0011196136474609375, -0.0010646581649780273, -0.0010097026824951172, -0.000954747200012207, -0.0008997917175292969, -0.0008448362350463867, -0.0007898807525634766, -0.0007349252700805664, -0.0006799697875976562, -0.0006250143051147461, -0.0005700588226318359, -0.0005151033401489258, -0.0004601478576660156, -0.00040519237518310547, -0.0003502368927001953, -0.00029528141021728516, -0.000240325927734375, -0.00018537044525146484, -0.0001304149627685547, -7.545948028564453e-05, -2.0503997802734375e-05, 3.445148468017578e-05, 8.940696716308594e-05, 0.0001443624496459961, 0.00019931793212890625, 0.0002542734146118164, 0.00030922889709472656, 0.0003641843795776367, 0.0004191398620605469, 0.00047409534454345703, 0.0005290508270263672, 0.0005840063095092773, 0.0006389617919921875, 0.0006939172744750977, 0.0007488727569580078, 0.000803828239440918, 0.0008587837219238281, 0.0009137392044067383, 0.0009686946868896484, 0.0010236501693725586, 0.0010786056518554688, 0.001133561134338379, 0.001188516616821289, 0.0012434720993041992, 0.0012984275817871094, 0.0013533830642700195, 0.0014083385467529297, 0.0014632940292358398, 0.00151824951171875]}, "gradients/model.layers.16.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 2.0, 6.0, 9.0, 9.0, 9.0, 14.0, 30.0, 34.0, 75.0, 82.0, 165.0, 238.0, 461.0, 888.0, 1621.0, 3158.0, 6913.0, 17136.0, 61593.0, 353387.0, 2476243.0, 9673446.0, 3544180.0, 516515.0, 83666.0, 21243.0, 8314.0, 3809.0, 1785.0, 956.0, 498.0, 297.0, 160.0, 109.0, 53.0, 35.0, 23.0, 18.0, 12.0, 8.0, 6.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.00299072265625, -0.002913236618041992, -0.0028357505798339844, -0.0027582645416259766, -0.0026807785034179688, -0.002603292465209961, -0.002525806427001953, -0.0024483203887939453, -0.0023708343505859375, -0.0022933483123779297, -0.002215862274169922, -0.002138376235961914, -0.0020608901977539062, -0.0019834041595458984, -0.0019059181213378906, -0.0018284320831298828, -0.001750946044921875, -0.0016734600067138672, -0.0015959739685058594, -0.0015184879302978516, -0.0014410018920898438, -0.001363515853881836, -0.0012860298156738281, -0.0012085437774658203, -0.0011310577392578125, -0.0010535717010498047, -0.0009760856628417969, -0.0008985996246337891, -0.0008211135864257812, -0.0007436275482177734, -0.0006661415100097656, -0.0005886554718017578, -0.00051116943359375, -0.0004336833953857422, -0.0003561973571777344, -0.00027871131896972656, -0.00020122528076171875, -0.00012373924255371094, -4.6253204345703125e-05, 3.123283386230469e-05, 0.0001087188720703125, 0.0001862049102783203, 0.0002636909484863281, 0.00034117698669433594, 0.00041866302490234375, 0.0004961490631103516, 0.0005736351013183594, 0.0006511211395263672, 0.000728607177734375, 0.0008060932159423828, 0.0008835792541503906, 0.0009610652923583984, 0.0010385513305664062, 0.001116037368774414, 0.0011935234069824219, 0.0012710094451904297, 0.0013484954833984375, 0.0014259815216064453, 0.0015034675598144531, 0.001580953598022461, 0.0016584396362304688, 0.0017359256744384766, 0.0018134117126464844, 0.0018908977508544922, 0.0019683837890625]}, "gradients/model.layers.16.self_attn.v_proj.weight": {"_type": "histogram", "values": [5.0, 2.0, 6.0, 1.0, 6.0, 3.0, 2.0, 9.0, 9.0, 28.0, 18.0, 25.0, 26.0, 34.0, 62.0, 75.0, 119.0, 141.0, 181.0, 230.0, 361.0, 483.0, 673.0, 1030.0, 1413.0, 1945.0, 3249.0, 6107.0, 18240.0, 217398.0, 2545728.0, 1301004.0, 71662.0, 10763.0, 4601.0, 2746.0, 1764.0, 1159.0, 866.0, 555.0, 407.0, 314.0, 234.0, 137.0, 126.0, 79.0, 72.0, 44.0, 40.0, 26.0, 14.0, 18.0, 15.0, 16.0, 8.0, 5.0, 3.0, 7.0, 2.0, 3.0, 2.0, 0.0, 1.0, 2.0], "bins": [-0.005523681640625, -0.005344390869140625, -0.00516510009765625, -0.004985809326171875, -0.0048065185546875, -0.004627227783203125, -0.00444793701171875, -0.004268646240234375, -0.00408935546875, -0.003910064697265625, -0.00373077392578125, -0.003551483154296875, -0.0033721923828125, -0.003192901611328125, -0.00301361083984375, -0.002834320068359375, -0.002655029296875, -0.002475738525390625, -0.00229644775390625, -0.002117156982421875, -0.0019378662109375, -0.001758575439453125, -0.00157928466796875, -0.001399993896484375, -0.001220703125, -0.001041412353515625, -0.00086212158203125, -0.000682830810546875, -0.0005035400390625, -0.000324249267578125, -0.00014495849609375, 3.4332275390625e-05, 0.000213623046875, 0.000392913818359375, 0.00057220458984375, 0.000751495361328125, 0.0009307861328125, 0.001110076904296875, 0.00128936767578125, 0.001468658447265625, 0.00164794921875, 0.001827239990234375, 0.00200653076171875, 0.002185821533203125, 0.0023651123046875, 0.002544403076171875, 0.00272369384765625, 0.002902984619140625, 0.003082275390625, 0.003261566162109375, 0.00344085693359375, 0.003620147705078125, 0.0037994384765625, 0.003978729248046875, 0.00415802001953125, 0.004337310791015625, 0.0045166015625, 0.004695892333984375, 0.00487518310546875, 0.005054473876953125, 0.0052337646484375, 0.005413055419921875, 0.00559234619140625, 0.005771636962890625, 0.005950927734375]}, "gradients/model.layers.16.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 3.0, 4.0, 4.0, 3.0, 4.0, 5.0, 10.0, 7.0, 13.0, 21.0, 19.0, 39.0, 84.0, 150.0, 318.0, 822.0, 2744.0, 14154.0, 2077978.0, 2079039.0, 14503.0, 2888.0, 806.0, 324.0, 137.0, 83.0, 45.0, 23.0, 14.0, 9.0, 12.0, 5.0, 3.0, 6.0, 2.0, 5.0, 3.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.0032806396484375, -0.003178119659423828, -0.0030755996704101562, -0.0029730796813964844, -0.0028705596923828125, -0.0027680397033691406, -0.0026655197143554688, -0.002562999725341797, -0.002460479736328125, -0.002357959747314453, -0.0022554397583007812, -0.0021529197692871094, -0.0020503997802734375, -0.0019478797912597656, -0.0018453598022460938, -0.0017428398132324219, -0.00164031982421875, -0.0015377998352050781, -0.0014352798461914062, -0.0013327598571777344, -0.0012302398681640625, -0.0011277198791503906, -0.0010251998901367188, -0.0009226799011230469, -0.000820159912109375, -0.0007176399230957031, -0.0006151199340820312, -0.0005125999450683594, -0.0004100799560546875, -0.0003075599670410156, -0.00020503997802734375, -0.00010251998901367188, 0.0, 0.00010251998901367188, 0.00020503997802734375, 0.0003075599670410156, 0.0004100799560546875, 0.0005125999450683594, 0.0006151199340820312, 0.0007176399230957031, 0.000820159912109375, 0.0009226799011230469, 0.0010251998901367188, 0.0011277198791503906, 0.0012302398681640625, 0.0013327598571777344, 0.0014352798461914062, 0.0015377998352050781, 0.00164031982421875, 0.0017428398132324219, 0.0018453598022460938, 0.0019478797912597656, 0.0020503997802734375, 0.0021529197692871094, 0.0022554397583007812, 0.002357959747314453, 0.002460479736328125, 0.002562999725341797, 0.0026655197143554688, 0.0027680397033691406, 0.0028705596923828125, 0.0029730796813964844, 0.0030755996704101562, 0.003178119659423828, 0.0032806396484375]}, "gradients/model.layers.16.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 3.0, 5.0, 2.0, 2.0, 2.0, 6.0, 4.0, 9.0, 16.0, 16.0, 21.0, 52.0, 108.0, 211.0, 509.0, 1909.0, 12813.0, 13220350.0, 3527936.0, 10583.0, 1755.0, 479.0, 169.0, 104.0, 44.0, 18.0, 26.0, 12.0, 9.0, 10.0, 2.0, 5.0, 4.0, 4.0, 2.0, 4.0, 0.0, 1.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.006805419921875, -0.006598949432373047, -0.006392478942871094, -0.006186008453369141, -0.0059795379638671875, -0.005773067474365234, -0.005566596984863281, -0.005360126495361328, -0.005153656005859375, -0.004947185516357422, -0.004740715026855469, -0.004534244537353516, -0.0043277740478515625, -0.004121303558349609, -0.003914833068847656, -0.003708362579345703, -0.00350189208984375, -0.003295421600341797, -0.0030889511108398438, -0.0028824806213378906, -0.0026760101318359375, -0.0024695396423339844, -0.0022630691528320312, -0.002056598663330078, -0.001850128173828125, -0.0016436576843261719, -0.0014371871948242188, -0.0012307167053222656, -0.0010242462158203125, -0.0008177757263183594, -0.0006113052368164062, -0.0004048347473144531, -0.0001983642578125, 8.106231689453125e-06, 0.00021457672119140625, 0.0004210472106933594, 0.0006275177001953125, 0.0008339881896972656, 0.0010404586791992188, 0.0012469291687011719, 0.001453399658203125, 0.0016598701477050781, 0.0018663406372070312, 0.0020728111267089844, 0.0022792816162109375, 0.0024857521057128906, 0.0026922225952148438, 0.002898693084716797, 0.00310516357421875, 0.003311634063720703, 0.0035181045532226562, 0.0037245750427246094, 0.0039310455322265625, 0.004137516021728516, 0.004343986511230469, 0.004550457000732422, 0.004756927490234375, 0.004963397979736328, 0.005169868469238281, 0.005376338958740234, 0.0055828094482421875, 0.005789279937744141, 0.005995750427246094, 0.006202220916748047, 0.00640869140625]}, "gradients/model.layers.16.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 1.0, 4.0, 1.0, 5.0, 6.0, 6.0, 19.0, 20.0, 107.0, 436.0, 1045.0, 1308.0, 750.0, 282.0, 58.0, 18.0, 7.0, 5.0, 3.0, 2.0, 0.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00156402587890625, -0.0015045404434204102, -0.0014450550079345703, -0.0013855695724487305, -0.0013260841369628906, -0.0012665987014770508, -0.001207113265991211, -0.001147627830505371, -0.0010881423950195312, -0.0010286569595336914, -0.0009691715240478516, -0.0009096860885620117, -0.0008502006530761719, -0.000790715217590332, -0.0007312297821044922, -0.0006717443466186523, -0.0006122589111328125, -0.0005527734756469727, -0.0004932880401611328, -0.00043380260467529297, -0.0003743171691894531, -0.0003148317337036133, -0.00025534629821777344, -0.0001958608627319336, -0.00013637542724609375, -7.68899917602539e-05, -1.7404556274414062e-05, 4.208087921142578e-05, 0.00010156631469726562, 0.00016105175018310547, 0.0002205371856689453, 0.00028002262115478516, 0.000339508056640625, 0.00039899349212646484, 0.0004584789276123047, 0.0005179643630981445, 0.0005774497985839844, 0.0006369352340698242, 0.0006964206695556641, 0.0007559061050415039, 0.0008153915405273438, 0.0008748769760131836, 0.0009343624114990234, 0.0009938478469848633, 0.0010533332824707031, 0.001112818717956543, 0.0011723041534423828, 0.0012317895889282227, 0.0012912750244140625, 0.0013507604598999023, 0.0014102458953857422, 0.001469731330871582, 0.0015292167663574219, 0.0015887022018432617, 0.0016481876373291016, 0.0017076730728149414, 0.0017671585083007812, 0.001826643943786621, 0.001886129379272461, 0.0019456148147583008, 0.0020051002502441406, 0.0020645856857299805, 0.0021240711212158203, 0.00218355655670166, 0.0022430419921875]}, "gradients/model.layers.15.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 4.0, 4.0, 5.0, 5.0, 4.0, 8.0, 6.0, 11.0, 23.0, 24.0, 39.0, 47.0, 63.0, 108.0, 171.0, 263.0, 422.0, 766.0, 1415.0, 2586.0, 5207.0, 11017.0, 25456.0, 71702.0, 292408.0, 2045184.0, 14737780.0, 16777216.0, 9990799.0, 1275663.0, 207384.0, 56342.0, 20701.0, 9287.0, 4549.0, 2357.0, 1226.0, 675.0, 366.0, 260.0, 161.0, 89.0, 54.0, 53.0, 29.0, 27.0, 18.0, 9.0, 8.0, 6.0, 4.0, 5.0, 5.0, 0.0, 4.0, 3.0], "bins": [-0.0017547607421875, -0.001705169677734375, -0.00165557861328125, -0.001605987548828125, -0.001556396484375, -0.001506805419921875, -0.00145721435546875, -0.001407623291015625, -0.0013580322265625, -0.001308441162109375, -0.00125885009765625, -0.001209259033203125, -0.00115966796875, -0.001110076904296875, -0.00106048583984375, -0.001010894775390625, -0.0009613037109375, -0.000911712646484375, -0.00086212158203125, -0.000812530517578125, -0.000762939453125, -0.000713348388671875, -0.00066375732421875, -0.000614166259765625, -0.0005645751953125, -0.000514984130859375, -0.00046539306640625, -0.000415802001953125, -0.0003662109375, -0.000316619873046875, -0.00026702880859375, -0.000217437744140625, -0.0001678466796875, -0.000118255615234375, -6.866455078125e-05, -1.9073486328125e-05, 3.0517578125e-05, 8.0108642578125e-05, 0.00012969970703125, 0.000179290771484375, 0.0002288818359375, 0.000278472900390625, 0.00032806396484375, 0.000377655029296875, 0.00042724609375, 0.000476837158203125, 0.00052642822265625, 0.000576019287109375, 0.0006256103515625, 0.000675201416015625, 0.00072479248046875, 0.000774383544921875, 0.000823974609375, 0.000873565673828125, 0.00092315673828125, 0.000972747802734375, 0.0010223388671875, 0.001071929931640625, 0.00112152099609375, 0.001171112060546875, 0.001220703125, 0.001270294189453125, 0.00131988525390625, 0.001369476318359375, 0.0014190673828125]}, "gradients/model.layers.15.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [3.0, 1.0, 1.0, 1.0, 0.0, 6.0, 1.0, 7.0, 5.0, 6.0, 13.0, 17.0, 37.0, 72.0, 90.0, 165.0, 354.0, 673.0, 1547.0, 3940.0, 20240.0, 1279383.0, 16777216.0, 366272.0, 13722.0, 3289.0, 1268.0, 555.0, 331.0, 151.0, 92.0, 52.0, 31.0, 16.0, 14.0, 12.0, 3.0, 2.0, 1.0, 1.0, 0.0, 0.0, 3.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.004241943359375, -0.004052639007568359, -0.0038633346557617188, -0.003674030303955078, -0.0034847259521484375, -0.003295421600341797, -0.0031061172485351562, -0.0029168128967285156, -0.002727508544921875, -0.0025382041931152344, -0.0023488998413085938, -0.002159595489501953, -0.0019702911376953125, -0.0017809867858886719, -0.0015916824340820312, -0.0014023780822753906, -0.00121307373046875, -0.0010237693786621094, -0.0008344650268554688, -0.0006451606750488281, -0.0004558563232421875, -0.0002665519714355469, -7.724761962890625e-05, 0.00011205673217773438, 0.000301361083984375, 0.0004906654357910156, 0.0006799697875976562, 0.0008692741394042969, 0.0010585784912109375, 0.0012478828430175781, 0.0014371871948242188, 0.0016264915466308594, 0.0018157958984375, 0.0020051002502441406, 0.0021944046020507812, 0.002383708953857422, 0.0025730133056640625, 0.002762317657470703, 0.0029516220092773438, 0.0031409263610839844, 0.003330230712890625, 0.0035195350646972656, 0.0037088394165039062, 0.003898143768310547, 0.0040874481201171875, 0.004276752471923828, 0.004466056823730469, 0.004655361175537109, 0.00484466552734375, 0.005033969879150391, 0.005223274230957031, 0.005412578582763672, 0.0056018829345703125, 0.005791187286376953, 0.005980491638183594, 0.006169795989990234, 0.006359100341796875, 0.006548404693603516, 0.006737709045410156, 0.006927013397216797, 0.0071163177490234375, 0.007305622100830078, 0.007494926452636719, 0.007684230804443359, 0.00787353515625]}, "gradients/model.layers.15.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 5.0, 3.0, 16.0, 15.0, 34.0, 63.0, 97.0, 175.0, 240.0, 331.0, 435.0, 491.0, 506.0, 467.0, 381.0, 319.0, 229.0, 102.0, 70.0, 40.0, 25.0, 11.0, 8.0, 4.0, 5.0, 2.0, 2.0, 3.0, 0.0, 1.0, 0.0, 2.0, 1.0, 0.0, 1.0], "bins": [-0.0019683837890625, -0.001920461654663086, -0.0018725395202636719, -0.0018246173858642578, -0.0017766952514648438, -0.0017287731170654297, -0.0016808509826660156, -0.0016329288482666016, -0.0015850067138671875, -0.0015370845794677734, -0.0014891624450683594, -0.0014412403106689453, -0.0013933181762695312, -0.0013453960418701172, -0.0012974739074707031, -0.001249551773071289, -0.001201629638671875, -0.001153707504272461, -0.0011057853698730469, -0.0010578632354736328, -0.0010099411010742188, -0.0009620189666748047, -0.0009140968322753906, -0.0008661746978759766, -0.0008182525634765625, -0.0007703304290771484, -0.0007224082946777344, -0.0006744861602783203, -0.0006265640258789062, -0.0005786418914794922, -0.0005307197570800781, -0.00048279762268066406, -0.00043487548828125, -0.00038695335388183594, -0.0003390312194824219, -0.0002911090850830078, -0.00024318695068359375, -0.0001952648162841797, -0.00014734268188476562, -9.942054748535156e-05, -5.14984130859375e-05, -3.5762786865234375e-06, 4.4345855712890625e-05, 9.226799011230469e-05, 0.00014019012451171875, 0.0001881122589111328, 0.00023603439331054688, 0.00028395652770996094, 0.000331878662109375, 0.00037980079650878906, 0.0004277229309082031, 0.0004756450653076172, 0.0005235671997070312, 0.0005714893341064453, 0.0006194114685058594, 0.0006673336029052734, 0.0007152557373046875, 0.0007631778717041016, 0.0008111000061035156, 0.0008590221405029297, 0.0009069442749023438, 0.0009548664093017578, 0.0010027885437011719, 0.001050710678100586, 0.0010986328125]}, "gradients/model.layers.15.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 0.0, 2.0, 0.0, 1.0, 3.0, 3.0, 5.0, 8.0, 6.0, 13.0, 23.0, 35.0, 55.0, 100.0, 170.0, 303.0, 511.0, 1057.0, 2150.0, 5235.0, 14896.0, 60076.0, 436085.0, 3888238.0, 10420657.0, 1706405.0, 190363.0, 33959.0, 9737.0, 3905.0, 1560.0, 748.0, 385.0, 219.0, 110.0, 67.0, 44.0, 26.0, 14.0, 8.0, 4.0, 2.0, 4.0, 4.0, 7.0, 4.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.002838134765625, -0.0027341842651367188, -0.0026302337646484375, -0.0025262832641601562, -0.002422332763671875, -0.0023183822631835938, -0.0022144317626953125, -0.0021104812622070312, -0.00200653076171875, -0.0019025802612304688, -0.0017986297607421875, -0.0016946792602539062, -0.001590728759765625, -0.0014867782592773438, -0.0013828277587890625, -0.0012788772583007812, -0.0011749267578125, -0.0010709762573242188, -0.0009670257568359375, -0.0008630752563476562, -0.000759124755859375, -0.0006551742553710938, -0.0005512237548828125, -0.00044727325439453125, -0.00034332275390625, -0.00023937225341796875, -0.0001354217529296875, -3.147125244140625e-05, 7.2479248046875e-05, 0.00017642974853515625, 0.0002803802490234375, 0.00038433074951171875, 0.00048828125, 0.0005922317504882812, 0.0006961822509765625, 0.0008001327514648438, 0.000904083251953125, 0.0010080337524414062, 0.0011119842529296875, 0.0012159347534179688, 0.00131988525390625, 0.0014238357543945312, 0.0015277862548828125, 0.0016317367553710938, 0.001735687255859375, 0.0018396377563476562, 0.0019435882568359375, 0.0020475387573242188, 0.0021514892578125, 0.0022554397583007812, 0.0023593902587890625, 0.0024633407592773438, 0.002567291259765625, 0.0026712417602539062, 0.0027751922607421875, 0.0028791427612304688, 0.00298309326171875, 0.0030870437622070312, 0.0031909942626953125, 0.0032949447631835938, 0.003398895263671875, 0.0035028457641601562, 0.0036067962646484375, 0.0037107467651367188, 0.003814697265625]}, "gradients/model.layers.15.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 2.0, 1.0, 0.0, 3.0, 2.0, 3.0, 0.0, 10.0, 5.0, 5.0, 4.0, 6.0, 13.0, 12.0, 20.0, 22.0, 24.0, 51.0, 48.0, 71.0, 83.0, 127.0, 241.0, 415.0, 730.0, 1356.0, 2592.0, 6102.0, 50635.0, 3552752.0, 560734.0, 10527.0, 3597.0, 1782.0, 912.0, 511.0, 324.0, 155.0, 91.0, 75.0, 55.0, 44.0, 30.0, 32.0, 25.0, 16.0, 11.0, 2.0, 8.0, 8.0, 6.0, 2.0, 6.0, 2.0, 3.0, 3.0, 0.0, 1.0, 0.0, 3.0, 2.0], "bins": [-0.0111083984375, -0.010768890380859375, -0.01042938232421875, -0.010089874267578125, -0.0097503662109375, -0.009410858154296875, -0.00907135009765625, -0.008731842041015625, -0.008392333984375, -0.008052825927734375, -0.00771331787109375, -0.007373809814453125, -0.0070343017578125, -0.006694793701171875, -0.00635528564453125, -0.006015777587890625, -0.00567626953125, -0.005336761474609375, -0.00499725341796875, -0.004657745361328125, -0.0043182373046875, -0.003978729248046875, -0.00363922119140625, -0.003299713134765625, -0.002960205078125, -0.002620697021484375, -0.00228118896484375, -0.001941680908203125, -0.0016021728515625, -0.001262664794921875, -0.00092315673828125, -0.000583648681640625, -0.000244140625, 9.5367431640625e-05, 0.00043487548828125, 0.000774383544921875, 0.0011138916015625, 0.001453399658203125, 0.00179290771484375, 0.002132415771484375, 0.002471923828125, 0.002811431884765625, 0.00315093994140625, 0.003490447998046875, 0.0038299560546875, 0.004169464111328125, 0.00450897216796875, 0.004848480224609375, 0.00518798828125, 0.005527496337890625, 0.00586700439453125, 0.006206512451171875, 0.0065460205078125, 0.006885528564453125, 0.00722503662109375, 0.007564544677734375, 0.007904052734375, 0.008243560791015625, 0.00858306884765625, 0.008922576904296875, 0.0092620849609375, 0.009601593017578125, 0.00994110107421875, 0.010280609130859375, 0.0106201171875]}, "gradients/model.layers.15.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 3.0, 0.0, 1.0, 4.0, 7.0, 3.0, 7.0, 13.0, 18.0, 25.0, 27.0, 39.0, 66.0, 99.0, 162.0, 328.0, 834.0, 2494.0, 8517.0, 41847.0, 3196542.0, 911211.0, 23474.0, 5477.0, 1776.0, 655.0, 270.0, 142.0, 82.0, 38.0, 37.0, 26.0, 16.0, 13.0, 10.0, 10.0, 5.0, 3.0, 4.0, 2.0, 3.0, 1.0, 0.0, 3.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.00194549560546875, -0.0018824338912963867, -0.0018193721771240234, -0.0017563104629516602, -0.0016932487487792969, -0.0016301870346069336, -0.0015671253204345703, -0.001504063606262207, -0.0014410018920898438, -0.0013779401779174805, -0.0013148784637451172, -0.001251816749572754, -0.0011887550354003906, -0.0011256933212280273, -0.001062631607055664, -0.0009995698928833008, -0.0009365081787109375, -0.0008734464645385742, -0.0008103847503662109, -0.0007473230361938477, -0.0006842613220214844, -0.0006211996078491211, -0.0005581378936767578, -0.0004950761795043945, -0.00043201446533203125, -0.00036895275115966797, -0.0003058910369873047, -0.0002428293228149414, -0.00017976760864257812, -0.00011670589447021484, -5.364418029785156e-05, 9.417533874511719e-06, 7.2479248046875e-05, 0.00013554096221923828, 0.00019860267639160156, 0.00026166439056396484, 0.0003247261047363281, 0.0003877878189086914, 0.0004508495330810547, 0.000513911247253418, 0.0005769729614257812, 0.0006400346755981445, 0.0007030963897705078, 0.0007661581039428711, 0.0008292198181152344, 0.0008922815322875977, 0.0009553432464599609, 0.0010184049606323242, 0.0010814666748046875, 0.0011445283889770508, 0.001207590103149414, 0.0012706518173217773, 0.0013337135314941406, 0.001396775245666504, 0.0014598369598388672, 0.0015228986740112305, 0.0015859603881835938, 0.001649022102355957, 0.0017120838165283203, 0.0017751455307006836, 0.0018382072448730469, 0.0019012689590454102, 0.0019643306732177734, 0.0020273923873901367, 0.0020904541015625]}, "gradients/model.layers.15.self_attn.q_proj.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 0.0, 1.0, 3.0, 1.0, 4.0, 4.0, 1.0, 5.0, 2.0, 11.0, 13.0, 30.0, 31.0, 44.0, 78.0, 130.0, 266.0, 576.0, 1414.0, 4633.0, 23569.0, 797001.0, 15827617.0, 106685.0, 10585.0, 2693.0, 921.0, 433.0, 195.0, 107.0, 55.0, 35.0, 23.0, 9.0, 7.0, 10.0, 10.0, 5.0, 2.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0022125244140625, -0.0021250247955322266, -0.002037525177001953, -0.0019500255584716797, -0.0018625259399414062, -0.0017750263214111328, -0.0016875267028808594, -0.001600027084350586, -0.0015125274658203125, -0.001425027847290039, -0.0013375282287597656, -0.0012500286102294922, -0.0011625289916992188, -0.0010750293731689453, -0.0009875297546386719, -0.0009000301361083984, -0.000812530517578125, -0.0007250308990478516, -0.0006375312805175781, -0.0005500316619873047, -0.00046253204345703125, -0.0003750324249267578, -0.0002875328063964844, -0.00020003318786621094, -0.0001125335693359375, -2.5033950805664062e-05, 6.246566772460938e-05, 0.0001499652862548828, 0.00023746490478515625, 0.0003249645233154297, 0.0004124641418457031, 0.0004999637603759766, 0.00058746337890625, 0.0006749629974365234, 0.0007624626159667969, 0.0008499622344970703, 0.0009374618530273438, 0.0010249614715576172, 0.0011124610900878906, 0.001199960708618164, 0.0012874603271484375, 0.001374959945678711, 0.0014624595642089844, 0.0015499591827392578, 0.0016374588012695312, 0.0017249584197998047, 0.0018124580383300781, 0.0018999576568603516, 0.001987457275390625, 0.0020749568939208984, 0.002162456512451172, 0.0022499561309814453, 0.0023374557495117188, 0.002424955368041992, 0.0025124549865722656, 0.002599954605102539, 0.0026874542236328125, 0.002774953842163086, 0.0028624534606933594, 0.002949953079223633, 0.0030374526977539062, 0.0031249523162841797, 0.003212451934814453, 0.0032999515533447266, 0.003387451171875]}, "gradients/model.layers.15.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 3.0, 4.0, 4.0, 7.0, 14.0, 32.0, 128.0, 393.0, 872.0, 1116.0, 901.0, 420.0, 112.0, 41.0, 16.0, 6.0, 4.0, 7.0, 1.0, 2.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.00125885009765625, -0.0012074708938598633, -0.0011560916900634766, -0.0011047124862670898, -0.0010533332824707031, -0.0010019540786743164, -0.0009505748748779297, -0.000899195671081543, -0.0008478164672851562, -0.0007964372634887695, -0.0007450580596923828, -0.0006936788558959961, -0.0006422996520996094, -0.0005909204483032227, -0.0005395412445068359, -0.0004881620407104492, -0.0004367828369140625, -0.0003854036331176758, -0.00033402442932128906, -0.00028264522552490234, -0.00023126602172851562, -0.0001798868179321289, -0.0001285076141357422, -7.712841033935547e-05, -2.574920654296875e-05, 2.562999725341797e-05, 7.700920104980469e-05, 0.0001283884048461914, 0.00017976760864257812, 0.00023114681243896484, 0.00028252601623535156, 0.0003339052200317383, 0.000385284423828125, 0.0004366636276245117, 0.00048804283142089844, 0.0005394220352172852, 0.0005908012390136719, 0.0006421804428100586, 0.0006935596466064453, 0.000744938850402832, 0.0007963180541992188, 0.0008476972579956055, 0.0008990764617919922, 0.0009504556655883789, 0.0010018348693847656, 0.0010532140731811523, 0.001104593276977539, 0.0011559724807739258, 0.0012073516845703125, 0.0012587308883666992, 0.001310110092163086, 0.0013614892959594727, 0.0014128684997558594, 0.001464247703552246, 0.0015156269073486328, 0.0015670061111450195, 0.0016183853149414062, 0.001669764518737793, 0.0017211437225341797, 0.0017725229263305664, 0.0018239021301269531, 0.0018752813339233398, 0.0019266605377197266, 0.0019780397415161133, 0.0020294189453125]}, "gradients/model.layers.14.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 2.0, 0.0, 0.0, 1.0, 6.0, 5.0, 3.0, 12.0, 8.0, 12.0, 31.0, 31.0, 53.0, 72.0, 101.0, 175.0, 356.0, 709.0, 1692.0, 4717.0, 17870.0, 138908.0, 7023059.0, 16777216.0, 2554032.0, 78577.0, 12862.0, 3657.0, 1357.0, 564.0, 280.0, 178.0, 96.0, 49.0, 36.0, 32.0, 23.0, 17.0, 3.0, 8.0, 1.0, 5.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0025177001953125, -0.0024144649505615234, -0.002311229705810547, -0.0022079944610595703, -0.0021047592163085938, -0.002001523971557617, -0.0018982887268066406, -0.001795053482055664, -0.0016918182373046875, -0.001588582992553711, -0.0014853477478027344, -0.0013821125030517578, -0.0012788772583007812, -0.0011756420135498047, -0.0010724067687988281, -0.0009691715240478516, -0.000865936279296875, -0.0007627010345458984, -0.0006594657897949219, -0.0005562305450439453, -0.00045299530029296875, -0.0003497600555419922, -0.0002465248107910156, -0.00014328956604003906, -4.00543212890625e-05, 6.318092346191406e-05, 0.00016641616821289062, 0.0002696514129638672, 0.00037288665771484375, 0.0004761219024658203, 0.0005793571472167969, 0.0006825923919677734, 0.00078582763671875, 0.0008890628814697266, 0.0009922981262207031, 0.0010955333709716797, 0.0011987686157226562, 0.0013020038604736328, 0.0014052391052246094, 0.001508474349975586, 0.0016117095947265625, 0.001714944839477539, 0.0018181800842285156, 0.0019214153289794922, 0.0020246505737304688, 0.0021278858184814453, 0.002231121063232422, 0.0023343563079833984, 0.002437591552734375, 0.0025408267974853516, 0.002644062042236328, 0.0027472972869873047, 0.0028505325317382812, 0.002953767776489258, 0.0030570030212402344, 0.003160238265991211, 0.0032634735107421875, 0.003366708755493164, 0.0034699440002441406, 0.003573179244995117, 0.0036764144897460938, 0.0037796497344970703, 0.003882884979248047, 0.0039861202239990234, 0.00408935546875]}, "gradients/model.layers.14.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 1.0, 4.0, 2.0, 4.0, 2.0, 13.0, 12.0, 21.0, 32.0, 60.0, 110.0, 177.0, 336.0, 666.0, 1552.0, 3783.0, 16985.0, 380570.0, 16777216.0, 2842774.0, 33434.0, 5494.0, 1917.0, 790.0, 444.0, 215.0, 131.0, 82.0, 43.0, 33.0, 10.0, 18.0, 10.0, 4.0, 5.0, 3.0, 3.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.00665283203125, -0.006480693817138672, -0.006308555603027344, -0.006136417388916016, -0.0059642791748046875, -0.005792140960693359, -0.005620002746582031, -0.005447864532470703, -0.005275726318359375, -0.005103588104248047, -0.004931449890136719, -0.004759311676025391, -0.0045871734619140625, -0.004415035247802734, -0.004242897033691406, -0.004070758819580078, -0.00389862060546875, -0.003726482391357422, -0.0035543441772460938, -0.0033822059631347656, -0.0032100677490234375, -0.0030379295349121094, -0.0028657913208007812, -0.002693653106689453, -0.002521514892578125, -0.002349376678466797, -0.0021772384643554688, -0.0020051002502441406, -0.0018329620361328125, -0.0016608238220214844, -0.0014886856079101562, -0.0013165473937988281, -0.0011444091796875, -0.0009722709655761719, -0.0008001327514648438, -0.0006279945373535156, -0.0004558563232421875, -0.0002837181091308594, -0.00011157989501953125, 6.0558319091796875e-05, 0.000232696533203125, 0.0004048347473144531, 0.0005769729614257812, 0.0007491111755371094, 0.0009212493896484375, 0.0010933876037597656, 0.0012655258178710938, 0.0014376640319824219, 0.00160980224609375, 0.0017819404602050781, 0.0019540786743164062, 0.0021262168884277344, 0.0022983551025390625, 0.0024704933166503906, 0.0026426315307617188, 0.002814769744873047, 0.002986907958984375, 0.003159046173095703, 0.0033311843872070312, 0.0035033226013183594, 0.0036754608154296875, 0.0038475990295410156, 0.004019737243652344, 0.004191875457763672, 0.004364013671875]}, "gradients/model.layers.14.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 2.0, 0.0, 2.0, 0.0, 3.0, 3.0, 15.0, 29.0, 74.0, 175.0, 345.0, 572.0, 766.0, 770.0, 599.0, 409.0, 188.0, 76.0, 30.0, 17.0, 4.0, 4.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0034332275390625, -0.003355264663696289, -0.003277301788330078, -0.003199338912963867, -0.0031213760375976562, -0.0030434131622314453, -0.0029654502868652344, -0.0028874874114990234, -0.0028095245361328125, -0.0027315616607666016, -0.0026535987854003906, -0.0025756359100341797, -0.0024976730346679688, -0.002419710159301758, -0.002341747283935547, -0.002263784408569336, -0.002185821533203125, -0.002107858657836914, -0.002029895782470703, -0.0019519329071044922, -0.0018739700317382812, -0.0017960071563720703, -0.0017180442810058594, -0.0016400814056396484, -0.0015621185302734375, -0.0014841556549072266, -0.0014061927795410156, -0.0013282299041748047, -0.0012502670288085938, -0.0011723041534423828, -0.0010943412780761719, -0.001016378402709961, -0.00093841552734375, -0.0008604526519775391, -0.0007824897766113281, -0.0007045269012451172, -0.0006265640258789062, -0.0005486011505126953, -0.0004706382751464844, -0.00039267539978027344, -0.0003147125244140625, -0.00023674964904785156, -0.00015878677368164062, -8.082389831542969e-05, -2.86102294921875e-06, 7.510185241699219e-05, 0.00015306472778320312, 0.00023102760314941406, 0.000308990478515625, 0.00038695335388183594, 0.0004649162292480469, 0.0005428791046142578, 0.0006208419799804688, 0.0006988048553466797, 0.0007767677307128906, 0.0008547306060791016, 0.0009326934814453125, 0.0010106563568115234, 0.0010886192321777344, 0.0011665821075439453, 0.0012445449829101562, 0.0013225078582763672, 0.0014004707336425781, 0.001478433609008789, 0.001556396484375]}, "gradients/model.layers.14.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 3.0, 1.0, 4.0, 1.0, 2.0, 7.0, 3.0, 5.0, 13.0, 11.0, 28.0, 28.0, 58.0, 94.0, 161.0, 282.0, 569.0, 1108.0, 2054.0, 4630.0, 10358.0, 29531.0, 119594.0, 727429.0, 4633793.0, 8583797.0, 2243873.0, 325292.0, 61653.0, 18657.0, 7329.0, 3270.0, 1721.0, 834.0, 459.0, 233.0, 102.0, 81.0, 61.0, 27.0, 9.0, 10.0, 15.0, 6.0, 7.0, 2.0, 2.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.003021240234375, -0.002924680709838867, -0.0028281211853027344, -0.0027315616607666016, -0.0026350021362304688, -0.002538442611694336, -0.002441883087158203, -0.0023453235626220703, -0.0022487640380859375, -0.0021522045135498047, -0.002055644989013672, -0.001959085464477539, -0.0018625259399414062, -0.0017659664154052734, -0.0016694068908691406, -0.0015728473663330078, -0.001476287841796875, -0.0013797283172607422, -0.0012831687927246094, -0.0011866092681884766, -0.0010900497436523438, -0.000993490219116211, -0.0008969306945800781, -0.0008003711700439453, -0.0007038116455078125, -0.0006072521209716797, -0.0005106925964355469, -0.00041413307189941406, -0.00031757354736328125, -0.00022101402282714844, -0.00012445449829101562, -2.7894973754882812e-05, 6.866455078125e-05, 0.0001652240753173828, 0.0002617835998535156, 0.00035834312438964844, 0.00045490264892578125, 0.0005514621734619141, 0.0006480216979980469, 0.0007445812225341797, 0.0008411407470703125, 0.0009377002716064453, 0.0010342597961425781, 0.001130819320678711, 0.0012273788452148438, 0.0013239383697509766, 0.0014204978942871094, 0.0015170574188232422, 0.001613616943359375, 0.0017101764678955078, 0.0018067359924316406, 0.0019032955169677734, 0.0019998550415039062, 0.002096414566040039, 0.002192974090576172, 0.0022895336151123047, 0.0023860931396484375, 0.0024826526641845703, 0.002579212188720703, 0.002675771713256836, 0.0027723312377929688, 0.0028688907623291016, 0.0029654502868652344, 0.003062009811401367, 0.0031585693359375]}, "gradients/model.layers.14.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 3.0, 4.0, 6.0, 7.0, 8.0, 18.0, 18.0, 28.0, 36.0, 58.0, 68.0, 113.0, 143.0, 164.0, 244.0, 261.0, 444.0, 559.0, 826.0, 1108.0, 1532.0, 2273.0, 3415.0, 7082.0, 36297.0, 824338.0, 2929079.0, 351282.0, 19536.0, 5394.0, 2974.0, 2019.0, 1331.0, 900.0, 705.0, 524.0, 393.0, 301.0, 189.0, 155.0, 130.0, 78.0, 83.0, 53.0, 36.0, 24.0, 19.0, 12.0, 10.0, 4.0, 3.0, 4.0, 4.0, 2.0, 0.0, 2.0, 1.0, 0.0, 1.0], "bins": [-0.006011962890625, -0.005820274353027344, -0.0056285858154296875, -0.005436897277832031, -0.005245208740234375, -0.005053520202636719, -0.0048618316650390625, -0.004670143127441406, -0.00447845458984375, -0.004286766052246094, -0.0040950775146484375, -0.0039033889770507812, -0.003711700439453125, -0.0035200119018554688, -0.0033283233642578125, -0.0031366348266601562, -0.0029449462890625, -0.0027532577514648438, -0.0025615692138671875, -0.0023698806762695312, -0.002178192138671875, -0.0019865036010742188, -0.0017948150634765625, -0.0016031265258789062, -0.00141143798828125, -0.0012197494506835938, -0.0010280609130859375, -0.0008363723754882812, -0.000644683837890625, -0.00045299530029296875, -0.0002613067626953125, -6.961822509765625e-05, 0.0001220703125, 0.00031375885009765625, 0.0005054473876953125, 0.0006971359252929688, 0.000888824462890625, 0.0010805130004882812, 0.0012722015380859375, 0.0014638900756835938, 0.00165557861328125, 0.0018472671508789062, 0.0020389556884765625, 0.0022306442260742188, 0.002422332763671875, 0.0026140213012695312, 0.0028057098388671875, 0.0029973983764648438, 0.0031890869140625, 0.0033807754516601562, 0.0035724639892578125, 0.0037641525268554688, 0.003955841064453125, 0.004147529602050781, 0.0043392181396484375, 0.004530906677246094, 0.00472259521484375, 0.004914283752441406, 0.0051059722900390625, 0.005297660827636719, 0.005489349365234375, 0.005681037902832031, 0.0058727264404296875, 0.006064414978027344, 0.006256103515625]}, "gradients/model.layers.14.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 3.0, 0.0, 3.0, 1.0, 3.0, 7.0, 11.0, 10.0, 4.0, 9.0, 9.0, 25.0, 25.0, 28.0, 39.0, 66.0, 84.0, 133.0, 270.0, 541.0, 1356.0, 4030.0, 17770.0, 743766.0, 3382928.0, 33923.0, 5956.0, 1744.0, 704.0, 336.0, 173.0, 103.0, 66.0, 36.0, 48.0, 20.0, 18.0, 9.0, 11.0, 6.0, 3.0, 9.0, 4.0, 1.0, 2.0, 3.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.002838134765625, -0.0027637481689453125, -0.002689361572265625, -0.0026149749755859375, -0.00254058837890625, -0.0024662017822265625, -0.002391815185546875, -0.0023174285888671875, -0.0022430419921875, -0.0021686553955078125, -0.002094268798828125, -0.0020198822021484375, -0.00194549560546875, -0.0018711090087890625, -0.001796722412109375, -0.0017223358154296875, -0.00164794921875, -0.0015735626220703125, -0.001499176025390625, -0.0014247894287109375, -0.00135040283203125, -0.0012760162353515625, -0.001201629638671875, -0.0011272430419921875, -0.0010528564453125, -0.0009784698486328125, -0.000904083251953125, -0.0008296966552734375, -0.00075531005859375, -0.0006809234619140625, -0.000606536865234375, -0.0005321502685546875, -0.000457763671875, -0.0003833770751953125, -0.000308990478515625, -0.0002346038818359375, -0.00016021728515625, -8.58306884765625e-05, -1.1444091796875e-05, 6.29425048828125e-05, 0.0001373291015625, 0.0002117156982421875, 0.000286102294921875, 0.0003604888916015625, 0.00043487548828125, 0.0005092620849609375, 0.000583648681640625, 0.0006580352783203125, 0.000732421875, 0.0008068084716796875, 0.000881195068359375, 0.0009555816650390625, 0.00102996826171875, 0.0011043548583984375, 0.001178741455078125, 0.0012531280517578125, 0.0013275146484375, 0.0014019012451171875, 0.001476287841796875, 0.0015506744384765625, 0.00162506103515625, 0.0016994476318359375, 0.001773834228515625, 0.0018482208251953125, 0.001922607421875]}, "gradients/model.layers.14.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 2.0, 1.0, 1.0, 0.0, 2.0, 0.0, 2.0, 5.0, 4.0, 4.0, 9.0, 9.0, 15.0, 15.0, 20.0, 26.0, 37.0, 65.0, 74.0, 119.0, 206.0, 403.0, 795.0, 2189.0, 7281.0, 46754.0, 14715106.0, 1969949.0, 25806.0, 5041.0, 1704.0, 659.0, 350.0, 176.0, 111.0, 82.0, 50.0, 36.0, 15.0, 20.0, 14.0, 18.0, 5.0, 7.0, 6.0, 3.0, 1.0, 5.0, 3.0, 1.0, 1.0, 1.0, 0.0, 2.0, 1.0, 0.0, 2.0], "bins": [-0.004974365234375, -0.0048274993896484375, -0.004680633544921875, -0.0045337677001953125, -0.00438690185546875, -0.0042400360107421875, -0.004093170166015625, -0.0039463043212890625, -0.0037994384765625, -0.0036525726318359375, -0.003505706787109375, -0.0033588409423828125, -0.00321197509765625, -0.0030651092529296875, -0.002918243408203125, -0.0027713775634765625, -0.00262451171875, -0.0024776458740234375, -0.002330780029296875, -0.0021839141845703125, -0.00203704833984375, -0.0018901824951171875, -0.001743316650390625, -0.0015964508056640625, -0.0014495849609375, -0.0013027191162109375, -0.001155853271484375, -0.0010089874267578125, -0.00086212158203125, -0.0007152557373046875, -0.000568389892578125, -0.0004215240478515625, -0.000274658203125, -0.0001277923583984375, 1.9073486328125e-05, 0.0001659393310546875, 0.00031280517578125, 0.0004596710205078125, 0.000606536865234375, 0.0007534027099609375, 0.0009002685546875, 0.0010471343994140625, 0.001194000244140625, 0.0013408660888671875, 0.00148773193359375, 0.0016345977783203125, 0.001781463623046875, 0.0019283294677734375, 0.0020751953125, 0.0022220611572265625, 0.002368927001953125, 0.0025157928466796875, 0.00266265869140625, 0.0028095245361328125, 0.002956390380859375, 0.0031032562255859375, 0.0032501220703125, 0.0033969879150390625, 0.003543853759765625, 0.0036907196044921875, 0.00383758544921875, 0.0039844512939453125, 0.004131317138671875, 0.0042781829833984375, 0.004425048828125]}, "gradients/model.layers.14.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 2.0, 3.0, 5.0, 9.0, 34.0, 91.0, 251.0, 551.0, 927.0, 960.0, 690.0, 338.0, 117.0, 44.0, 27.0, 9.0, 7.0, 5.0, 7.0, 4.0, 3.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0010223388671875, -0.0009779930114746094, -0.0009336471557617188, -0.0008893013000488281, -0.0008449554443359375, -0.0008006095886230469, -0.0007562637329101562, -0.0007119178771972656, -0.000667572021484375, -0.0006232261657714844, -0.0005788803100585938, -0.0005345344543457031, -0.0004901885986328125, -0.0004458427429199219, -0.00040149688720703125, -0.0003571510314941406, -0.00031280517578125, -0.0002684593200683594, -0.00022411346435546875, -0.00017976760864257812, -0.0001354217529296875, -9.107589721679688e-05, -4.673004150390625e-05, -2.384185791015625e-06, 4.1961669921875e-05, 8.630752563476562e-05, 0.00013065338134765625, 0.00017499923706054688, 0.0002193450927734375, 0.0002636909484863281, 0.00030803680419921875, 0.0003523826599121094, 0.000396728515625, 0.0004410743713378906, 0.00048542022705078125, 0.0005297660827636719, 0.0005741119384765625, 0.0006184577941894531, 0.0006628036499023438, 0.0007071495056152344, 0.000751495361328125, 0.0007958412170410156, 0.0008401870727539062, 0.0008845329284667969, 0.0009288787841796875, 0.0009732246398925781, 0.0010175704956054688, 0.0010619163513183594, 0.00110626220703125, 0.0011506080627441406, 0.0011949539184570312, 0.0012392997741699219, 0.0012836456298828125, 0.0013279914855957031, 0.0013723373413085938, 0.0014166831970214844, 0.001461029052734375, 0.0015053749084472656, 0.0015497207641601562, 0.0015940666198730469, 0.0016384124755859375, 0.0016827583312988281, 0.0017271041870117188, 0.0017714500427246094, 0.0018157958984375]}, "gradients/model.layers.13.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 9.0, 3.0, 6.0, 22.0, 21.0, 31.0, 38.0, 60.0, 84.0, 120.0, 157.0, 228.0, 393.0, 996.0, 6457.0, 313071.0, 16777216.0, 899509.0, 9083.0, 1217.0, 439.0, 221.0, 170.0, 140.0, 91.0, 66.0, 49.0, 39.0, 16.0, 6.0, 4.0, 4.0, 7.0, 2.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00494384765625, -0.004742622375488281, -0.0045413970947265625, -0.004340171813964844, -0.004138946533203125, -0.003937721252441406, -0.0037364959716796875, -0.0035352706909179688, -0.00333404541015625, -0.0031328201293945312, -0.0029315948486328125, -0.0027303695678710938, -0.002529144287109375, -0.0023279190063476562, -0.0021266937255859375, -0.0019254684448242188, -0.0017242431640625, -0.0015230178833007812, -0.0013217926025390625, -0.0011205673217773438, -0.000919342041015625, -0.0007181167602539062, -0.0005168914794921875, -0.00031566619873046875, -0.00011444091796875, 8.678436279296875e-05, 0.0002880096435546875, 0.0004892349243164062, 0.000690460205078125, 0.0008916854858398438, 0.0010929107666015625, 0.0012941360473632812, 0.001495361328125, 0.0016965866088867188, 0.0018978118896484375, 0.0020990371704101562, 0.002300262451171875, 0.0025014877319335938, 0.0027027130126953125, 0.0029039382934570312, 0.00310516357421875, 0.0033063888549804688, 0.0035076141357421875, 0.0037088394165039062, 0.003910064697265625, 0.004111289978027344, 0.0043125152587890625, 0.004513740539550781, 0.0047149658203125, 0.004916191101074219, 0.0051174163818359375, 0.005318641662597656, 0.005519866943359375, 0.005721092224121094, 0.0059223175048828125, 0.006123542785644531, 0.00632476806640625, 0.006525993347167969, 0.0067272186279296875, 0.006928443908691406, 0.007129669189453125, 0.007330894470214844, 0.0075321197509765625, 0.007733345031738281, 0.0079345703125]}, "gradients/model.layers.13.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 2.0, 5.0, 9.0, 18.0, 34.0, 73.0, 160.0, 449.0, 1237.0, 5814.0, 144324.0, 16777216.0, 113026.0, 5447.0, 1193.0, 407.0, 144.0, 76.0, 37.0, 10.0, 9.0, 2.0, 4.0, 1.0, 1.0, 2.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.005401611328125, -0.005109310150146484, -0.004817008972167969, -0.004524707794189453, -0.0042324066162109375, -0.003940105438232422, -0.0036478042602539062, -0.0033555030822753906, -0.003063201904296875, -0.0027709007263183594, -0.0024785995483398438, -0.002186298370361328, -0.0018939971923828125, -0.0016016960144042969, -0.0013093948364257812, -0.0010170936584472656, -0.00072479248046875, -0.0004324913024902344, -0.00014019012451171875, 0.00015211105346679688, 0.0004444122314453125, 0.0007367134094238281, 0.0010290145874023438, 0.0013213157653808594, 0.001613616943359375, 0.0019059181213378906, 0.0021982192993164062, 0.002490520477294922, 0.0027828216552734375, 0.003075122833251953, 0.0033674240112304688, 0.0036597251892089844, 0.0039520263671875, 0.004244327545166016, 0.004536628723144531, 0.004828929901123047, 0.0051212310791015625, 0.005413532257080078, 0.005705833435058594, 0.005998134613037109, 0.006290435791015625, 0.006582736968994141, 0.006875038146972656, 0.007167339324951172, 0.0074596405029296875, 0.007751941680908203, 0.008044242858886719, 0.008336544036865234, 0.00862884521484375, 0.008921146392822266, 0.009213447570800781, 0.009505748748779297, 0.009798049926757812, 0.010090351104736328, 0.010382652282714844, 0.01067495346069336, 0.010967254638671875, 0.01125955581665039, 0.011551856994628906, 0.011844158172607422, 0.012136459350585938, 0.012428760528564453, 0.012721061706542969, 0.013013362884521484, 0.0133056640625]}, "gradients/model.layers.13.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 10.0, 16.0, 24.0, 146.0, 430.0, 1023.0, 1262.0, 807.0, 297.0, 47.0, 14.0, 6.0, 5.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.005126953125, -0.004999637603759766, -0.004872322082519531, -0.004745006561279297, -0.0046176910400390625, -0.004490375518798828, -0.004363059997558594, -0.004235744476318359, -0.004108428955078125, -0.003981113433837891, -0.0038537979125976562, -0.003726482391357422, -0.0035991668701171875, -0.003471851348876953, -0.0033445358276367188, -0.0032172203063964844, -0.00308990478515625, -0.0029625892639160156, -0.0028352737426757812, -0.002707958221435547, -0.0025806427001953125, -0.002453327178955078, -0.0023260116577148438, -0.0021986961364746094, -0.002071380615234375, -0.0019440650939941406, -0.0018167495727539062, -0.0016894340515136719, -0.0015621185302734375, -0.0014348030090332031, -0.0013074874877929688, -0.0011801719665527344, -0.0010528564453125, -0.0009255409240722656, -0.0007982254028320312, -0.0006709098815917969, -0.0005435943603515625, -0.0004162788391113281, -0.00028896331787109375, -0.00016164779663085938, -3.4332275390625e-05, 9.298324584960938e-05, 0.00022029876708984375, 0.0003476142883300781, 0.0004749298095703125, 0.0006022453308105469, 0.0007295608520507812, 0.0008568763732910156, 0.00098419189453125, 0.0011115074157714844, 0.0012388229370117188, 0.0013661384582519531, 0.0014934539794921875, 0.0016207695007324219, 0.0017480850219726562, 0.0018754005432128906, 0.002002716064453125, 0.0021300315856933594, 0.0022573471069335938, 0.002384662628173828, 0.0025119781494140625, 0.002639293670654297, 0.0027666091918945312, 0.0028939247131347656, 0.003021240234375]}, "gradients/model.layers.13.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 2.0, 2.0, 1.0, 3.0, 5.0, 14.0, 13.0, 30.0, 47.0, 90.0, 173.0, 338.0, 772.0, 2002.0, 6585.0, 28394.0, 208294.0, 4967175.0, 10892247.0, 602446.0, 52796.0, 10716.0, 2983.0, 1144.0, 429.0, 254.0, 114.0, 59.0, 37.0, 16.0, 13.0, 4.0, 3.0, 4.0, 1.0, 3.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00439453125, -0.004220008850097656, -0.0040454864501953125, -0.0038709640502929688, -0.003696441650390625, -0.0035219192504882812, -0.0033473968505859375, -0.0031728744506835938, -0.00299835205078125, -0.0028238296508789062, -0.0026493072509765625, -0.0024747848510742188, -0.002300262451171875, -0.0021257400512695312, -0.0019512176513671875, -0.0017766952514648438, -0.0016021728515625, -0.0014276504516601562, -0.0012531280517578125, -0.0010786056518554688, -0.000904083251953125, -0.0007295608520507812, -0.0005550384521484375, -0.00038051605224609375, -0.00020599365234375, -3.147125244140625e-05, 0.0001430511474609375, 0.00031757354736328125, 0.000492095947265625, 0.0006666183471679688, 0.0008411407470703125, 0.0010156631469726562, 0.001190185546875, 0.0013647079467773438, 0.0015392303466796875, 0.0017137527465820312, 0.001888275146484375, 0.0020627975463867188, 0.0022373199462890625, 0.0024118423461914062, 0.00258636474609375, 0.0027608871459960938, 0.0029354095458984375, 0.0031099319458007812, 0.003284454345703125, 0.0034589767456054688, 0.0036334991455078125, 0.0038080215454101562, 0.0039825439453125, 0.004157066345214844, 0.0043315887451171875, 0.004506111145019531, 0.004680633544921875, 0.004855155944824219, 0.0050296783447265625, 0.005204200744628906, 0.00537872314453125, 0.005553245544433594, 0.0057277679443359375, 0.005902290344238281, 0.006076812744140625, 0.006251335144042969, 0.0064258575439453125, 0.006600379943847656, 0.00677490234375]}, "gradients/model.layers.13.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 3.0, 4.0, 7.0, 5.0, 14.0, 16.0, 12.0, 12.0, 25.0, 42.0, 79.0, 74.0, 108.0, 149.0, 205.0, 251.0, 413.0, 529.0, 787.0, 1112.0, 1615.0, 2558.0, 4208.0, 12785.0, 626327.0, 3360013.0, 164461.0, 7753.0, 3653.0, 2280.0, 1434.0, 951.0, 732.0, 475.0, 347.0, 215.0, 182.0, 130.0, 88.0, 75.0, 38.0, 36.0, 26.0, 30.0, 11.0, 6.0, 10.0, 7.0, 2.0, 3.0, 0.0, 0.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0084228515625, -0.008145332336425781, -0.007867813110351562, -0.007590293884277344, -0.007312774658203125, -0.007035255432128906, -0.0067577362060546875, -0.006480216979980469, -0.00620269775390625, -0.005925178527832031, -0.0056476593017578125, -0.005370140075683594, -0.005092620849609375, -0.004815101623535156, -0.0045375823974609375, -0.004260063171386719, -0.0039825439453125, -0.0037050247192382812, -0.0034275054931640625, -0.0031499862670898438, -0.002872467041015625, -0.0025949478149414062, -0.0023174285888671875, -0.0020399093627929688, -0.00176239013671875, -0.0014848709106445312, -0.0012073516845703125, -0.0009298324584960938, -0.000652313232421875, -0.00037479400634765625, -9.72747802734375e-05, 0.00018024444580078125, 0.000457763671875, 0.0007352828979492188, 0.0010128021240234375, 0.0012903213500976562, 0.001567840576171875, 0.0018453598022460938, 0.0021228790283203125, 0.0024003982543945312, 0.00267791748046875, 0.0029554367065429688, 0.0032329559326171875, 0.0035104751586914062, 0.003787994384765625, 0.004065513610839844, 0.0043430328369140625, 0.004620552062988281, 0.0048980712890625, 0.005175590515136719, 0.0054531097412109375, 0.005730628967285156, 0.006008148193359375, 0.006285667419433594, 0.0065631866455078125, 0.006840705871582031, 0.00711822509765625, 0.007395744323730469, 0.0076732635498046875, 0.007950782775878906, 0.008228302001953125, 0.008505821228027344, 0.008783340454101562, 0.009060859680175781, 0.00933837890625]}, "gradients/model.layers.13.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 2.0, 1.0, 1.0, 1.0, 0.0, 1.0, 4.0, 3.0, 3.0, 8.0, 4.0, 9.0, 13.0, 11.0, 7.0, 16.0, 17.0, 25.0, 29.0, 43.0, 56.0, 81.0, 168.0, 277.0, 741.0, 2015.0, 6720.0, 27273.0, 1570398.0, 2540827.0, 34014.0, 7635.0, 2275.0, 748.0, 341.0, 160.0, 93.0, 66.0, 46.0, 45.0, 22.0, 29.0, 9.0, 8.0, 8.0, 12.0, 6.0, 3.0, 4.0, 11.0, 2.0, 1.0, 2.0, 4.0, 1.0, 1.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.0021514892578125, -0.0020864009857177734, -0.002021312713623047, -0.0019562244415283203, -0.0018911361694335938, -0.0018260478973388672, -0.0017609596252441406, -0.001695871353149414, -0.0016307830810546875, -0.001565694808959961, -0.0015006065368652344, -0.0014355182647705078, -0.0013704299926757812, -0.0013053417205810547, -0.0012402534484863281, -0.0011751651763916016, -0.001110076904296875, -0.0010449886322021484, -0.0009799003601074219, -0.0009148120880126953, -0.0008497238159179688, -0.0007846355438232422, -0.0007195472717285156, -0.0006544589996337891, -0.0005893707275390625, -0.0005242824554443359, -0.0004591941833496094, -0.0003941059112548828, -0.00032901763916015625, -0.0002639293670654297, -0.00019884109497070312, -0.00013375282287597656, -6.866455078125e-05, -3.5762786865234375e-06, 6.151199340820312e-05, 0.0001266002655029297, 0.00019168853759765625, 0.0002567768096923828, 0.0003218650817871094, 0.00038695335388183594, 0.0004520416259765625, 0.0005171298980712891, 0.0005822181701660156, 0.0006473064422607422, 0.0007123947143554688, 0.0007774829864501953, 0.0008425712585449219, 0.0009076595306396484, 0.000972747802734375, 0.0010378360748291016, 0.0011029243469238281, 0.0011680126190185547, 0.0012331008911132812, 0.0012981891632080078, 0.0013632774353027344, 0.001428365707397461, 0.0014934539794921875, 0.001558542251586914, 0.0016236305236816406, 0.0016887187957763672, 0.0017538070678710938, 0.0018188953399658203, 0.0018839836120605469, 0.0019490718841552734, 0.00201416015625]}, "gradients/model.layers.13.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 4.0, 2.0, 4.0, 5.0, 8.0, 12.0, 19.0, 13.0, 11.0, 16.0, 28.0, 46.0, 78.0, 98.0, 109.0, 175.0, 306.0, 433.0, 825.0, 1619.0, 3890.0, 12462.0, 84847.0, 8807493.0, 7764945.0, 79976.0, 11979.0, 3825.0, 1720.0, 841.0, 482.0, 302.0, 229.0, 114.0, 70.0, 49.0, 48.0, 39.0, 20.0, 17.0, 11.0, 8.0, 13.0, 3.0, 2.0, 4.0, 7.0, 2.0, 1.0, 0.0, 1.0, 2.0], "bins": [-0.0030975341796875, -0.003011465072631836, -0.002925395965576172, -0.002839326858520508, -0.0027532577514648438, -0.0026671886444091797, -0.0025811195373535156, -0.0024950504302978516, -0.0024089813232421875, -0.0023229122161865234, -0.0022368431091308594, -0.0021507740020751953, -0.0020647048950195312, -0.001978635787963867, -0.0018925666809082031, -0.001806497573852539, -0.001720428466796875, -0.001634359359741211, -0.0015482902526855469, -0.0014622211456298828, -0.0013761520385742188, -0.0012900829315185547, -0.0012040138244628906, -0.0011179447174072266, -0.0010318756103515625, -0.0009458065032958984, -0.0008597373962402344, -0.0007736682891845703, -0.0006875991821289062, -0.0006015300750732422, -0.0005154609680175781, -0.00042939186096191406, -0.00034332275390625, -0.00025725364685058594, -0.00017118453979492188, -8.511543273925781e-05, 9.5367431640625e-07, 8.702278137207031e-05, 0.00017309188842773438, 0.00025916099548339844, 0.0003452301025390625, 0.00043129920959472656, 0.0005173683166503906, 0.0006034374237060547, 0.0006895065307617188, 0.0007755756378173828, 0.0008616447448730469, 0.0009477138519287109, 0.001033782958984375, 0.001119852066040039, 0.0012059211730957031, 0.0012919902801513672, 0.0013780593872070312, 0.0014641284942626953, 0.0015501976013183594, 0.0016362667083740234, 0.0017223358154296875, 0.0018084049224853516, 0.0018944740295410156, 0.0019805431365966797, 0.0020666122436523438, 0.002152681350708008, 0.002238750457763672, 0.002324819564819336, 0.002410888671875]}, "gradients/model.layers.13.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 4.0, 1.0, 2.0, 6.0, 9.0, 51.0, 309.0, 891.0, 1328.0, 1031.0, 340.0, 86.0, 15.0, 6.0, 5.0, 4.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00147247314453125, -0.0014072656631469727, -0.0013420581817626953, -0.001276850700378418, -0.0012116432189941406, -0.0011464357376098633, -0.001081228256225586, -0.0010160207748413086, -0.0009508132934570312, -0.0008856058120727539, -0.0008203983306884766, -0.0007551908493041992, -0.0006899833679199219, -0.0006247758865356445, -0.0005595684051513672, -0.0004943609237670898, -0.0004291534423828125, -0.00036394596099853516, -0.0002987384796142578, -0.00023353099822998047, -0.00016832351684570312, -0.00010311603546142578, -3.790855407714844e-05, 2.7298927307128906e-05, 9.250640869140625e-05, 0.0001577138900756836, 0.00022292137145996094, 0.0002881288528442383, 0.0003533363342285156, 0.00041854381561279297, 0.0004837512969970703, 0.0005489587783813477, 0.000614166259765625, 0.0006793737411499023, 0.0007445812225341797, 0.000809788703918457, 0.0008749961853027344, 0.0009402036666870117, 0.001005411148071289, 0.0010706186294555664, 0.0011358261108398438, 0.001201033592224121, 0.0012662410736083984, 0.0013314485549926758, 0.0013966560363769531, 0.0014618635177612305, 0.0015270709991455078, 0.0015922784805297852, 0.0016574859619140625, 0.0017226934432983398, 0.0017879009246826172, 0.0018531084060668945, 0.0019183158874511719, 0.0019835233688354492, 0.0020487308502197266, 0.002113938331604004, 0.0021791458129882812, 0.0022443532943725586, 0.002309560775756836, 0.0023747682571411133, 0.0024399757385253906, 0.002505183219909668, 0.0025703907012939453, 0.0026355981826782227, 0.0027008056640625]}, "gradients/model.layers.12.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 3.0, 5.0, 13.0, 13.0, 19.0, 27.0, 44.0, 70.0, 94.0, 142.0, 215.0, 270.0, 466.0, 2219.0, 60532.0, 16777216.0, 7316151.0, 12526.0, 1056.0, 392.0, 241.0, 150.0, 120.0, 95.0, 64.0, 38.0, 32.0, 29.0, 11.0, 8.0, 4.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.008056640625, -0.007859230041503906, -0.0076618194580078125, -0.007464408874511719, -0.007266998291015625, -0.007069587707519531, -0.0068721771240234375, -0.006674766540527344, -0.00647735595703125, -0.006279945373535156, -0.0060825347900390625, -0.005885124206542969, -0.005687713623046875, -0.005490303039550781, -0.0052928924560546875, -0.005095481872558594, -0.0048980712890625, -0.004700660705566406, -0.0045032501220703125, -0.004305839538574219, -0.004108428955078125, -0.003911018371582031, -0.0037136077880859375, -0.0035161972045898438, -0.00331878662109375, -0.0031213760375976562, -0.0029239654541015625, -0.0027265548706054688, -0.002529144287109375, -0.0023317337036132812, -0.0021343231201171875, -0.0019369125366210938, -0.001739501953125, -0.0015420913696289062, -0.0013446807861328125, -0.0011472702026367188, -0.000949859619140625, -0.0007524490356445312, -0.0005550384521484375, -0.00035762786865234375, -0.00016021728515625, 3.719329833984375e-05, 0.0002346038818359375, 0.00043201446533203125, 0.000629425048828125, 0.0008268356323242188, 0.0010242462158203125, 0.0012216567993164062, 0.0014190673828125, 0.0016164779663085938, 0.0018138885498046875, 0.0020112991333007812, 0.002208709716796875, 0.0024061203002929688, 0.0026035308837890625, 0.0028009414672851562, 0.00299835205078125, 0.0031957626342773438, 0.0033931732177734375, 0.0035905838012695312, 0.003787994384765625, 0.003985404968261719, 0.0041828155517578125, 0.004380226135253906, 0.00457763671875]}, "gradients/model.layers.12.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 3.0, 5.0, 2.0, 2.0, 5.0, 9.0, 10.0, 23.0, 31.0, 38.0, 84.0, 106.0, 155.0, 275.0, 503.0, 925.0, 1889.0, 4832.0, 19798.0, 205890.0, 16777216.0, 16777216.0, 106191.0, 14323.0, 3975.0, 1505.0, 792.0, 443.0, 260.0, 149.0, 98.0, 63.0, 39.0, 32.0, 18.0, 11.0, 6.0, 7.0, 0.0, 8.0, 2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.004119873046875, -0.003982067108154297, -0.0038442611694335938, -0.0037064552307128906, -0.0035686492919921875, -0.0034308433532714844, -0.0032930374145507812, -0.003155231475830078, -0.003017425537109375, -0.002879619598388672, -0.0027418136596679688, -0.0026040077209472656, -0.0024662017822265625, -0.0023283958435058594, -0.0021905899047851562, -0.002052783966064453, -0.00191497802734375, -0.0017771720886230469, -0.0016393661499023438, -0.0015015602111816406, -0.0013637542724609375, -0.0012259483337402344, -0.0010881423950195312, -0.0009503364562988281, -0.000812530517578125, -0.0006747245788574219, -0.0005369186401367188, -0.0003991127014160156, -0.0002613067626953125, -0.00012350082397460938, 1.430511474609375e-05, 0.00015211105346679688, 0.0002899169921875, 0.0004277229309082031, 0.0005655288696289062, 0.0007033348083496094, 0.0008411407470703125, 0.0009789466857910156, 0.0011167526245117188, 0.0012545585632324219, 0.001392364501953125, 0.0015301704406738281, 0.0016679763793945312, 0.0018057823181152344, 0.0019435882568359375, 0.0020813941955566406, 0.0022192001342773438, 0.002357006072998047, 0.00249481201171875, 0.002632617950439453, 0.0027704238891601562, 0.0029082298278808594, 0.0030460357666015625, 0.0031838417053222656, 0.0033216476440429688, 0.003459453582763672, 0.003597259521484375, 0.003735065460205078, 0.0038728713989257812, 0.004010677337646484, 0.0041484832763671875, 0.004286289215087891, 0.004424095153808594, 0.004561901092529297, 0.00469970703125]}, "gradients/model.layers.12.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 3.0, 5.0, 1.0, 3.0, 10.0, 24.0, 78.0, 244.0, 655.0, 1000.0, 1019.0, 664.0, 275.0, 80.0, 14.0, 2.0, 2.0, 2.0, 1.0, 0.0, 1.0, 1.0, 0.0, 3.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.005035400390625, -0.004923343658447266, -0.004811286926269531, -0.004699230194091797, -0.0045871734619140625, -0.004475116729736328, -0.004363059997558594, -0.004251003265380859, -0.004138946533203125, -0.004026889801025391, -0.003914833068847656, -0.003802776336669922, -0.0036907196044921875, -0.003578662872314453, -0.0034666061401367188, -0.0033545494079589844, -0.00324249267578125, -0.0031304359436035156, -0.0030183792114257812, -0.002906322479248047, -0.0027942657470703125, -0.002682209014892578, -0.0025701522827148438, -0.0024580955505371094, -0.002346038818359375, -0.0022339820861816406, -0.0021219253540039062, -0.002009868621826172, -0.0018978118896484375, -0.0017857551574707031, -0.0016736984252929688, -0.0015616416931152344, -0.0014495849609375, -0.0013375282287597656, -0.0012254714965820312, -0.0011134147644042969, -0.0010013580322265625, -0.0008893013000488281, -0.0007772445678710938, -0.0006651878356933594, -0.000553131103515625, -0.0004410743713378906, -0.00032901763916015625, -0.00021696090698242188, -0.0001049041748046875, 7.152557373046875e-06, 0.00011920928955078125, 0.00023126602172851562, 0.00034332275390625, 0.0004553794860839844, 0.0005674362182617188, 0.0006794929504394531, 0.0007915496826171875, 0.0009036064147949219, 0.0010156631469726562, 0.0011277198791503906, 0.001239776611328125, 0.0013518333435058594, 0.0014638900756835938, 0.0015759468078613281, 0.0016880035400390625, 0.0018000602722167969, 0.0019121170043945312, 0.0020241737365722656, 0.00213623046875]}, "gradients/model.layers.12.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 3.0, 3.0, 3.0, 11.0, 3.0, 16.0, 31.0, 51.0, 107.0, 135.0, 241.0, 364.0, 814.0, 2029.0, 8986.0, 74565.0, 3385237.0, 12834696.0, 438931.0, 24233.0, 4182.0, 1188.0, 577.0, 325.0, 196.0, 104.0, 75.0, 47.0, 26.0, 12.0, 9.0, 3.0, 2.0, 3.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.00885009765625, -0.00864553451538086, -0.008440971374511719, -0.008236408233642578, -0.008031845092773438, -0.007827281951904297, -0.007622718811035156, -0.007418155670166016, -0.007213592529296875, -0.007009029388427734, -0.006804466247558594, -0.006599903106689453, -0.0063953399658203125, -0.006190776824951172, -0.005986213684082031, -0.005781650543212891, -0.00557708740234375, -0.005372524261474609, -0.005167961120605469, -0.004963397979736328, -0.0047588348388671875, -0.004554271697998047, -0.004349708557128906, -0.004145145416259766, -0.003940582275390625, -0.0037360191345214844, -0.0035314559936523438, -0.003326892852783203, -0.0031223297119140625, -0.002917766571044922, -0.0027132034301757812, -0.0025086402893066406, -0.0023040771484375, -0.0020995140075683594, -0.0018949508666992188, -0.0016903877258300781, -0.0014858245849609375, -0.0012812614440917969, -0.0010766983032226562, -0.0008721351623535156, -0.000667572021484375, -0.0004630088806152344, -0.00025844573974609375, -5.3882598876953125e-05, 0.0001506805419921875, 0.0003552436828613281, 0.0005598068237304688, 0.0007643699645996094, 0.00096893310546875, 0.0011734962463378906, 0.0013780593872070312, 0.0015826225280761719, 0.0017871856689453125, 0.001991748809814453, 0.0021963119506835938, 0.0024008750915527344, 0.002605438232421875, 0.0028100013732910156, 0.0030145645141601562, 0.003219127655029297, 0.0034236907958984375, 0.003628253936767578, 0.0038328170776367188, 0.004037380218505859, 0.004241943359375]}, "gradients/model.layers.12.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 5.0, 7.0, 3.0, 5.0, 14.0, 21.0, 15.0, 27.0, 34.0, 42.0, 84.0, 112.0, 134.0, 237.0, 349.0, 558.0, 998.0, 1774.0, 3497.0, 9312.0, 816991.0, 3313377.0, 36031.0, 4952.0, 2338.0, 1256.0, 753.0, 463.0, 277.0, 202.0, 133.0, 76.0, 73.0, 36.0, 30.0, 22.0, 8.0, 16.0, 8.0, 6.0, 8.0, 4.0, 0.0, 1.0, 3.0, 3.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.010986328125, -0.010634422302246094, -0.010282516479492188, -0.009930610656738281, -0.009578704833984375, -0.009226799011230469, -0.008874893188476562, -0.008522987365722656, -0.00817108154296875, -0.007819175720214844, -0.0074672698974609375, -0.007115364074707031, -0.006763458251953125, -0.006411552429199219, -0.0060596466064453125, -0.005707740783691406, -0.0053558349609375, -0.005003929138183594, -0.0046520233154296875, -0.004300117492675781, -0.003948211669921875, -0.0035963058471679688, -0.0032444000244140625, -0.0028924942016601562, -0.00254058837890625, -0.0021886825561523438, -0.0018367767333984375, -0.0014848709106445312, -0.001132965087890625, -0.0007810592651367188, -0.0004291534423828125, -7.724761962890625e-05, 0.000274658203125, 0.0006265640258789062, 0.0009784698486328125, 0.0013303756713867188, 0.001682281494140625, 0.0020341873168945312, 0.0023860931396484375, 0.0027379989624023438, 0.00308990478515625, 0.0034418106079101562, 0.0037937164306640625, 0.004145622253417969, 0.004497528076171875, 0.004849433898925781, 0.0052013397216796875, 0.005553245544433594, 0.0059051513671875, 0.006257057189941406, 0.0066089630126953125, 0.006960868835449219, 0.007312774658203125, 0.007664680480957031, 0.008016586303710938, 0.008368492126464844, 0.00872039794921875, 0.009072303771972656, 0.009424209594726562, 0.009776115417480469, 0.010128021240234375, 0.010479927062988281, 0.010831832885742188, 0.011183738708496094, 0.01153564453125]}, "gradients/model.layers.12.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 3.0, 0.0, 2.0, 1.0, 2.0, 3.0, 2.0, 2.0, 5.0, 8.0, 10.0, 9.0, 12.0, 13.0, 23.0, 33.0, 86.0, 147.0, 398.0, 1624.0, 6606.0, 69077.0, 4074481.0, 34800.0, 5067.0, 1254.0, 333.0, 114.0, 55.0, 30.0, 30.0, 17.0, 12.0, 14.0, 8.0, 6.0, 2.0, 1.0, 1.0, 1.0, 2.0, 2.0, 3.0, 0.0, 1.0, 0.0, 2.0], "bins": [-0.004302978515625, -0.004193782806396484, -0.004084587097167969, -0.003975391387939453, -0.0038661956787109375, -0.003756999969482422, -0.0036478042602539062, -0.0035386085510253906, -0.003429412841796875, -0.0033202171325683594, -0.0032110214233398438, -0.003101825714111328, -0.0029926300048828125, -0.002883434295654297, -0.0027742385864257812, -0.0026650428771972656, -0.00255584716796875, -0.0024466514587402344, -0.0023374557495117188, -0.002228260040283203, -0.0021190643310546875, -0.002009868621826172, -0.0019006729125976562, -0.0017914772033691406, -0.001682281494140625, -0.0015730857849121094, -0.0014638900756835938, -0.0013546943664550781, -0.0012454986572265625, -0.0011363029479980469, -0.0010271072387695312, -0.0009179115295410156, -0.0008087158203125, -0.0006995201110839844, -0.0005903244018554688, -0.0004811286926269531, -0.0003719329833984375, -0.0002627372741699219, -0.00015354156494140625, -4.4345855712890625e-05, 6.4849853515625e-05, 0.00017404556274414062, 0.00028324127197265625, 0.0003924369812011719, 0.0005016326904296875, 0.0006108283996582031, 0.0007200241088867188, 0.0008292198181152344, 0.00093841552734375, 0.0010476112365722656, 0.0011568069458007812, 0.0012660026550292969, 0.0013751983642578125, 0.0014843940734863281, 0.0015935897827148438, 0.0017027854919433594, 0.001811981201171875, 0.0019211769104003906, 0.0020303726196289062, 0.002139568328857422, 0.0022487640380859375, 0.002357959747314453, 0.0024671554565429688, 0.0025763511657714844, 0.002685546875]}, "gradients/model.layers.12.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 1.0, 3.0, 2.0, 2.0, 2.0, 0.0, 2.0, 7.0, 10.0, 10.0, 11.0, 19.0, 19.0, 32.0, 61.0, 67.0, 149.0, 263.0, 593.0, 1933.0, 10950.0, 187672.0, 16492109.0, 74020.0, 6846.0, 1350.0, 468.0, 225.0, 136.0, 70.0, 56.0, 32.0, 20.0, 17.0, 19.0, 9.0, 4.0, 7.0, 2.0, 2.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.004791259765625, -0.004647731781005859, -0.004504203796386719, -0.004360675811767578, -0.0042171478271484375, -0.004073619842529297, -0.003930091857910156, -0.0037865638732910156, -0.003643035888671875, -0.0034995079040527344, -0.0033559799194335938, -0.003212451934814453, -0.0030689239501953125, -0.002925395965576172, -0.0027818679809570312, -0.0026383399963378906, -0.00249481201171875, -0.0023512840270996094, -0.0022077560424804688, -0.002064228057861328, -0.0019207000732421875, -0.0017771720886230469, -0.0016336441040039062, -0.0014901161193847656, -0.001346588134765625, -0.0012030601501464844, -0.0010595321655273438, -0.0009160041809082031, -0.0007724761962890625, -0.0006289482116699219, -0.00048542022705078125, -0.0003418922424316406, -0.0001983642578125, -5.4836273193359375e-05, 8.869171142578125e-05, 0.00023221969604492188, 0.0003757476806640625, 0.0005192756652832031, 0.0006628036499023438, 0.0008063316345214844, 0.000949859619140625, 0.0010933876037597656, 0.0012369155883789062, 0.0013804435729980469, 0.0015239715576171875, 0.0016674995422363281, 0.0018110275268554688, 0.0019545555114746094, 0.00209808349609375, 0.0022416114807128906, 0.0023851394653320312, 0.002528667449951172, 0.0026721954345703125, 0.002815723419189453, 0.0029592514038085938, 0.0031027793884277344, 0.003246307373046875, 0.0033898353576660156, 0.0035333633422851562, 0.003676891326904297, 0.0038204193115234375, 0.003963947296142578, 0.004107475280761719, 0.004251003265380859, 0.00439453125]}, "gradients/model.layers.12.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 1.0, 1.0, 3.0, 5.0, 18.0, 77.0, 357.0, 1033.0, 1411.0, 846.0, 271.0, 48.0, 7.0, 5.0, 2.0, 4.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0034637451171875, -0.0033922195434570312, -0.0033206939697265625, -0.0032491683959960938, -0.003177642822265625, -0.0031061172485351562, -0.0030345916748046875, -0.0029630661010742188, -0.00289154052734375, -0.0028200149536132812, -0.0027484893798828125, -0.0026769638061523438, -0.002605438232421875, -0.0025339126586914062, -0.0024623870849609375, -0.0023908615112304688, -0.0023193359375, -0.0022478103637695312, -0.0021762847900390625, -0.0021047592163085938, -0.002033233642578125, -0.0019617080688476562, -0.0018901824951171875, -0.0018186569213867188, -0.00174713134765625, -0.0016756057739257812, -0.0016040802001953125, -0.0015325546264648438, -0.001461029052734375, -0.0013895034790039062, -0.0013179779052734375, -0.0012464523315429688, -0.0011749267578125, -0.0011034011840820312, -0.0010318756103515625, -0.0009603500366210938, -0.000888824462890625, -0.0008172988891601562, -0.0007457733154296875, -0.0006742477416992188, -0.00060272216796875, -0.0005311965942382812, -0.0004596710205078125, -0.00038814544677734375, -0.000316619873046875, -0.00024509429931640625, -0.0001735687255859375, -0.00010204315185546875, -3.0517578125e-05, 4.100799560546875e-05, 0.0001125335693359375, 0.00018405914306640625, 0.000255584716796875, 0.00032711029052734375, 0.0003986358642578125, 0.00047016143798828125, 0.00054168701171875, 0.0006132125854492188, 0.0006847381591796875, 0.0007562637329101562, 0.000827789306640625, 0.0008993148803710938, 0.0009708404541015625, 0.0010423660278320312, 0.0011138916015625]}, "gradients/model.layers.11.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0, 4.0, 5.0, 9.0, 14.0, 24.0, 53.0, 73.0, 89.0, 163.0, 235.0, 353.0, 656.0, 2388.0, 660944.0, 16777216.0, 43015.0, 1299.0, 544.0, 348.0, 185.0, 142.0, 105.0, 67.0, 36.0, 25.0, 14.0, 5.0, 11.0, 1.0, 3.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.01068115234375, -0.010428905487060547, -0.010176658630371094, -0.00992441177368164, -0.009672164916992188, -0.009419918060302734, -0.009167671203613281, -0.008915424346923828, -0.008663177490234375, -0.008410930633544922, -0.008158683776855469, -0.007906436920166016, -0.0076541900634765625, -0.007401943206787109, -0.007149696350097656, -0.006897449493408203, -0.00664520263671875, -0.006392955780029297, -0.006140708923339844, -0.005888462066650391, -0.0056362152099609375, -0.005383968353271484, -0.005131721496582031, -0.004879474639892578, -0.004627227783203125, -0.004374980926513672, -0.004122734069824219, -0.0038704872131347656, -0.0036182403564453125, -0.0033659934997558594, -0.0031137466430664062, -0.002861499786376953, -0.0026092529296875, -0.002357006072998047, -0.0021047592163085938, -0.0018525123596191406, -0.0016002655029296875, -0.0013480186462402344, -0.0010957717895507812, -0.0008435249328613281, -0.000591278076171875, -0.0003390312194824219, -8.678436279296875e-05, 0.00016546249389648438, 0.0004177093505859375, 0.0006699562072753906, 0.0009222030639648438, 0.0011744499206542969, 0.00142669677734375, 0.0016789436340332031, 0.0019311904907226562, 0.0021834373474121094, 0.0024356842041015625, 0.0026879310607910156, 0.0029401779174804688, 0.003192424774169922, 0.003444671630859375, 0.003696918487548828, 0.003949165344238281, 0.004201412200927734, 0.0044536590576171875, 0.004705905914306641, 0.004958152770996094, 0.005210399627685547, 0.005462646484375]}, "gradients/model.layers.11.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 4.0, 1.0, 0.0, 0.0, 1.0, 5.0, 5.0, 4.0, 15.0, 31.0, 47.0, 82.0, 131.0, 314.0, 616.0, 1446.0, 4484.0, 24648.0, 1767185.0, 16777216.0, 231915.0, 14644.0, 3189.0, 1100.0, 455.0, 233.0, 114.0, 60.0, 26.0, 20.0, 19.0, 11.0, 4.0, 3.0, 2.0, 3.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0062255859375, -0.0060443878173828125, -0.005863189697265625, -0.0056819915771484375, -0.00550079345703125, -0.0053195953369140625, -0.005138397216796875, -0.0049571990966796875, -0.0047760009765625, -0.0045948028564453125, -0.004413604736328125, -0.0042324066162109375, -0.00405120849609375, -0.0038700103759765625, -0.003688812255859375, -0.0035076141357421875, -0.003326416015625, -0.0031452178955078125, -0.002964019775390625, -0.0027828216552734375, -0.00260162353515625, -0.0024204254150390625, -0.002239227294921875, -0.0020580291748046875, -0.0018768310546875, -0.0016956329345703125, -0.001514434814453125, -0.0013332366943359375, -0.00115203857421875, -0.0009708404541015625, -0.000789642333984375, -0.0006084442138671875, -0.00042724609375, -0.0002460479736328125, -6.4849853515625e-05, 0.0001163482666015625, 0.00029754638671875, 0.0004787445068359375, 0.000659942626953125, 0.0008411407470703125, 0.0010223388671875, 0.0012035369873046875, 0.001384735107421875, 0.0015659332275390625, 0.00174713134765625, 0.0019283294677734375, 0.002109527587890625, 0.0022907257080078125, 0.002471923828125, 0.0026531219482421875, 0.002834320068359375, 0.0030155181884765625, 0.00319671630859375, 0.0033779144287109375, 0.003559112548828125, 0.0037403106689453125, 0.0039215087890625, 0.0041027069091796875, 0.004283905029296875, 0.0044651031494140625, 0.00464630126953125, 0.0048274993896484375, 0.005008697509765625, 0.0051898956298828125, 0.00537109375]}, "gradients/model.layers.11.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 2.0, 2.0, 2.0, 2.0, 7.0, 25.0, 36.0, 103.0, 208.0, 456.0, 602.0, 750.0, 746.0, 551.0, 325.0, 161.0, 54.0, 20.0, 17.0, 6.0, 4.0, 1.0, 3.0, 1.0, 2.0, 0.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0014190673828125, -0.0013432502746582031, -0.0012674331665039062, -0.0011916160583496094, -0.0011157989501953125, -0.0010399818420410156, -0.0009641647338867188, -0.0008883476257324219, -0.000812530517578125, -0.0007367134094238281, -0.0006608963012695312, -0.0005850791931152344, -0.0005092620849609375, -0.0004334449768066406, -0.00035762786865234375, -0.0002818107604980469, -0.00020599365234375, -0.00013017654418945312, -5.435943603515625e-05, 2.1457672119140625e-05, 9.72747802734375e-05, 0.00017309188842773438, 0.00024890899658203125, 0.0003247261047363281, 0.000400543212890625, 0.0004763603210449219, 0.0005521774291992188, 0.0006279945373535156, 0.0007038116455078125, 0.0007796287536621094, 0.0008554458618164062, 0.0009312629699707031, 0.001007080078125, 0.0010828971862792969, 0.0011587142944335938, 0.0012345314025878906, 0.0013103485107421875, 0.0013861656188964844, 0.0014619827270507812, 0.0015377998352050781, 0.001613616943359375, 0.0016894340515136719, 0.0017652511596679688, 0.0018410682678222656, 0.0019168853759765625, 0.0019927024841308594, 0.0020685195922851562, 0.002144336700439453, 0.00222015380859375, 0.002295970916748047, 0.0023717880249023438, 0.0024476051330566406, 0.0025234222412109375, 0.0025992393493652344, 0.0026750564575195312, 0.002750873565673828, 0.002826690673828125, 0.002902507781982422, 0.0029783248901367188, 0.0030541419982910156, 0.0031299591064453125, 0.0032057762145996094, 0.0032815933227539062, 0.003357410430908203, 0.0034332275390625]}, "gradients/model.layers.11.self_attn.o_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 4.0, 5.0, 2.0, 7.0, 5.0, 7.0, 10.0, 11.0, 20.0, 15.0, 20.0, 49.0, 63.0, 127.0, 218.0, 388.0, 660.0, 1222.0, 2393.0, 4763.0, 9243.0, 20258.0, 54172.0, 229266.0, 1642038.0, 8194246.0, 5557902.0, 856064.0, 133070.0, 38944.0, 15942.0, 7928.0, 3695.0, 2104.0, 1043.0, 583.0, 304.0, 161.0, 98.0, 64.0, 30.0, 23.0, 9.0, 9.0, 7.0, 5.0, 4.0, 3.0, 2.0, 0.0, 3.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.00299072265625, -0.002899646759033203, -0.0028085708618164062, -0.0027174949645996094, -0.0026264190673828125, -0.0025353431701660156, -0.0024442672729492188, -0.002353191375732422, -0.002262115478515625, -0.002171039581298828, -0.0020799636840820312, -0.0019888877868652344, -0.0018978118896484375, -0.0018067359924316406, -0.0017156600952148438, -0.0016245841979980469, -0.00153350830078125, -0.0014424324035644531, -0.0013513565063476562, -0.0012602806091308594, -0.0011692047119140625, -0.0010781288146972656, -0.0009870529174804688, -0.0008959770202636719, -0.000804901123046875, -0.0007138252258300781, -0.0006227493286132812, -0.0005316734313964844, -0.0004405975341796875, -0.0003495216369628906, -0.00025844573974609375, -0.00016736984252929688, -7.62939453125e-05, 1.4781951904296875e-05, 0.00010585784912109375, 0.00019693374633789062, 0.0002880096435546875, 0.0003790855407714844, 0.00047016143798828125, 0.0005612373352050781, 0.000652313232421875, 0.0007433891296386719, 0.0008344650268554688, 0.0009255409240722656, 0.0010166168212890625, 0.0011076927185058594, 0.0011987686157226562, 0.0012898445129394531, 0.00138092041015625, 0.0014719963073730469, 0.0015630722045898438, 0.0016541481018066406, 0.0017452239990234375, 0.0018362998962402344, 0.0019273757934570312, 0.002018451690673828, 0.002109527587890625, 0.002200603485107422, 0.0022916793823242188, 0.0023827552795410156, 0.0024738311767578125, 0.0025649070739746094, 0.0026559829711914062, 0.002747058868408203, 0.002838134765625]}, "gradients/model.layers.11.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 8.0, 2.0, 5.0, 2.0, 6.0, 4.0, 7.0, 7.0, 10.0, 19.0, 32.0, 36.0, 72.0, 116.0, 192.0, 314.0, 651.0, 1179.0, 2561.0, 5526.0, 116796.0, 4028031.0, 29705.0, 4596.0, 2131.0, 1017.0, 529.0, 294.0, 147.0, 99.0, 59.0, 46.0, 27.0, 16.0, 12.0, 9.0, 2.0, 2.0, 9.0, 3.0, 3.0, 2.0, 4.0, 0.0, 1.0, 5.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.01434326171875, -0.013886451721191406, -0.013429641723632812, -0.012972831726074219, -0.012516021728515625, -0.012059211730957031, -0.011602401733398438, -0.011145591735839844, -0.01068878173828125, -0.010231971740722656, -0.009775161743164062, -0.009318351745605469, -0.008861541748046875, -0.008404731750488281, -0.007947921752929688, -0.007491111755371094, -0.0070343017578125, -0.006577491760253906, -0.0061206817626953125, -0.005663871765136719, -0.005207061767578125, -0.004750251770019531, -0.0042934417724609375, -0.0038366317749023438, -0.00337982177734375, -0.0029230117797851562, -0.0024662017822265625, -0.0020093917846679688, -0.001552581787109375, -0.0010957717895507812, -0.0006389617919921875, -0.00018215179443359375, 0.000274658203125, 0.0007314682006835938, 0.0011882781982421875, 0.0016450881958007812, 0.002101898193359375, 0.0025587081909179688, 0.0030155181884765625, 0.0034723281860351562, 0.00392913818359375, 0.004385948181152344, 0.0048427581787109375, 0.005299568176269531, 0.005756378173828125, 0.006213188171386719, 0.0066699981689453125, 0.007126808166503906, 0.0075836181640625, 0.008040428161621094, 0.008497238159179688, 0.008954048156738281, 0.009410858154296875, 0.009867668151855469, 0.010324478149414062, 0.010781288146972656, 0.01123809814453125, 0.011694908142089844, 0.012151718139648438, 0.012608528137207031, 0.013065338134765625, 0.013522148132324219, 0.013978958129882812, 0.014435768127441406, 0.014892578125]}, "gradients/model.layers.11.self_attn.k_proj.weight": {"_type": "histogram", "values": [3.0, 3.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 1.0, 1.0, 1.0, 2.0, 3.0, 5.0, 6.0, 6.0, 10.0, 9.0, 9.0, 16.0, 23.0, 30.0, 38.0, 79.0, 150.0, 474.0, 1537.0, 5165.0, 27451.0, 3769276.0, 373775.0, 11688.0, 2937.0, 929.0, 332.0, 128.0, 81.0, 34.0, 26.0, 14.0, 10.0, 3.0, 11.0, 11.0, 3.0, 2.0, 2.0, 3.0, 3.0, 2.0, 0.0, 0.0, 1.0, 0.0, 2.0, 2.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.002349853515625, -0.002270936965942383, -0.0021920204162597656, -0.0021131038665771484, -0.0020341873168945312, -0.001955270767211914, -0.0018763542175292969, -0.0017974376678466797, -0.0017185211181640625, -0.0016396045684814453, -0.0015606880187988281, -0.001481771469116211, -0.0014028549194335938, -0.0013239383697509766, -0.0012450218200683594, -0.0011661052703857422, -0.001087188720703125, -0.0010082721710205078, -0.0009293556213378906, -0.0008504390716552734, -0.0007715225219726562, -0.0006926059722900391, -0.0006136894226074219, -0.0005347728729248047, -0.0004558563232421875, -0.0003769397735595703, -0.0002980232238769531, -0.00021910667419433594, -0.00014019012451171875, -6.127357482910156e-05, 1.7642974853515625e-05, 9.655952453613281e-05, 0.00017547607421875, 0.0002543926239013672, 0.0003333091735839844, 0.00041222572326660156, 0.0004911422729492188, 0.0005700588226318359, 0.0006489753723144531, 0.0007278919219970703, 0.0008068084716796875, 0.0008857250213623047, 0.0009646415710449219, 0.001043558120727539, 0.0011224746704101562, 0.0012013912200927734, 0.0012803077697753906, 0.0013592243194580078, 0.001438140869140625, 0.0015170574188232422, 0.0015959739685058594, 0.0016748905181884766, 0.0017538070678710938, 0.001832723617553711, 0.0019116401672363281, 0.0019905567169189453, 0.0020694732666015625, 0.0021483898162841797, 0.002227306365966797, 0.002306222915649414, 0.0023851394653320312, 0.0024640560150146484, 0.0025429725646972656, 0.002621889114379883, 0.0027008056640625]}, "gradients/model.layers.11.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 0.0, 2.0, 3.0, 1.0, 1.0, 4.0, 6.0, 8.0, 3.0, 4.0, 7.0, 12.0, 7.0, 17.0, 17.0, 22.0, 41.0, 39.0, 55.0, 61.0, 88.0, 149.0, 228.0, 398.0, 696.0, 1408.0, 3748.0, 13901.0, 82499.0, 14379352.0, 2236083.0, 43407.0, 9274.0, 2862.0, 1143.0, 609.0, 343.0, 219.0, 133.0, 91.0, 53.0, 47.0, 38.0, 20.0, 22.0, 15.0, 15.0, 12.0, 11.0, 5.0, 7.0, 6.0, 4.0, 6.0, 6.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0], "bins": [-0.0032806396484375, -0.003177642822265625, -0.00307464599609375, -0.002971649169921875, -0.00286865234375, -0.002765655517578125, -0.00266265869140625, -0.002559661865234375, -0.0024566650390625, -0.002353668212890625, -0.00225067138671875, -0.002147674560546875, -0.002044677734375, -0.001941680908203125, -0.00183868408203125, -0.001735687255859375, -0.0016326904296875, -0.001529693603515625, -0.00142669677734375, -0.001323699951171875, -0.001220703125, -0.001117706298828125, -0.00101470947265625, -0.000911712646484375, -0.0008087158203125, -0.000705718994140625, -0.00060272216796875, -0.000499725341796875, -0.000396728515625, -0.000293731689453125, -0.00019073486328125, -8.7738037109375e-05, 1.52587890625e-05, 0.000118255615234375, 0.00022125244140625, 0.000324249267578125, 0.00042724609375, 0.000530242919921875, 0.00063323974609375, 0.000736236572265625, 0.0008392333984375, 0.000942230224609375, 0.00104522705078125, 0.001148223876953125, 0.001251220703125, 0.001354217529296875, 0.00145721435546875, 0.001560211181640625, 0.0016632080078125, 0.001766204833984375, 0.00186920166015625, 0.001972198486328125, 0.0020751953125, 0.002178192138671875, 0.00228118896484375, 0.002384185791015625, 0.0024871826171875, 0.002590179443359375, 0.00269317626953125, 0.002796173095703125, 0.002899169921875, 0.003002166748046875, 0.00310516357421875, 0.003208160400390625, 0.0033111572265625]}, "gradients/model.layers.11.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 2.0, 3.0, 4.0, 16.0, 29.0, 76.0, 142.0, 259.0, 455.0, 589.0, 655.0, 608.0, 498.0, 345.0, 196.0, 117.0, 47.0, 17.0, 8.0, 5.0, 4.0, 2.0, 3.0, 1.0, 2.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.000614166259765625, -0.0005882382392883301, -0.0005623102188110352, -0.0005363821983337402, -0.0005104541778564453, -0.0004845261573791504, -0.00045859813690185547, -0.00043267011642456055, -0.0004067420959472656, -0.0003808140754699707, -0.0003548860549926758, -0.00032895803451538086, -0.00030303001403808594, -0.000277101993560791, -0.0002511739730834961, -0.00022524595260620117, -0.00019931793212890625, -0.00017338991165161133, -0.0001474618911743164, -0.00012153387069702148, -9.560585021972656e-05, -6.967782974243164e-05, -4.374980926513672e-05, -1.7821788787841797e-05, 8.106231689453125e-06, 3.403425216674805e-05, 5.996227264404297e-05, 8.589029312133789e-05, 0.00011181831359863281, 0.00013774633407592773, 0.00016367435455322266, 0.00018960237503051758, 0.0002155303955078125, 0.00024145841598510742, 0.00026738643646240234, 0.00029331445693969727, 0.0003192424774169922, 0.0003451704978942871, 0.00037109851837158203, 0.00039702653884887695, 0.0004229545593261719, 0.0004488825798034668, 0.0004748106002807617, 0.0005007386207580566, 0.0005266666412353516, 0.0005525946617126465, 0.0005785226821899414, 0.0006044507026672363, 0.0006303787231445312, 0.0006563067436218262, 0.0006822347640991211, 0.000708162784576416, 0.0007340908050537109, 0.0007600188255310059, 0.0007859468460083008, 0.0008118748664855957, 0.0008378028869628906, 0.0008637309074401855, 0.0008896589279174805, 0.0009155869483947754, 0.0009415149688720703, 0.0009674429893493652, 0.0009933710098266602, 0.001019299030303955, 0.00104522705078125]}, "gradients/model.layers.10.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 3.0, 1.0, 3.0, 4.0, 7.0, 11.0, 12.0, 14.0, 53.0, 46.0, 107.0, 123.0, 201.0, 366.0, 635.0, 1453.0, 10689.0, 4560789.0, 16777216.0, 124514.0, 3389.0, 868.0, 442.0, 290.0, 179.0, 117.0, 58.0, 50.0, 27.0, 18.0, 14.0, 7.0, 3.0, 1.0, 1.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.003875732421875, -0.0037016868591308594, -0.0035276412963867188, -0.003353595733642578, -0.0031795501708984375, -0.003005504608154297, -0.0028314590454101562, -0.0026574134826660156, -0.002483367919921875, -0.0023093223571777344, -0.0021352767944335938, -0.001961231231689453, -0.0017871856689453125, -0.0016131401062011719, -0.0014390945434570312, -0.0012650489807128906, -0.00109100341796875, -0.0009169578552246094, -0.0007429122924804688, -0.0005688667297363281, -0.0003948211669921875, -0.00022077560424804688, -4.673004150390625e-05, 0.00012731552124023438, 0.000301361083984375, 0.0004754066467285156, 0.0006494522094726562, 0.0008234977722167969, 0.0009975433349609375, 0.0011715888977050781, 0.0013456344604492188, 0.0015196800231933594, 0.0016937255859375, 0.0018677711486816406, 0.0020418167114257812, 0.002215862274169922, 0.0023899078369140625, 0.002563953399658203, 0.0027379989624023438, 0.0029120445251464844, 0.003086090087890625, 0.0032601356506347656, 0.0034341812133789062, 0.003608226776123047, 0.0037822723388671875, 0.003956317901611328, 0.004130363464355469, 0.004304409027099609, 0.00447845458984375, 0.004652500152587891, 0.004826545715332031, 0.005000591278076172, 0.0051746368408203125, 0.005348682403564453, 0.005522727966308594, 0.005696773529052734, 0.005870819091796875, 0.006044864654541016, 0.006218910217285156, 0.006392955780029297, 0.0065670013427734375, 0.006741046905517578, 0.006915092468261719, 0.007089138031005859, 0.00726318359375]}, "gradients/model.layers.10.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 2.0, 7.0, 6.0, 14.0, 9.0, 13.0, 24.0, 53.0, 99.0, 106.0, 184.0, 236.0, 477.0, 798.0, 1519.0, 3718.0, 11130.0, 51506.0, 1740936.0, 16777216.0, 6329909.0, 83825.0, 14979.0, 4370.0, 1897.0, 848.0, 479.0, 332.0, 192.0, 122.0, 82.0, 53.0, 36.0, 21.0, 12.0, 11.0, 10.0, 8.0, 3.0, 0.0, 0.0, 0.0, 4.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.003021240234375, -0.0029191970825195312, -0.0028171539306640625, -0.0027151107788085938, -0.002613067626953125, -0.0025110244750976562, -0.0024089813232421875, -0.0023069381713867188, -0.00220489501953125, -0.0021028518676757812, -0.0020008087158203125, -0.0018987655639648438, -0.001796722412109375, -0.0016946792602539062, -0.0015926361083984375, -0.0014905929565429688, -0.0013885498046875, -0.0012865066528320312, -0.0011844635009765625, -0.0010824203491210938, -0.000980377197265625, -0.0008783340454101562, -0.0007762908935546875, -0.0006742477416992188, -0.00057220458984375, -0.00047016143798828125, -0.0003681182861328125, -0.00026607513427734375, -0.000164031982421875, -6.198883056640625e-05, 4.00543212890625e-05, 0.00014209747314453125, 0.000244140625, 0.00034618377685546875, 0.0004482269287109375, 0.0005502700805664062, 0.000652313232421875, 0.0007543563842773438, 0.0008563995361328125, 0.0009584426879882812, 0.00106048583984375, 0.0011625289916992188, 0.0012645721435546875, 0.0013666152954101562, 0.001468658447265625, 0.0015707015991210938, 0.0016727447509765625, 0.0017747879028320312, 0.0018768310546875, 0.0019788742065429688, 0.0020809173583984375, 0.0021829605102539062, 0.002285003662109375, 0.0023870468139648438, 0.0024890899658203125, 0.0025911331176757812, 0.00269317626953125, 0.0027952194213867188, 0.0028972625732421875, 0.0029993057250976562, 0.003101348876953125, 0.0032033920288085938, 0.0033054351806640625, 0.0034074783325195312, 0.003509521484375]}, "gradients/model.layers.10.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 4.0, 5.0, 5.0, 18.0, 25.0, 80.0, 209.0, 431.0, 684.0, 824.0, 761.0, 566.0, 284.0, 125.0, 38.0, 16.0, 5.0, 2.0, 3.0, 0.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0028839111328125, -0.0028030872344970703, -0.0027222633361816406, -0.002641439437866211, -0.0025606155395507812, -0.0024797916412353516, -0.002398967742919922, -0.002318143844604492, -0.0022373199462890625, -0.002156496047973633, -0.002075672149658203, -0.0019948482513427734, -0.0019140243530273438, -0.001833200454711914, -0.0017523765563964844, -0.0016715526580810547, -0.001590728759765625, -0.0015099048614501953, -0.0014290809631347656, -0.001348257064819336, -0.0012674331665039062, -0.0011866092681884766, -0.0011057853698730469, -0.0010249614715576172, -0.0009441375732421875, -0.0008633136749267578, -0.0007824897766113281, -0.0007016658782958984, -0.0006208419799804688, -0.0005400180816650391, -0.0004591941833496094, -0.0003783702850341797, -0.00029754638671875, -0.0002167224884033203, -0.00013589859008789062, -5.507469177246094e-05, 2.574920654296875e-05, 0.00010657310485839844, 0.00018739700317382812, 0.0002682209014892578, 0.0003490447998046875, 0.0004298686981201172, 0.0005106925964355469, 0.0005915164947509766, 0.0006723403930664062, 0.0007531642913818359, 0.0008339881896972656, 0.0009148120880126953, 0.000995635986328125, 0.0010764598846435547, 0.0011572837829589844, 0.001238107681274414, 0.0013189315795898438, 0.0013997554779052734, 0.0014805793762207031, 0.0015614032745361328, 0.0016422271728515625, 0.0017230510711669922, 0.0018038749694824219, 0.0018846988677978516, 0.0019655227661132812, 0.002046346664428711, 0.0021271705627441406, 0.0022079944610595703, 0.002288818359375]}, "gradients/model.layers.10.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 1.0, 3.0, 3.0, 4.0, 7.0, 5.0, 9.0, 12.0, 21.0, 26.0, 39.0, 67.0, 136.0, 202.0, 335.0, 635.0, 1220.0, 2444.0, 5268.0, 12804.0, 40622.0, 195909.0, 1666845.0, 9895993.0, 4344582.0, 499469.0, 75666.0, 20246.0, 7802.0, 3393.0, 1614.0, 795.0, 436.0, 238.0, 144.0, 73.0, 51.0, 24.0, 14.0, 16.0, 9.0, 4.0, 7.0, 7.0, 2.0, 2.0, 0.0, 0.0, 3.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 2.0], "bins": [-0.00286865234375, -0.002765178680419922, -0.0026617050170898438, -0.0025582313537597656, -0.0024547576904296875, -0.0023512840270996094, -0.0022478103637695312, -0.002144336700439453, -0.002040863037109375, -0.0019373893737792969, -0.0018339157104492188, -0.0017304420471191406, -0.0016269683837890625, -0.0015234947204589844, -0.0014200210571289062, -0.0013165473937988281, -0.00121307373046875, -0.0011096000671386719, -0.0010061264038085938, -0.0009026527404785156, -0.0007991790771484375, -0.0006957054138183594, -0.0005922317504882812, -0.0004887580871582031, -0.000385284423828125, -0.0002818107604980469, -0.00017833709716796875, -7.486343383789062e-05, 2.86102294921875e-05, 0.00013208389282226562, 0.00023555755615234375, 0.0003390312194824219, 0.0004425048828125, 0.0005459785461425781, 0.0006494522094726562, 0.0007529258728027344, 0.0008563995361328125, 0.0009598731994628906, 0.0010633468627929688, 0.0011668205261230469, 0.001270294189453125, 0.0013737678527832031, 0.0014772415161132812, 0.0015807151794433594, 0.0016841888427734375, 0.0017876625061035156, 0.0018911361694335938, 0.001994609832763672, 0.00209808349609375, 0.002201557159423828, 0.0023050308227539062, 0.0024085044860839844, 0.0025119781494140625, 0.0026154518127441406, 0.0027189254760742188, 0.002822399139404297, 0.002925872802734375, 0.003029346466064453, 0.0031328201293945312, 0.0032362937927246094, 0.0033397674560546875, 0.0034432411193847656, 0.0035467147827148438, 0.003650188446044922, 0.003753662109375]}, "gradients/model.layers.10.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 3.0, 5.0, 1.0, 0.0, 2.0, 4.0, 3.0, 8.0, 10.0, 13.0, 13.0, 21.0, 24.0, 33.0, 49.0, 55.0, 92.0, 122.0, 165.0, 235.0, 323.0, 521.0, 773.0, 1135.0, 1957.0, 3420.0, 7526.0, 120282.0, 3690277.0, 347295.0, 10076.0, 3722.0, 2152.0, 1298.0, 855.0, 513.0, 363.0, 270.0, 189.0, 132.0, 84.0, 71.0, 59.0, 39.0, 19.0, 20.0, 20.0, 14.0, 9.0, 10.0, 3.0, 5.0, 2.0, 5.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 2.0], "bins": [-0.00970458984375, -0.009397506713867188, -0.009090423583984375, -0.008783340454101562, -0.00847625732421875, -0.008169174194335938, -0.007862091064453125, -0.0075550079345703125, -0.0072479248046875, -0.0069408416748046875, -0.006633758544921875, -0.0063266754150390625, -0.00601959228515625, -0.0057125091552734375, -0.005405426025390625, -0.0050983428955078125, -0.004791259765625, -0.0044841766357421875, -0.004177093505859375, -0.0038700103759765625, -0.00356292724609375, -0.0032558441162109375, -0.002948760986328125, -0.0026416778564453125, -0.0023345947265625, -0.0020275115966796875, -0.001720428466796875, -0.0014133453369140625, -0.00110626220703125, -0.0007991790771484375, -0.000492095947265625, -0.0001850128173828125, 0.0001220703125, 0.0004291534423828125, 0.000736236572265625, 0.0010433197021484375, 0.00135040283203125, 0.0016574859619140625, 0.001964569091796875, 0.0022716522216796875, 0.0025787353515625, 0.0028858184814453125, 0.003192901611328125, 0.0034999847412109375, 0.00380706787109375, 0.0041141510009765625, 0.004421234130859375, 0.0047283172607421875, 0.005035400390625, 0.0053424835205078125, 0.005649566650390625, 0.0059566497802734375, 0.00626373291015625, 0.0065708160400390625, 0.006877899169921875, 0.0071849822998046875, 0.0074920654296875, 0.0077991485595703125, 0.008106231689453125, 0.008413314819335938, 0.00872039794921875, 0.009027481079101562, 0.009334564208984375, 0.009641647338867188, 0.00994873046875]}, "gradients/model.layers.10.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 1.0, 0.0, 3.0, 2.0, 3.0, 2.0, 6.0, 3.0, 0.0, 4.0, 3.0, 10.0, 20.0, 29.0, 45.0, 51.0, 69.0, 110.0, 191.0, 324.0, 631.0, 1443.0, 3729.0, 11416.0, 83245.0, 3726039.0, 339825.0, 18086.0, 5124.0, 1972.0, 844.0, 425.0, 229.0, 131.0, 68.0, 52.0, 39.0, 34.0, 31.0, 10.0, 7.0, 8.0, 7.0, 8.0, 7.0, 6.0, 2.0, 0.0, 3.0, 0.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.0017852783203125, -0.0017352104187011719, -0.0016851425170898438, -0.0016350746154785156, -0.0015850067138671875, -0.0015349388122558594, -0.0014848709106445312, -0.0014348030090332031, -0.001384735107421875, -0.0013346672058105469, -0.0012845993041992188, -0.0012345314025878906, -0.0011844635009765625, -0.0011343955993652344, -0.0010843276977539062, -0.0010342597961425781, -0.00098419189453125, -0.0009341239929199219, -0.0008840560913085938, -0.0008339881896972656, -0.0007839202880859375, -0.0007338523864746094, -0.0006837844848632812, -0.0006337165832519531, -0.000583648681640625, -0.0005335807800292969, -0.00048351287841796875, -0.0004334449768066406, -0.0003833770751953125, -0.0003333091735839844, -0.00028324127197265625, -0.00023317337036132812, -0.00018310546875, -0.00013303756713867188, -8.296966552734375e-05, -3.2901763916015625e-05, 1.71661376953125e-05, 6.723403930664062e-05, 0.00011730194091796875, 0.00016736984252929688, 0.000217437744140625, 0.0002675056457519531, 0.00031757354736328125, 0.0003676414489746094, 0.0004177093505859375, 0.0004677772521972656, 0.0005178451538085938, 0.0005679130554199219, 0.00061798095703125, 0.0006680488586425781, 0.0007181167602539062, 0.0007681846618652344, 0.0008182525634765625, 0.0008683204650878906, 0.0009183883666992188, 0.0009684562683105469, 0.001018524169921875, 0.0010685920715332031, 0.0011186599731445312, 0.0011687278747558594, 0.0012187957763671875, 0.0012688636779785156, 0.0013189315795898438, 0.0013689994812011719, 0.0014190673828125]}, "gradients/model.layers.10.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 2.0, 0.0, 0.0, 2.0, 3.0, 1.0, 3.0, 2.0, 4.0, 4.0, 5.0, 3.0, 13.0, 10.0, 14.0, 18.0, 26.0, 33.0, 44.0, 82.0, 134.0, 204.0, 391.0, 999.0, 3679.0, 24783.0, 4601402.0, 12108805.0, 30052.0, 4296.0, 1082.0, 466.0, 215.0, 136.0, 90.0, 65.0, 33.0, 19.0, 20.0, 18.0, 8.0, 8.0, 11.0, 6.0, 2.0, 6.0, 1.0, 4.0, 4.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.00421142578125, -0.004071235656738281, -0.0039310455322265625, -0.0037908554077148438, -0.003650665283203125, -0.0035104751586914062, -0.0033702850341796875, -0.0032300949096679688, -0.00308990478515625, -0.0029497146606445312, -0.0028095245361328125, -0.0026693344116210938, -0.002529144287109375, -0.0023889541625976562, -0.0022487640380859375, -0.0021085739135742188, -0.0019683837890625, -0.0018281936645507812, -0.0016880035400390625, -0.0015478134155273438, -0.001407623291015625, -0.0012674331665039062, -0.0011272430419921875, -0.0009870529174804688, -0.00084686279296875, -0.0007066726684570312, -0.0005664825439453125, -0.00042629241943359375, -0.000286102294921875, -0.00014591217041015625, -5.7220458984375e-06, 0.00013446807861328125, 0.000274658203125, 0.00041484832763671875, 0.0005550384521484375, 0.0006952285766601562, 0.000835418701171875, 0.0009756088256835938, 0.0011157989501953125, 0.0012559890747070312, 0.00139617919921875, 0.0015363693237304688, 0.0016765594482421875, 0.0018167495727539062, 0.001956939697265625, 0.0020971298217773438, 0.0022373199462890625, 0.0023775100708007812, 0.0025177001953125, 0.0026578903198242188, 0.0027980804443359375, 0.0029382705688476562, 0.003078460693359375, 0.0032186508178710938, 0.0033588409423828125, 0.0034990310668945312, 0.00363922119140625, 0.0037794113159179688, 0.0039196014404296875, 0.004059791564941406, 0.004199981689453125, 0.004340171813964844, 0.0044803619384765625, 0.004620552062988281, 0.0047607421875]}, "gradients/model.layers.10.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 2.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 2.0, 1.0, 2.0, 10.0, 21.0, 52.0, 122.0, 260.0, 486.0, 723.0, 833.0, 670.0, 467.0, 244.0, 120.0, 42.0, 12.0, 1.0, 2.0, 0.0, 4.0, 1.0, 0.0, 2.0, 2.0, 0.0, 0.0, 0.0, 2.0, 1.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.000621795654296875, -0.000589907169342041, -0.000558018684387207, -0.000526130199432373, -0.0004942417144775391, -0.0004623532295227051, -0.0004304647445678711, -0.0003985762596130371, -0.0003666877746582031, -0.00033479928970336914, -0.00030291080474853516, -0.00027102231979370117, -0.0002391338348388672, -0.0002072453498840332, -0.00017535686492919922, -0.00014346837997436523, -0.00011157989501953125, -7.969141006469727e-05, -4.780292510986328e-05, -1.5914440155029297e-05, 1.5974044799804688e-05, 4.786252975463867e-05, 7.975101470947266e-05, 0.00011163949966430664, 0.00014352798461914062, 0.0001754164695739746, 0.0002073049545288086, 0.00023919343948364258, 0.00027108192443847656, 0.00030297040939331055, 0.00033485889434814453, 0.0003667473793029785, 0.0003986358642578125, 0.0004305243492126465, 0.00046241283416748047, 0.0004943013191223145, 0.0005261898040771484, 0.0005580782890319824, 0.0005899667739868164, 0.0006218552589416504, 0.0006537437438964844, 0.0006856322288513184, 0.0007175207138061523, 0.0007494091987609863, 0.0007812976837158203, 0.0008131861686706543, 0.0008450746536254883, 0.0008769631385803223, 0.0009088516235351562, 0.0009407401084899902, 0.0009726285934448242, 0.0010045170783996582, 0.0010364055633544922, 0.0010682940483093262, 0.0011001825332641602, 0.0011320710182189941, 0.0011639595031738281, 0.0011958479881286621, 0.001227736473083496, 0.00125962495803833, 0.001291513442993164, 0.001323401927947998, 0.001355290412902832, 0.001387178897857666, 0.0014190673828125]}, "gradients/model.layers.9.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 2.0, 2.0, 12.0, 15.0, 41.0, 40.0, 59.0, 112.0, 138.0, 185.0, 277.0, 467.0, 1161.0, 7793.0, 2363088.0, 16777216.0, 135628.0, 2953.0, 773.0, 320.0, 216.0, 166.0, 128.0, 87.0, 57.0, 51.0, 25.0, 17.0, 16.0, 1.0, 1.0, 2.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.00811767578125, -0.007930278778076172, -0.007742881774902344, -0.007555484771728516, -0.0073680877685546875, -0.007180690765380859, -0.006993293762207031, -0.006805896759033203, -0.006618499755859375, -0.006431102752685547, -0.006243705749511719, -0.006056308746337891, -0.0058689117431640625, -0.005681514739990234, -0.005494117736816406, -0.005306720733642578, -0.00511932373046875, -0.004931926727294922, -0.004744529724121094, -0.004557132720947266, -0.0043697357177734375, -0.004182338714599609, -0.003994941711425781, -0.003807544708251953, -0.003620147705078125, -0.003432750701904297, -0.0032453536987304688, -0.0030579566955566406, -0.0028705596923828125, -0.0026831626892089844, -0.0024957656860351562, -0.002308368682861328, -0.0021209716796875, -0.0019335746765136719, -0.0017461776733398438, -0.0015587806701660156, -0.0013713836669921875, -0.0011839866638183594, -0.0009965896606445312, -0.0008091926574707031, -0.000621795654296875, -0.0004343986511230469, -0.00024700164794921875, -5.9604644775390625e-05, 0.0001277923583984375, 0.0003151893615722656, 0.0005025863647460938, 0.0006899833679199219, 0.00087738037109375, 0.0010647773742675781, 0.0012521743774414062, 0.0014395713806152344, 0.0016269683837890625, 0.0018143653869628906, 0.0020017623901367188, 0.002189159393310547, 0.002376556396484375, 0.002563953399658203, 0.0027513504028320312, 0.0029387474060058594, 0.0031261444091796875, 0.0033135414123535156, 0.0035009384155273438, 0.003688335418701172, 0.003875732421875]}, "gradients/model.layers.9.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 3.0, 0.0, 2.0, 4.0, 5.0, 5.0, 5.0, 10.0, 14.0, 22.0, 25.0, 28.0, 58.0, 66.0, 150.0, 210.0, 358.0, 577.0, 1180.0, 2452.0, 6385.0, 19205.0, 103009.0, 5502323.0, 16777216.0, 3040126.0, 82597.0, 16895.0, 5632.0, 2230.0, 1088.0, 539.0, 315.0, 211.0, 114.0, 97.0, 70.0, 36.0, 22.0, 22.0, 9.0, 10.0, 6.0, 6.0, 3.0, 2.0, 0.0, 0.0, 1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.002685546875, -0.0025911331176757812, -0.0024967193603515625, -0.0024023056030273438, -0.002307891845703125, -0.0022134780883789062, -0.0021190643310546875, -0.0020246505737304688, -0.00193023681640625, -0.0018358230590820312, -0.0017414093017578125, -0.0016469955444335938, -0.001552581787109375, -0.0014581680297851562, -0.0013637542724609375, -0.0012693405151367188, -0.0011749267578125, -0.0010805130004882812, -0.0009860992431640625, -0.0008916854858398438, -0.000797271728515625, -0.0007028579711914062, -0.0006084442138671875, -0.0005140304565429688, -0.00041961669921875, -0.00032520294189453125, -0.0002307891845703125, -0.00013637542724609375, -4.1961669921875e-05, 5.245208740234375e-05, 0.0001468658447265625, 0.00024127960205078125, 0.000335693359375, 0.00043010711669921875, 0.0005245208740234375, 0.0006189346313476562, 0.000713348388671875, 0.0008077621459960938, 0.0009021759033203125, 0.0009965896606445312, 0.00109100341796875, 0.0011854171752929688, 0.0012798309326171875, 0.0013742446899414062, 0.001468658447265625, 0.0015630722045898438, 0.0016574859619140625, 0.0017518997192382812, 0.0018463134765625, 0.0019407272338867188, 0.0020351409912109375, 0.0021295547485351562, 0.002223968505859375, 0.0023183822631835938, 0.0024127960205078125, 0.0025072097778320312, 0.00260162353515625, 0.0026960372924804688, 0.0027904510498046875, 0.0028848648071289062, 0.002979278564453125, 0.0030736923217773438, 0.0031681060791015625, 0.0032625198364257812, 0.00335693359375]}, "gradients/model.layers.9.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 6.0, 3.0, 7.0, 12.0, 22.0, 59.0, 214.0, 534.0, 921.0, 1029.0, 724.0, 374.0, 135.0, 29.0, 12.0, 4.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0028228759765625, -0.0027227401733398438, -0.0026226043701171875, -0.0025224685668945312, -0.002422332763671875, -0.0023221969604492188, -0.0022220611572265625, -0.0021219253540039062, -0.00202178955078125, -0.0019216537475585938, -0.0018215179443359375, -0.0017213821411132812, -0.001621246337890625, -0.0015211105346679688, -0.0014209747314453125, -0.0013208389282226562, -0.001220703125, -0.0011205673217773438, -0.0010204315185546875, -0.0009202957153320312, -0.000820159912109375, -0.0007200241088867188, -0.0006198883056640625, -0.0005197525024414062, -0.00041961669921875, -0.00031948089599609375, -0.0002193450927734375, -0.00011920928955078125, -1.9073486328125e-05, 8.106231689453125e-05, 0.0001811981201171875, 0.00028133392333984375, 0.0003814697265625, 0.00048160552978515625, 0.0005817413330078125, 0.0006818771362304688, 0.000782012939453125, 0.0008821487426757812, 0.0009822845458984375, 0.0010824203491210938, 0.00118255615234375, 0.0012826919555664062, 0.0013828277587890625, 0.0014829635620117188, 0.001583099365234375, 0.0016832351684570312, 0.0017833709716796875, 0.0018835067749023438, 0.001983642578125, 0.0020837783813476562, 0.0021839141845703125, 0.0022840499877929688, 0.002384185791015625, 0.0024843215942382812, 0.0025844573974609375, 0.0026845932006835938, 0.00278472900390625, 0.0028848648071289062, 0.0029850006103515625, 0.0030851364135742188, 0.003185272216796875, 0.0032854080200195312, 0.0033855438232421875, 0.0034856796264648438, 0.0035858154296875]}, "gradients/model.layers.9.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 0.0, 1.0, 1.0, 2.0, 1.0, 0.0, 1.0, 1.0, 2.0, 7.0, 9.0, 5.0, 24.0, 36.0, 48.0, 88.0, 186.0, 349.0, 648.0, 1465.0, 3234.0, 8248.0, 25844.0, 128450.0, 1762922.0, 11183945.0, 3387441.0, 221302.0, 34853.0, 10536.0, 4188.0, 1733.0, 774.0, 430.0, 187.0, 120.0, 60.0, 31.0, 16.0, 11.0, 4.0, 1.0, 2.0, 2.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.003173828125, -0.003058910369873047, -0.0029439926147460938, -0.0028290748596191406, -0.0027141571044921875, -0.0025992393493652344, -0.0024843215942382812, -0.002369403839111328, -0.002254486083984375, -0.002139568328857422, -0.0020246505737304688, -0.0019097328186035156, -0.0017948150634765625, -0.0016798973083496094, -0.0015649795532226562, -0.0014500617980957031, -0.00133514404296875, -0.0012202262878417969, -0.0011053085327148438, -0.0009903907775878906, -0.0008754730224609375, -0.0007605552673339844, -0.0006456375122070312, -0.0005307197570800781, -0.000415802001953125, -0.0003008842468261719, -0.00018596649169921875, -7.104873657226562e-05, 4.38690185546875e-05, 0.00015878677368164062, 0.00027370452880859375, 0.0003886222839355469, 0.0005035400390625, 0.0006184577941894531, 0.0007333755493164062, 0.0008482933044433594, 0.0009632110595703125, 0.0010781288146972656, 0.0011930465698242188, 0.0013079643249511719, 0.001422882080078125, 0.0015377998352050781, 0.0016527175903320312, 0.0017676353454589844, 0.0018825531005859375, 0.0019974708557128906, 0.0021123886108398438, 0.002227306365966797, 0.00234222412109375, 0.002457141876220703, 0.0025720596313476562, 0.0026869773864746094, 0.0028018951416015625, 0.0029168128967285156, 0.0030317306518554688, 0.003146648406982422, 0.003261566162109375, 0.003376483917236328, 0.0034914016723632812, 0.0036063194274902344, 0.0037212371826171875, 0.0038361549377441406, 0.003951072692871094, 0.004065990447998047, 0.004180908203125]}, "gradients/model.layers.9.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0, 0.0, 0.0, 2.0, 1.0, 2.0, 2.0, 7.0, 3.0, 4.0, 9.0, 10.0, 8.0, 22.0, 34.0, 28.0, 60.0, 65.0, 141.0, 210.0, 423.0, 852.0, 1858.0, 4528.0, 46772.0, 4120125.0, 12562.0, 3553.0, 1461.0, 698.0, 335.0, 196.0, 81.0, 61.0, 40.0, 33.0, 21.0, 22.0, 15.0, 13.0, 8.0, 6.0, 5.0, 7.0, 1.0, 0.0, 5.0, 1.0, 4.0, 0.0, 0.0, 1.0, 1.0, 0.0, 2.0], "bins": [-0.0203857421875, -0.019792556762695312, -0.019199371337890625, -0.018606185913085938, -0.01801300048828125, -0.017419815063476562, -0.016826629638671875, -0.016233444213867188, -0.0156402587890625, -0.015047073364257812, -0.014453887939453125, -0.013860702514648438, -0.01326751708984375, -0.012674331665039062, -0.012081146240234375, -0.011487960815429688, -0.010894775390625, -0.010301589965820312, -0.009708404541015625, -0.009115219116210938, -0.00852203369140625, -0.007928848266601562, -0.007335662841796875, -0.0067424774169921875, -0.0061492919921875, -0.0055561065673828125, -0.004962921142578125, -0.0043697357177734375, -0.00377655029296875, -0.0031833648681640625, -0.002590179443359375, -0.0019969940185546875, -0.00140380859375, -0.0008106231689453125, -0.000217437744140625, 0.0003757476806640625, 0.00096893310546875, 0.0015621185302734375, 0.002155303955078125, 0.0027484893798828125, 0.0033416748046875, 0.0039348602294921875, 0.004528045654296875, 0.0051212310791015625, 0.00571441650390625, 0.0063076019287109375, 0.006900787353515625, 0.0074939727783203125, 0.008087158203125, 0.008680343627929688, 0.009273529052734375, 0.009866714477539062, 0.01045989990234375, 0.011053085327148438, 0.011646270751953125, 0.012239456176757812, 0.0128326416015625, 0.013425827026367188, 0.014019012451171875, 0.014612197875976562, 0.01520538330078125, 0.015798568725585938, 0.016391754150390625, 0.016984939575195312, 0.017578125]}, "gradients/model.layers.9.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 2.0, 2.0, 1.0, 1.0, 0.0, 1.0, 1.0, 7.0, 6.0, 7.0, 5.0, 4.0, 14.0, 10.0, 14.0, 11.0, 20.0, 24.0, 56.0, 74.0, 118.0, 242.0, 549.0, 1402.0, 5151.0, 46284.0, 4072464.0, 59438.0, 5647.0, 1568.0, 552.0, 218.0, 125.0, 85.0, 41.0, 40.0, 26.0, 21.0, 18.0, 7.0, 8.0, 7.0, 5.0, 2.0, 4.0, 4.0, 2.0, 1.0, 4.0, 2.0, 1.0, 0.0, 0.0, 1.0, 2.0, 1.0, 0.0, 2.0], "bins": [-0.003265380859375, -0.003165006637573242, -0.0030646324157714844, -0.0029642581939697266, -0.0028638839721679688, -0.002763509750366211, -0.002663135528564453, -0.0025627613067626953, -0.0024623870849609375, -0.0023620128631591797, -0.002261638641357422, -0.002161264419555664, -0.0020608901977539062, -0.0019605159759521484, -0.0018601417541503906, -0.0017597675323486328, -0.001659393310546875, -0.0015590190887451172, -0.0014586448669433594, -0.0013582706451416016, -0.0012578964233398438, -0.001157522201538086, -0.0010571479797363281, -0.0009567737579345703, -0.0008563995361328125, -0.0007560253143310547, -0.0006556510925292969, -0.0005552768707275391, -0.00045490264892578125, -0.00035452842712402344, -0.0002541542053222656, -0.0001537799835205078, -5.340576171875e-05, 4.696846008300781e-05, 0.00014734268188476562, 0.00024771690368652344, 0.00034809112548828125, 0.00044846534729003906, 0.0005488395690917969, 0.0006492137908935547, 0.0007495880126953125, 0.0008499622344970703, 0.0009503364562988281, 0.001050710678100586, 0.0011510848999023438, 0.0012514591217041016, 0.0013518333435058594, 0.0014522075653076172, 0.001552581787109375, 0.0016529560089111328, 0.0017533302307128906, 0.0018537044525146484, 0.0019540786743164062, 0.002054452896118164, 0.002154827117919922, 0.0022552013397216797, 0.0023555755615234375, 0.0024559497833251953, 0.002556324005126953, 0.002656698226928711, 0.0027570724487304688, 0.0028574466705322266, 0.0029578208923339844, 0.003058195114135742, 0.0031585693359375]}, "gradients/model.layers.9.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 3.0, 2.0, 0.0, 1.0, 0.0, 1.0, 4.0, 3.0, 4.0, 6.0, 8.0, 4.0, 15.0, 18.0, 42.0, 72.0, 136.0, 326.0, 857.0, 3537.0, 38254.0, 16653584.0, 74028.0, 4614.0, 1001.0, 334.0, 156.0, 69.0, 50.0, 17.0, 18.0, 9.0, 6.0, 5.0, 2.0, 3.0, 3.0, 2.0, 5.0, 2.0, 2.0, 0.0, 3.0, 1.0, 1.0, 0.0, 1.0, 1.0], "bins": [-0.00750732421875, -0.007307529449462891, -0.007107734680175781, -0.006907939910888672, -0.0067081451416015625, -0.006508350372314453, -0.006308555603027344, -0.006108760833740234, -0.005908966064453125, -0.005709171295166016, -0.005509376525878906, -0.005309581756591797, -0.0051097869873046875, -0.004909992218017578, -0.004710197448730469, -0.004510402679443359, -0.00431060791015625, -0.004110813140869141, -0.003911018371582031, -0.003711223602294922, -0.0035114288330078125, -0.003311634063720703, -0.0031118392944335938, -0.0029120445251464844, -0.002712249755859375, -0.0025124549865722656, -0.0023126602172851562, -0.002112865447998047, -0.0019130706787109375, -0.0017132759094238281, -0.0015134811401367188, -0.0013136863708496094, -0.0011138916015625, -0.0009140968322753906, -0.0007143020629882812, -0.0005145072937011719, -0.0003147125244140625, -0.00011491775512695312, 8.487701416015625e-05, 0.0002846717834472656, 0.000484466552734375, 0.0006842613220214844, 0.0008840560913085938, 0.0010838508605957031, 0.0012836456298828125, 0.0014834403991699219, 0.0016832351684570312, 0.0018830299377441406, 0.00208282470703125, 0.0022826194763183594, 0.0024824142456054688, 0.002682209014892578, 0.0028820037841796875, 0.003081798553466797, 0.0032815933227539062, 0.0034813880920410156, 0.003681182861328125, 0.0038809776306152344, 0.004080772399902344, 0.004280567169189453, 0.0044803619384765625, 0.004680156707763672, 0.004879951477050781, 0.005079746246337891, 0.005279541015625]}, "gradients/model.layers.9.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 4.0, 21.0, 551.0, 2678.0, 778.0, 41.0, 9.0, 2.0, 1.0, 1.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00494384765625, -0.004808664321899414, -0.004673480987548828, -0.004538297653198242, -0.004403114318847656, -0.00426793098449707, -0.004132747650146484, -0.0039975643157958984, -0.0038623809814453125, -0.0037271976470947266, -0.0035920143127441406, -0.0034568309783935547, -0.0033216476440429688, -0.003186464309692383, -0.003051280975341797, -0.002916097640991211, -0.002780914306640625, -0.002645730972290039, -0.002510547637939453, -0.002375364303588867, -0.0022401809692382812, -0.0021049976348876953, -0.0019698143005371094, -0.0018346309661865234, -0.0016994476318359375, -0.0015642642974853516, -0.0014290809631347656, -0.0012938976287841797, -0.0011587142944335938, -0.0010235309600830078, -0.0008883476257324219, -0.0007531642913818359, -0.00061798095703125, -0.00048279762268066406, -0.0003476142883300781, -0.0002124309539794922, -7.724761962890625e-05, 5.793571472167969e-05, 0.00019311904907226562, 0.00032830238342285156, 0.0004634857177734375, 0.0005986690521240234, 0.0007338523864746094, 0.0008690357208251953, 0.0010042190551757812, 0.0011394023895263672, 0.0012745857238769531, 0.001409769058227539, 0.001544952392578125, 0.001680135726928711, 0.0018153190612792969, 0.0019505023956298828, 0.0020856857299804688, 0.0022208690643310547, 0.0023560523986816406, 0.0024912357330322266, 0.0026264190673828125, 0.0027616024017333984, 0.0028967857360839844, 0.0030319690704345703, 0.0031671524047851562, 0.003302335739135742, 0.003437519073486328, 0.003572702407836914, 0.0037078857421875]}, "gradients/model.layers.8.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 2.0, 5.0, 3.0, 8.0, 12.0, 23.0, 38.0, 95.0, 123.0, 240.0, 453.0, 881.0, 1775.0, 4321.0, 11265.0, 37144.0, 173699.0, 1234254.0, 9881094.0, 16777216.0, 15428955.0, 2270686.0, 288261.0, 56135.0, 15780.0, 5680.0, 2401.0, 1143.0, 536.0, 299.0, 155.0, 90.0, 52.0, 25.0, 19.0, 10.0, 7.0, 7.0, 4.0, 2.0, 2.0, 1.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00107574462890625, -0.001033782958984375, -0.0009918212890625, -0.000949859619140625, -0.00090789794921875, -0.000865936279296875, -0.000823974609375, -0.000782012939453125, -0.00074005126953125, -0.000698089599609375, -0.0006561279296875, -0.000614166259765625, -0.00057220458984375, -0.000530242919921875, -0.00048828125, -0.000446319580078125, -0.00040435791015625, -0.000362396240234375, -0.0003204345703125, -0.000278472900390625, -0.00023651123046875, -0.000194549560546875, -0.000152587890625, -0.000110626220703125, -6.866455078125e-05, -2.6702880859375e-05, 1.52587890625e-05, 5.7220458984375e-05, 9.918212890625e-05, 0.000141143798828125, 0.00018310546875, 0.000225067138671875, 0.00026702880859375, 0.000308990478515625, 0.0003509521484375, 0.000392913818359375, 0.00043487548828125, 0.000476837158203125, 0.000518798828125, 0.000560760498046875, 0.00060272216796875, 0.000644683837890625, 0.0006866455078125, 0.000728607177734375, 0.00077056884765625, 0.000812530517578125, 0.0008544921875, 0.000896453857421875, 0.00093841552734375, 0.000980377197265625, 0.0010223388671875, 0.001064300537109375, 0.00110626220703125, 0.001148223876953125, 0.001190185546875, 0.001232147216796875, 0.00127410888671875, 0.001316070556640625, 0.0013580322265625, 0.001399993896484375, 0.00144195556640625, 0.001483917236328125, 0.00152587890625, 0.001567840576171875, 0.00160980224609375]}, "gradients/model.layers.8.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 4.0, 3.0, 3.0, 4.0, 9.0, 14.0, 13.0, 20.0, 38.0, 67.0, 119.0, 173.0, 335.0, 615.0, 1234.0, 2853.0, 7973.0, 25348.0, 140829.0, 10966193.0, 16777216.0, 1517785.0, 66895.0, 16274.0, 5557.0, 2153.0, 953.0, 487.0, 264.0, 148.0, 78.0, 44.0, 28.0, 14.0, 15.0, 12.0, 5.0, 5.0, 3.0, 3.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.0030975341796875, -0.003001689910888672, -0.0029058456420898438, -0.0028100013732910156, -0.0027141571044921875, -0.0026183128356933594, -0.0025224685668945312, -0.002426624298095703, -0.002330780029296875, -0.002234935760498047, -0.0021390914916992188, -0.0020432472229003906, -0.0019474029541015625, -0.0018515586853027344, -0.0017557144165039062, -0.0016598701477050781, -0.00156402587890625, -0.0014681816101074219, -0.0013723373413085938, -0.0012764930725097656, -0.0011806488037109375, -0.0010848045349121094, -0.0009889602661132812, -0.0008931159973144531, -0.000797271728515625, -0.0007014274597167969, -0.0006055831909179688, -0.0005097389221191406, -0.0004138946533203125, -0.0003180503845214844, -0.00022220611572265625, -0.00012636184692382812, -3.0517578125e-05, 6.532669067382812e-05, 0.00016117095947265625, 0.0002570152282714844, 0.0003528594970703125, 0.0004487037658691406, 0.0005445480346679688, 0.0006403923034667969, 0.000736236572265625, 0.0008320808410644531, 0.0009279251098632812, 0.0010237693786621094, 0.0011196136474609375, 0.0012154579162597656, 0.0013113021850585938, 0.0014071464538574219, 0.00150299072265625, 0.0015988349914550781, 0.0016946792602539062, 0.0017905235290527344, 0.0018863677978515625, 0.0019822120666503906, 0.0020780563354492188, 0.002173900604248047, 0.002269744873046875, 0.002365589141845703, 0.0024614334106445312, 0.0025572776794433594, 0.0026531219482421875, 0.0027489662170410156, 0.0028448104858398438, 0.002940654754638672, 0.0030364990234375]}, "gradients/model.layers.8.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 2.0, 0.0, 2.0, 2.0, 2.0, 5.0, 10.0, 17.0, 40.0, 90.0, 202.0, 269.0, 487.0, 573.0, 642.0, 549.0, 478.0, 322.0, 195.0, 97.0, 49.0, 24.0, 12.0, 8.0, 0.0, 2.0, 4.0, 1.0, 2.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0020904541015625, -0.002029895782470703, -0.0019693374633789062, -0.0019087791442871094, -0.0018482208251953125, -0.0017876625061035156, -0.0017271041870117188, -0.0016665458679199219, -0.001605987548828125, -0.0015454292297363281, -0.0014848709106445312, -0.0014243125915527344, -0.0013637542724609375, -0.0013031959533691406, -0.0012426376342773438, -0.0011820793151855469, -0.00112152099609375, -0.0010609626770019531, -0.0010004043579101562, -0.0009398460388183594, -0.0008792877197265625, -0.0008187294006347656, -0.0007581710815429688, -0.0006976127624511719, -0.000637054443359375, -0.0005764961242675781, -0.0005159378051757812, -0.0004553794860839844, -0.0003948211669921875, -0.0003342628479003906, -0.00027370452880859375, -0.00021314620971679688, -0.000152587890625, -9.202957153320312e-05, -3.147125244140625e-05, 2.9087066650390625e-05, 8.96453857421875e-05, 0.00015020370483398438, 0.00021076202392578125, 0.0002713203430175781, 0.000331878662109375, 0.0003924369812011719, 0.00045299530029296875, 0.0005135536193847656, 0.0005741119384765625, 0.0006346702575683594, 0.0006952285766601562, 0.0007557868957519531, 0.00081634521484375, 0.0008769035339355469, 0.0009374618530273438, 0.0009980201721191406, 0.0010585784912109375, 0.0011191368103027344, 0.0011796951293945312, 0.0012402534484863281, 0.001300811767578125, 0.0013613700866699219, 0.0014219284057617188, 0.0014824867248535156, 0.0015430450439453125, 0.0016036033630371094, 0.0016641616821289062, 0.0017247200012207031, 0.0017852783203125]}, "gradients/model.layers.8.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 1.0, 0.0, 1.0, 1.0, 2.0, 10.0, 16.0, 12.0, 22.0, 33.0, 58.0, 104.0, 172.0, 319.0, 570.0, 1216.0, 2397.0, 5117.0, 12273.0, 35894.0, 179027.0, 1862153.0, 9908619.0, 4272049.0, 405687.0, 60377.0, 17729.0, 7037.0, 3084.0, 1543.0, 790.0, 397.0, 203.0, 125.0, 66.0, 42.0, 30.0, 12.0, 9.0, 5.0, 4.0, 2.0, 2.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00335693359375, -0.003257274627685547, -0.0031576156616210938, -0.0030579566955566406, -0.0029582977294921875, -0.0028586387634277344, -0.0027589797973632812, -0.002659320831298828, -0.002559661865234375, -0.002460002899169922, -0.0023603439331054688, -0.0022606849670410156, -0.0021610260009765625, -0.0020613670349121094, -0.0019617080688476562, -0.0018620491027832031, -0.00176239013671875, -0.0016627311706542969, -0.0015630722045898438, -0.0014634132385253906, -0.0013637542724609375, -0.0012640953063964844, -0.0011644363403320312, -0.0010647773742675781, -0.000965118408203125, -0.0008654594421386719, -0.0007658004760742188, -0.0006661415100097656, -0.0005664825439453125, -0.0004668235778808594, -0.00036716461181640625, -0.0002675056457519531, -0.0001678466796875, -6.818771362304688e-05, 3.147125244140625e-05, 0.00013113021850585938, 0.0002307891845703125, 0.0003304481506347656, 0.00043010711669921875, 0.0005297660827636719, 0.000629425048828125, 0.0007290840148925781, 0.0008287429809570312, 0.0009284019470214844, 0.0010280609130859375, 0.0011277198791503906, 0.0012273788452148438, 0.0013270378112792969, 0.00142669677734375, 0.0015263557434082031, 0.0016260147094726562, 0.0017256736755371094, 0.0018253326416015625, 0.0019249916076660156, 0.0020246505737304688, 0.002124309539794922, 0.002223968505859375, 0.002323627471923828, 0.0024232864379882812, 0.0025229454040527344, 0.0026226043701171875, 0.0027222633361816406, 0.0028219223022460938, 0.002921581268310547, 0.003021240234375]}, "gradients/model.layers.8.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 3.0, 1.0, 0.0, 2.0, 1.0, 0.0, 7.0, 7.0, 7.0, 10.0, 20.0, 16.0, 17.0, 32.0, 26.0, 20.0, 44.0, 67.0, 69.0, 112.0, 137.0, 204.0, 339.0, 655.0, 1231.0, 2257.0, 4308.0, 15425.0, 2711938.0, 1437815.0, 10894.0, 3792.0, 2073.0, 1095.0, 543.0, 344.0, 214.0, 142.0, 103.0, 72.0, 36.0, 30.0, 46.0, 24.0, 26.0, 25.0, 19.0, 17.0, 8.0, 2.0, 5.0, 2.0, 6.0, 6.0, 0.0, 2.0, 2.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.0115966796875, -0.011233329772949219, -0.010869979858398438, -0.010506629943847656, -0.010143280029296875, -0.009779930114746094, -0.009416580200195312, -0.009053230285644531, -0.00868988037109375, -0.008326530456542969, -0.007963180541992188, -0.007599830627441406, -0.007236480712890625, -0.006873130798339844, -0.0065097808837890625, -0.006146430969238281, -0.0057830810546875, -0.005419731140136719, -0.0050563812255859375, -0.004693031311035156, -0.004329681396484375, -0.003966331481933594, -0.0036029815673828125, -0.0032396316528320312, -0.00287628173828125, -0.0025129318237304688, -0.0021495819091796875, -0.0017862319946289062, -0.001422882080078125, -0.0010595321655273438, -0.0006961822509765625, -0.00033283233642578125, 3.0517578125e-05, 0.00039386749267578125, 0.0007572174072265625, 0.0011205673217773438, 0.001483917236328125, 0.0018472671508789062, 0.0022106170654296875, 0.0025739669799804688, 0.00293731689453125, 0.0033006668090820312, 0.0036640167236328125, 0.004027366638183594, 0.004390716552734375, 0.004754066467285156, 0.0051174163818359375, 0.005480766296386719, 0.0058441162109375, 0.006207466125488281, 0.0065708160400390625, 0.006934165954589844, 0.007297515869140625, 0.007660865783691406, 0.008024215698242188, 0.008387565612792969, 0.00875091552734375, 0.009114265441894531, 0.009477615356445312, 0.009840965270996094, 0.010204315185546875, 0.010567665100097656, 0.010931015014648438, 0.011294364929199219, 0.01165771484375]}, "gradients/model.layers.8.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 3.0, 4.0, 0.0, 5.0, 1.0, 4.0, 4.0, 7.0, 4.0, 3.0, 6.0, 16.0, 22.0, 28.0, 50.0, 76.0, 130.0, 312.0, 1015.0, 4301.0, 42982.0, 4092009.0, 47239.0, 4370.0, 1006.0, 322.0, 145.0, 60.0, 54.0, 26.0, 27.0, 19.0, 5.0, 8.0, 9.0, 5.0, 7.0, 2.0, 2.0, 1.0, 1.0, 3.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 3.0], "bins": [-0.0032501220703125, -0.003155946731567383, -0.0030617713928222656, -0.0029675960540771484, -0.0028734207153320312, -0.002779245376586914, -0.002685070037841797, -0.0025908946990966797, -0.0024967193603515625, -0.0024025440216064453, -0.002308368682861328, -0.002214193344116211, -0.0021200180053710938, -0.0020258426666259766, -0.0019316673278808594, -0.0018374919891357422, -0.001743316650390625, -0.0016491413116455078, -0.0015549659729003906, -0.0014607906341552734, -0.0013666152954101562, -0.001272439956665039, -0.0011782646179199219, -0.0010840892791748047, -0.0009899139404296875, -0.0008957386016845703, -0.0008015632629394531, -0.0007073879241943359, -0.0006132125854492188, -0.0005190372467041016, -0.0004248619079589844, -0.0003306865692138672, -0.00023651123046875, -0.0001423358917236328, -4.8160552978515625e-05, 4.601478576660156e-05, 0.00014019012451171875, 0.00023436546325683594, 0.0003285408020019531, 0.0004227161407470703, 0.0005168914794921875, 0.0006110668182373047, 0.0007052421569824219, 0.0007994174957275391, 0.0008935928344726562, 0.0009877681732177734, 0.0010819435119628906, 0.0011761188507080078, 0.001270294189453125, 0.0013644695281982422, 0.0014586448669433594, 0.0015528202056884766, 0.0016469955444335938, 0.001741170883178711, 0.0018353462219238281, 0.0019295215606689453, 0.0020236968994140625, 0.0021178722381591797, 0.002212047576904297, 0.002306222915649414, 0.0024003982543945312, 0.0024945735931396484, 0.0025887489318847656, 0.002682924270629883, 0.002777099609375]}, "gradients/model.layers.8.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 4.0, 5.0, 6.0, 2.0, 7.0, 6.0, 14.0, 13.0, 14.0, 24.0, 23.0, 39.0, 55.0, 57.0, 95.0, 147.0, 237.0, 376.0, 815.0, 2030.0, 5832.0, 24032.0, 282859.0, 16176370.0, 251637.0, 23133.0, 5590.0, 1906.0, 781.0, 398.0, 221.0, 125.0, 96.0, 59.0, 47.0, 33.0, 20.0, 16.0, 17.0, 16.0, 12.0, 9.0, 10.0, 6.0, 3.0, 4.0, 2.0, 1.0, 3.0, 2.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.002471923828125, -0.0023851394653320312, -0.0022983551025390625, -0.0022115707397460938, -0.002124786376953125, -0.0020380020141601562, -0.0019512176513671875, -0.0018644332885742188, -0.00177764892578125, -0.0016908645629882812, -0.0016040802001953125, -0.0015172958374023438, -0.001430511474609375, -0.0013437271118164062, -0.0012569427490234375, -0.0011701583862304688, -0.0010833740234375, -0.0009965896606445312, -0.0009098052978515625, -0.0008230209350585938, -0.000736236572265625, -0.0006494522094726562, -0.0005626678466796875, -0.00047588348388671875, -0.00038909912109375, -0.00030231475830078125, -0.0002155303955078125, -0.00012874603271484375, -4.1961669921875e-05, 4.482269287109375e-05, 0.0001316070556640625, 0.00021839141845703125, 0.00030517578125, 0.00039196014404296875, 0.0004787445068359375, 0.0005655288696289062, 0.000652313232421875, 0.0007390975952148438, 0.0008258819580078125, 0.0009126663208007812, 0.00099945068359375, 0.0010862350463867188, 0.0011730194091796875, 0.0012598037719726562, 0.001346588134765625, 0.0014333724975585938, 0.0015201568603515625, 0.0016069412231445312, 0.0016937255859375, 0.0017805099487304688, 0.0018672943115234375, 0.0019540786743164062, 0.002040863037109375, 0.0021276473999023438, 0.0022144317626953125, 0.0023012161254882812, 0.00238800048828125, 0.0024747848510742188, 0.0025615692138671875, 0.0026483535766601562, 0.002735137939453125, 0.0028219223022460938, 0.0029087066650390625, 0.0029954910278320312, 0.003082275390625]}, "gradients/model.layers.8.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 3.0, 2.0, 3.0, 14.0, 89.0, 688.0, 2001.0, 1112.0, 163.0, 6.0, 3.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.004425048828125, -0.004337668418884277, -0.004250288009643555, -0.004162907600402832, -0.004075527191162109, -0.003988146781921387, -0.003900766372680664, -0.0038133859634399414, -0.0037260055541992188, -0.003638625144958496, -0.0035512447357177734, -0.0034638643264770508, -0.003376483917236328, -0.0032891035079956055, -0.003201723098754883, -0.00311434268951416, -0.0030269622802734375, -0.002939581871032715, -0.002852201461791992, -0.0027648210525512695, -0.002677440643310547, -0.0025900602340698242, -0.0025026798248291016, -0.002415299415588379, -0.0023279190063476562, -0.0022405385971069336, -0.002153158187866211, -0.0020657777786254883, -0.0019783973693847656, -0.001891016960144043, -0.0018036365509033203, -0.0017162561416625977, -0.001628875732421875, -0.0015414953231811523, -0.0014541149139404297, -0.001366734504699707, -0.0012793540954589844, -0.0011919736862182617, -0.001104593276977539, -0.0010172128677368164, -0.0009298324584960938, -0.0008424520492553711, -0.0007550716400146484, -0.0006676912307739258, -0.0005803108215332031, -0.0004929304122924805, -0.0004055500030517578, -0.00031816959381103516, -0.0002307891845703125, -0.00014340877532958984, -5.602836608886719e-05, 3.135204315185547e-05, 0.00011873245239257812, 0.00020611286163330078, 0.00029349327087402344, 0.0003808736801147461, 0.00046825408935546875, 0.0005556344985961914, 0.0006430149078369141, 0.0007303953170776367, 0.0008177757263183594, 0.000905156135559082, 0.0009925365447998047, 0.0010799169540405273, 0.00116729736328125]}, "gradients/model.layers.7.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [7.0, 4.0, 6.0, 6.0, 6.0, 14.0, 19.0, 18.0, 30.0, 33.0, 51.0, 55.0, 76.0, 93.0, 105.0, 98.0, 147.0, 206.0, 265.0, 742.0, 4689.0, 101881.0, 16777216.0, 16777216.0, 236133.0, 7148.0, 991.0, 317.0, 209.0, 131.0, 149.0, 109.0, 105.0, 72.0, 54.0, 66.0, 42.0, 30.0, 22.0, 16.0, 11.0, 7.0, 1.0, 4.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0028076171875, -0.002686023712158203, -0.0025644302368164062, -0.0024428367614746094, -0.0023212432861328125, -0.0021996498107910156, -0.0020780563354492188, -0.001956462860107422, -0.001834869384765625, -0.0017132759094238281, -0.0015916824340820312, -0.0014700889587402344, -0.0013484954833984375, -0.0012269020080566406, -0.0011053085327148438, -0.0009837150573730469, -0.00086212158203125, -0.0007405281066894531, -0.0006189346313476562, -0.0004973411560058594, -0.0003757476806640625, -0.0002541542053222656, -0.00013256072998046875, -1.0967254638671875e-05, 0.000110626220703125, 0.00023221969604492188, 0.00035381317138671875, 0.0004754066467285156, 0.0005970001220703125, 0.0007185935974121094, 0.0008401870727539062, 0.0009617805480957031, 0.0010833740234375, 0.0012049674987792969, 0.0013265609741210938, 0.0014481544494628906, 0.0015697479248046875, 0.0016913414001464844, 0.0018129348754882812, 0.0019345283508300781, 0.002056121826171875, 0.002177715301513672, 0.0022993087768554688, 0.0024209022521972656, 0.0025424957275390625, 0.0026640892028808594, 0.0027856826782226562, 0.002907276153564453, 0.00302886962890625, 0.003150463104248047, 0.0032720565795898438, 0.0033936500549316406, 0.0035152435302734375, 0.0036368370056152344, 0.0037584304809570312, 0.003880023956298828, 0.004001617431640625, 0.004123210906982422, 0.004244804382324219, 0.004366397857666016, 0.0044879913330078125, 0.004609584808349609, 0.004731178283691406, 0.004852771759033203, 0.004974365234375]}, "gradients/model.layers.7.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 1.0, 0.0, 0.0, 1.0, 3.0, 1.0, 1.0, 3.0, 4.0, 3.0, 3.0, 6.0, 13.0, 28.0, 30.0, 48.0, 55.0, 106.0, 159.0, 256.0, 386.0, 646.0, 1208.0, 2045.0, 3585.0, 6455.0, 12485.0, 28333.0, 77378.0, 460308.0, 7008526.0, 16777216.0, 16777216.0, 1865346.0, 173458.0, 46301.0, 18882.0, 9198.0, 4898.0, 2765.0, 1490.0, 839.0, 514.0, 293.0, 186.0, 135.0, 94.0, 50.0, 27.0, 20.0, 15.0, 7.0, 4.0, 6.0, 6.0, 9.0, 0.0, 3.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.0016326904296875, -0.0015828609466552734, -0.0015330314636230469, -0.0014832019805908203, -0.0014333724975585938, -0.0013835430145263672, -0.0013337135314941406, -0.001283884048461914, -0.0012340545654296875, -0.001184225082397461, -0.0011343955993652344, -0.0010845661163330078, -0.0010347366333007812, -0.0009849071502685547, -0.0009350776672363281, -0.0008852481842041016, -0.000835418701171875, -0.0007855892181396484, -0.0007357597351074219, -0.0006859302520751953, -0.0006361007690429688, -0.0005862712860107422, -0.0005364418029785156, -0.00048661231994628906, -0.0004367828369140625, -0.00038695335388183594, -0.0003371238708496094, -0.0002872943878173828, -0.00023746490478515625, -0.0001876354217529297, -0.00013780593872070312, -8.797645568847656e-05, -3.814697265625e-05, 1.1682510375976562e-05, 6.151199340820312e-05, 0.00011134147644042969, 0.00016117095947265625, 0.0002110004425048828, 0.0002608299255371094, 0.00031065940856933594, 0.0003604888916015625, 0.00041031837463378906, 0.0004601478576660156, 0.0005099773406982422, 0.0005598068237304688, 0.0006096363067626953, 0.0006594657897949219, 0.0007092952728271484, 0.000759124755859375, 0.0008089542388916016, 0.0008587837219238281, 0.0009086132049560547, 0.0009584426879882812, 0.0010082721710205078, 0.0010581016540527344, 0.001107931137084961, 0.0011577606201171875, 0.001207590103149414, 0.0012574195861816406, 0.0013072490692138672, 0.0013570785522460938, 0.0014069080352783203, 0.0014567375183105469, 0.0015065670013427734, 0.001556396484375]}, "gradients/model.layers.7.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 12.0, 64.0, 333.0, 1070.0, 1414.0, 856.0, 254.0, 51.0, 9.0, 7.0, 4.0, 3.0, 2.0, 4.0, 3.0, 1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00151824951171875, -0.0013715028762817383, -0.0012247562408447266, -0.0010780096054077148, -0.0009312629699707031, -0.0007845163345336914, -0.0006377696990966797, -0.000491023063659668, -0.00034427642822265625, -0.00019752979278564453, -5.078315734863281e-05, 9.59634780883789e-05, 0.00024271011352539062, 0.00038945674896240234, 0.0005362033843994141, 0.0006829500198364258, 0.0008296966552734375, 0.0009764432907104492, 0.001123189926147461, 0.0012699365615844727, 0.0014166831970214844, 0.001563429832458496, 0.0017101764678955078, 0.0018569231033325195, 0.0020036697387695312, 0.002150416374206543, 0.0022971630096435547, 0.0024439096450805664, 0.002590656280517578, 0.00273740291595459, 0.0028841495513916016, 0.0030308961868286133, 0.003177642822265625, 0.0033243894577026367, 0.0034711360931396484, 0.00361788272857666, 0.003764629364013672, 0.003911375999450684, 0.004058122634887695, 0.004204869270324707, 0.004351615905761719, 0.0044983625411987305, 0.004645109176635742, 0.004791855812072754, 0.004938602447509766, 0.005085349082946777, 0.005232095718383789, 0.005378842353820801, 0.0055255889892578125, 0.005672335624694824, 0.005819082260131836, 0.005965828895568848, 0.006112575531005859, 0.006259322166442871, 0.006406068801879883, 0.0065528154373168945, 0.006699562072753906, 0.006846308708190918, 0.00699305534362793, 0.007139801979064941, 0.007286548614501953, 0.007433295249938965, 0.0075800418853759766, 0.007726788520812988, 0.00787353515625]}, "gradients/model.layers.7.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 2.0, 2.0, 3.0, 2.0, 6.0, 10.0, 9.0, 15.0, 17.0, 33.0, 59.0, 93.0, 161.0, 267.0, 449.0, 850.0, 1611.0, 3038.0, 6155.0, 13877.0, 34430.0, 136737.0, 1049006.0, 6785413.0, 7291427.0, 1230997.0, 155981.0, 38520.0, 14302.0, 6628.0, 3306.0, 1796.0, 906.0, 458.0, 248.0, 144.0, 81.0, 65.0, 36.0, 20.0, 11.0, 16.0, 8.0, 6.0, 4.0, 2.0, 1.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.003143310546875, -0.0030584335327148438, -0.0029735565185546875, -0.0028886795043945312, -0.002803802490234375, -0.0027189254760742188, -0.0026340484619140625, -0.0025491714477539062, -0.00246429443359375, -0.0023794174194335938, -0.0022945404052734375, -0.0022096633911132812, -0.002124786376953125, -0.0020399093627929688, -0.0019550323486328125, -0.0018701553344726562, -0.0017852783203125, -0.0017004013061523438, -0.0016155242919921875, -0.0015306472778320312, -0.001445770263671875, -0.0013608932495117188, -0.0012760162353515625, -0.0011911392211914062, -0.00110626220703125, -0.0010213851928710938, -0.0009365081787109375, -0.0008516311645507812, -0.000766754150390625, -0.0006818771362304688, -0.0005970001220703125, -0.0005121231079101562, -0.00042724609375, -0.00034236907958984375, -0.0002574920654296875, -0.00017261505126953125, -8.7738037109375e-05, -2.86102294921875e-06, 8.20159912109375e-05, 0.00016689300537109375, 0.00025177001953125, 0.00033664703369140625, 0.0004215240478515625, 0.0005064010620117188, 0.000591278076171875, 0.0006761550903320312, 0.0007610321044921875, 0.0008459091186523438, 0.0009307861328125, 0.0010156631469726562, 0.0011005401611328125, 0.0011854171752929688, 0.001270294189453125, 0.0013551712036132812, 0.0014400482177734375, 0.0015249252319335938, 0.00160980224609375, 0.0016946792602539062, 0.0017795562744140625, 0.0018644332885742188, 0.001949310302734375, 0.0020341873168945312, 0.0021190643310546875, 0.0022039413452148438, 0.002288818359375]}, "gradients/model.layers.7.self_attn.v_proj.weight": {"_type": "histogram", "values": [3.0, 1.0, 1.0, 1.0, 1.0, 3.0, 4.0, 4.0, 4.0, 3.0, 8.0, 10.0, 13.0, 20.0, 22.0, 30.0, 36.0, 45.0, 54.0, 68.0, 102.0, 144.0, 260.0, 418.0, 673.0, 1277.0, 2726.0, 6420.0, 114183.0, 3921887.0, 133066.0, 6769.0, 2774.0, 1395.0, 731.0, 347.0, 235.0, 157.0, 128.0, 66.0, 54.0, 25.0, 24.0, 24.0, 20.0, 12.0, 16.0, 7.0, 10.0, 7.0, 3.0, 2.0, 1.0, 1.0, 1.0, 0.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 3.0], "bins": [-0.0101318359375, -0.00978851318359375, -0.0094451904296875, -0.00910186767578125, -0.008758544921875, -0.00841522216796875, -0.0080718994140625, -0.00772857666015625, -0.00738525390625, -0.00704193115234375, -0.0066986083984375, -0.00635528564453125, -0.006011962890625, -0.00566864013671875, -0.0053253173828125, -0.00498199462890625, -0.004638671875, -0.00429534912109375, -0.0039520263671875, -0.00360870361328125, -0.003265380859375, -0.00292205810546875, -0.0025787353515625, -0.00223541259765625, -0.00189208984375, -0.00154876708984375, -0.0012054443359375, -0.00086212158203125, -0.000518798828125, -0.00017547607421875, 0.0001678466796875, 0.00051116943359375, 0.0008544921875, 0.00119781494140625, 0.0015411376953125, 0.00188446044921875, 0.002227783203125, 0.00257110595703125, 0.0029144287109375, 0.00325775146484375, 0.00360107421875, 0.00394439697265625, 0.0042877197265625, 0.00463104248046875, 0.004974365234375, 0.00531768798828125, 0.0056610107421875, 0.00600433349609375, 0.00634765625, 0.00669097900390625, 0.0070343017578125, 0.00737762451171875, 0.007720947265625, 0.00806427001953125, 0.0084075927734375, 0.00875091552734375, 0.00909423828125, 0.00943756103515625, 0.0097808837890625, 0.01012420654296875, 0.010467529296875, 0.01081085205078125, 0.0111541748046875, 0.01149749755859375, 0.0118408203125]}, "gradients/model.layers.7.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 4.0, 4.0, 5.0, 3.0, 5.0, 7.0, 7.0, 12.0, 12.0, 12.0, 26.0, 23.0, 37.0, 61.0, 97.0, 159.0, 335.0, 899.0, 3457.0, 19823.0, 3451240.0, 701991.0, 12197.0, 2437.0, 707.0, 286.0, 146.0, 78.0, 64.0, 35.0, 28.0, 21.0, 15.0, 10.0, 15.0, 11.0, 4.0, 2.0, 3.0, 3.0, 3.0, 4.0, 1.0, 0.0, 1.0, 1.0, 2.0, 1.0, 0.0, 1.0], "bins": [-0.003143310546875, -0.0030531883239746094, -0.0029630661010742188, -0.002872943878173828, -0.0027828216552734375, -0.002692699432373047, -0.0026025772094726562, -0.0025124549865722656, -0.002422332763671875, -0.0023322105407714844, -0.0022420883178710938, -0.002151966094970703, -0.0020618438720703125, -0.001971721649169922, -0.0018815994262695312, -0.0017914772033691406, -0.00170135498046875, -0.0016112327575683594, -0.0015211105346679688, -0.0014309883117675781, -0.0013408660888671875, -0.0012507438659667969, -0.0011606216430664062, -0.0010704994201660156, -0.000980377197265625, -0.0008902549743652344, -0.0008001327514648438, -0.0007100105285644531, -0.0006198883056640625, -0.0005297660827636719, -0.00043964385986328125, -0.0003495216369628906, -0.0002593994140625, -0.00016927719116210938, -7.915496826171875e-05, 1.0967254638671875e-05, 0.0001010894775390625, 0.00019121170043945312, 0.00028133392333984375, 0.0003714561462402344, 0.000461578369140625, 0.0005517005920410156, 0.0006418228149414062, 0.0007319450378417969, 0.0008220672607421875, 0.0009121894836425781, 0.0010023117065429688, 0.0010924339294433594, 0.00118255615234375, 0.0012726783752441406, 0.0013628005981445312, 0.0014529228210449219, 0.0015430450439453125, 0.0016331672668457031, 0.0017232894897460938, 0.0018134117126464844, 0.001903533935546875, 0.0019936561584472656, 0.0020837783813476562, 0.002173900604248047, 0.0022640228271484375, 0.002354145050048828, 0.0024442672729492188, 0.0025343894958496094, 0.00262451171875]}, "gradients/model.layers.7.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 3.0, 1.0, 5.0, 5.0, 5.0, 3.0, 7.0, 8.0, 9.0, 15.0, 29.0, 28.0, 43.0, 68.0, 81.0, 133.0, 265.0, 551.0, 1586.0, 6495.0, 44241.0, 11997618.0, 4683210.0, 34776.0, 5492.0, 1351.0, 516.0, 235.0, 127.0, 105.0, 44.0, 41.0, 41.0, 15.0, 14.0, 17.0, 7.0, 4.0, 3.0, 2.0, 2.0, 3.0, 3.0, 1.0, 2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0038604736328125, -0.0037467479705810547, -0.0036330223083496094, -0.003519296646118164, -0.0034055709838867188, -0.0032918453216552734, -0.003178119659423828, -0.003064393997192383, -0.0029506683349609375, -0.002836942672729492, -0.002723217010498047, -0.0026094913482666016, -0.0024957656860351562, -0.002382040023803711, -0.0022683143615722656, -0.0021545886993408203, -0.002040863037109375, -0.0019271373748779297, -0.0018134117126464844, -0.001699686050415039, -0.0015859603881835938, -0.0014722347259521484, -0.0013585090637207031, -0.0012447834014892578, -0.0011310577392578125, -0.0010173320770263672, -0.0009036064147949219, -0.0007898807525634766, -0.0006761550903320312, -0.0005624294281005859, -0.0004487037658691406, -0.0003349781036376953, -0.00022125244140625, -0.00010752677917480469, 6.198883056640625e-06, 0.00011992454528808594, 0.00023365020751953125, 0.00034737586975097656, 0.0004611015319824219, 0.0005748271942138672, 0.0006885528564453125, 0.0008022785186767578, 0.0009160041809082031, 0.0010297298431396484, 0.0011434555053710938, 0.001257181167602539, 0.0013709068298339844, 0.0014846324920654297, 0.001598358154296875, 0.0017120838165283203, 0.0018258094787597656, 0.001939535140991211, 0.0020532608032226562, 0.0021669864654541016, 0.002280712127685547, 0.002394437789916992, 0.0025081634521484375, 0.002621889114379883, 0.002735614776611328, 0.0028493404388427734, 0.0029630661010742188, 0.003076791763305664, 0.0031905174255371094, 0.0033042430877685547, 0.00341796875]}, "gradients/model.layers.7.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 3.0, 4.0, 8.0, 5.0, 15.0, 719.0, 3204.0, 127.0, 4.0, 2.0], "bins": [-0.01202392578125, -0.011824727058410645, -0.011625528335571289, -0.011426329612731934, -0.011227130889892578, -0.011027932167053223, -0.010828733444213867, -0.010629534721374512, -0.010430335998535156, -0.0102311372756958, -0.010031938552856445, -0.00983273983001709, -0.009633541107177734, -0.009434342384338379, -0.009235143661499023, -0.009035944938659668, -0.008836746215820312, -0.008637547492980957, -0.008438348770141602, -0.008239150047302246, -0.00803995132446289, -0.007840752601623535, -0.00764155387878418, -0.007442355155944824, -0.007243156433105469, -0.007043957710266113, -0.006844758987426758, -0.006645560264587402, -0.006446361541748047, -0.006247162818908691, -0.006047964096069336, -0.0058487653732299805, -0.005649566650390625, -0.0054503679275512695, -0.005251169204711914, -0.005051970481872559, -0.004852771759033203, -0.004653573036193848, -0.004454374313354492, -0.004255175590515137, -0.004055976867675781, -0.0038567781448364258, -0.0036575794219970703, -0.003458380699157715, -0.0032591819763183594, -0.003059983253479004, -0.0028607845306396484, -0.002661585807800293, -0.0024623870849609375, -0.002263188362121582, -0.0020639896392822266, -0.001864790916442871, -0.0016655921936035156, -0.0014663934707641602, -0.0012671947479248047, -0.0010679960250854492, -0.0008687973022460938, -0.0006695985794067383, -0.0004703998565673828, -0.00027120113372802734, -7.200241088867188e-05, 0.0001271963119506836, 0.00032639503479003906, 0.0005255937576293945, 0.00072479248046875]}, "gradients/model.layers.6.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 1.0, 3.0, 2.0, 6.0, 5.0, 11.0, 14.0, 26.0, 18.0, 35.0, 39.0, 73.0, 71.0, 130.0, 110.0, 145.0, 235.0, 309.0, 629.0, 3204.0, 121932.0, 16777216.0, 8303066.0, 19047.0, 1350.0, 428.0, 260.0, 183.0, 146.0, 124.0, 111.0, 81.0, 73.0, 44.0, 31.0, 13.0, 17.0, 8.0, 6.0, 5.0, 1.0, 2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.003814697265625, -0.003647327423095703, -0.0034799575805664062, -0.0033125877380371094, -0.0031452178955078125, -0.0029778480529785156, -0.0028104782104492188, -0.002643108367919922, -0.002475738525390625, -0.002308368682861328, -0.0021409988403320312, -0.0019736289978027344, -0.0018062591552734375, -0.0016388893127441406, -0.0014715194702148438, -0.0013041496276855469, -0.00113677978515625, -0.0009694099426269531, -0.0008020401000976562, -0.0006346702575683594, -0.0004673004150390625, -0.0002999305725097656, -0.00013256072998046875, 3.4809112548828125e-05, 0.000202178955078125, 0.0003695487976074219, 0.0005369186401367188, 0.0007042884826660156, 0.0008716583251953125, 0.0010390281677246094, 0.0012063980102539062, 0.0013737678527832031, 0.0015411376953125, 0.0017085075378417969, 0.0018758773803710938, 0.0020432472229003906, 0.0022106170654296875, 0.0023779869079589844, 0.0025453567504882812, 0.002712726593017578, 0.002880096435546875, 0.003047466278076172, 0.0032148361206054688, 0.0033822059631347656, 0.0035495758056640625, 0.0037169456481933594, 0.0038843154907226562, 0.004051685333251953, 0.00421905517578125, 0.004386425018310547, 0.004553794860839844, 0.004721164703369141, 0.0048885345458984375, 0.005055904388427734, 0.005223274230957031, 0.005390644073486328, 0.005558013916015625, 0.005725383758544922, 0.005892753601074219, 0.006060123443603516, 0.0062274932861328125, 0.006394863128662109, 0.006562232971191406, 0.006729602813720703, 0.00689697265625]}, "gradients/model.layers.6.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 2.0, 2.0, 3.0, 2.0, 1.0, 2.0, 4.0, 3.0, 3.0, 5.0, 11.0, 9.0, 9.0, 22.0, 20.0, 39.0, 71.0, 154.0, 243.0, 462.0, 1005.0, 2299.0, 5171.0, 13787.0, 47380.0, 388187.0, 16777216.0, 16777216.0, 4104936.0, 152074.0, 28406.0, 9620.0, 3839.0, 1563.0, 781.0, 384.0, 185.0, 113.0, 54.0, 30.0, 17.0, 19.0, 7.0, 7.0, 6.0, 5.0, 2.0, 7.0, 2.0, 1.0, 2.0, 0.0, 3.0, 1.0, 2.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.0023040771484375, -0.002232789993286133, -0.0021615028381347656, -0.0020902156829833984, -0.0020189285278320312, -0.001947641372680664, -0.0018763542175292969, -0.0018050670623779297, -0.0017337799072265625, -0.0016624927520751953, -0.0015912055969238281, -0.001519918441772461, -0.0014486312866210938, -0.0013773441314697266, -0.0013060569763183594, -0.0012347698211669922, -0.001163482666015625, -0.0010921955108642578, -0.0010209083557128906, -0.0009496212005615234, -0.0008783340454101562, -0.0008070468902587891, -0.0007357597351074219, -0.0006644725799560547, -0.0005931854248046875, -0.0005218982696533203, -0.0004506111145019531, -0.00037932395935058594, -0.00030803680419921875, -0.00023674964904785156, -0.00016546249389648438, -9.417533874511719e-05, -2.288818359375e-05, 4.839897155761719e-05, 0.00011968612670898438, 0.00019097328186035156, 0.00026226043701171875, 0.00033354759216308594, 0.0004048347473144531, 0.0004761219024658203, 0.0005474090576171875, 0.0006186962127685547, 0.0006899833679199219, 0.0007612705230712891, 0.0008325576782226562, 0.0009038448333740234, 0.0009751319885253906, 0.0010464191436767578, 0.001117706298828125, 0.0011889934539794922, 0.0012602806091308594, 0.0013315677642822266, 0.0014028549194335938, 0.001474142074584961, 0.0015454292297363281, 0.0016167163848876953, 0.0016880035400390625, 0.0017592906951904297, 0.0018305778503417969, 0.001901865005493164, 0.0019731521606445312, 0.0020444393157958984, 0.0021157264709472656, 0.002187013626098633, 0.00225830078125]}, "gradients/model.layers.6.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 3.0, 6.0, 2.0, 4.0, 18.0, 38.0, 77.0, 140.0, 210.0, 387.0, 507.0, 626.0, 598.0, 536.0, 383.0, 244.0, 139.0, 71.0, 46.0, 22.0, 7.0, 7.0, 3.0, 0.0, 3.0, 0.0, 3.0, 0.0, 0.0, 3.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00189971923828125, -0.0018383264541625977, -0.0017769336700439453, -0.001715540885925293, -0.0016541481018066406, -0.0015927553176879883, -0.001531362533569336, -0.0014699697494506836, -0.0014085769653320312, -0.001347184181213379, -0.0012857913970947266, -0.0012243986129760742, -0.0011630058288574219, -0.0011016130447387695, -0.0010402202606201172, -0.0009788274765014648, -0.0009174346923828125, -0.0008560419082641602, -0.0007946491241455078, -0.0007332563400268555, -0.0006718635559082031, -0.0006104707717895508, -0.0005490779876708984, -0.0004876852035522461, -0.00042629241943359375, -0.0003648996353149414, -0.00030350685119628906, -0.00024211406707763672, -0.00018072128295898438, -0.00011932849884033203, -5.793571472167969e-05, 3.4570693969726562e-06, 6.4849853515625e-05, 0.00012624263763427734, 0.0001876354217529297, 0.00024902820587158203, 0.0003104209899902344, 0.0003718137741088867, 0.00043320655822753906, 0.0004945993423461914, 0.0005559921264648438, 0.0006173849105834961, 0.0006787776947021484, 0.0007401704788208008, 0.0008015632629394531, 0.0008629560470581055, 0.0009243488311767578, 0.0009857416152954102, 0.0010471343994140625, 0.0011085271835327148, 0.0011699199676513672, 0.0012313127517700195, 0.0012927055358886719, 0.0013540983200073242, 0.0014154911041259766, 0.001476883888244629, 0.0015382766723632812, 0.0015996694564819336, 0.001661062240600586, 0.0017224550247192383, 0.0017838478088378906, 0.001845240592956543, 0.0019066333770751953, 0.0019680261611938477, 0.0020294189453125]}, "gradients/model.layers.6.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 1.0, 2.0, 3.0, 5.0, 6.0, 11.0, 18.0, 20.0, 45.0, 79.0, 124.0, 215.0, 356.0, 606.0, 1276.0, 2215.0, 4470.0, 9472.0, 24056.0, 89232.0, 586104.0, 4394412.0, 9248159.0, 2063353.0, 272886.0, 49935.0, 15990.0, 6916.0, 3313.0, 1739.0, 955.0, 520.0, 299.0, 168.0, 92.0, 49.0, 38.0, 21.0, 16.0, 10.0, 5.0, 5.0, 0.0, 3.0, 4.0, 1.0, 1.0, 0.0, 2.0, 1.0, 0.0, 0.0, 0.0, 2.0, 0.0, 1.0], "bins": [-0.0023193359375, -0.0022428035736083984, -0.002166271209716797, -0.0020897388458251953, -0.0020132064819335938, -0.0019366741180419922, -0.0018601417541503906, -0.001783609390258789, -0.0017070770263671875, -0.001630544662475586, -0.0015540122985839844, -0.0014774799346923828, -0.0014009475708007812, -0.0013244152069091797, -0.0012478828430175781, -0.0011713504791259766, -0.001094818115234375, -0.0010182857513427734, -0.0009417533874511719, -0.0008652210235595703, -0.0007886886596679688, -0.0007121562957763672, -0.0006356239318847656, -0.0005590915679931641, -0.0004825592041015625, -0.00040602684020996094, -0.0003294944763183594, -0.0002529621124267578, -0.00017642974853515625, -9.989738464355469e-05, -2.3365020751953125e-05, 5.316734313964844e-05, 0.00012969970703125, 0.00020623207092285156, 0.0002827644348144531, 0.0003592967987060547, 0.00043582916259765625, 0.0005123615264892578, 0.0005888938903808594, 0.0006654262542724609, 0.0007419586181640625, 0.0008184909820556641, 0.0008950233459472656, 0.0009715557098388672, 0.0010480880737304688, 0.0011246204376220703, 0.0012011528015136719, 0.0012776851654052734, 0.001354217529296875, 0.0014307498931884766, 0.0015072822570800781, 0.0015838146209716797, 0.0016603469848632812, 0.0017368793487548828, 0.0018134117126464844, 0.001889944076538086, 0.0019664764404296875, 0.002043008804321289, 0.0021195411682128906, 0.002196073532104492, 0.0022726058959960938, 0.0023491382598876953, 0.002425670623779297, 0.0025022029876708984, 0.0025787353515625]}, "gradients/model.layers.6.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 2.0, 2.0, 1.0, 1.0, 0.0, 1.0, 4.0, 2.0, 8.0, 6.0, 3.0, 8.0, 9.0, 16.0, 22.0, 23.0, 40.0, 32.0, 50.0, 84.0, 83.0, 111.0, 173.0, 352.0, 715.0, 1660.0, 4957.0, 49631.0, 4073620.0, 54195.0, 5040.0, 1802.0, 627.0, 326.0, 179.0, 124.0, 95.0, 75.0, 56.0, 39.0, 28.0, 20.0, 13.0, 13.0, 10.0, 11.0, 7.0, 5.0, 5.0, 2.0, 3.0, 2.0, 2.0, 2.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.01336669921875, -0.012913703918457031, -0.012460708618164062, -0.012007713317871094, -0.011554718017578125, -0.011101722717285156, -0.010648727416992188, -0.010195732116699219, -0.00974273681640625, -0.009289741516113281, -0.008836746215820312, -0.008383750915527344, -0.007930755615234375, -0.007477760314941406, -0.0070247650146484375, -0.006571769714355469, -0.0061187744140625, -0.005665779113769531, -0.0052127838134765625, -0.004759788513183594, -0.004306793212890625, -0.0038537979125976562, -0.0034008026123046875, -0.0029478073120117188, -0.00249481201171875, -0.0020418167114257812, -0.0015888214111328125, -0.0011358261108398438, -0.000682830810546875, -0.00022983551025390625, 0.0002231597900390625, 0.0006761550903320312, 0.001129150390625, 0.0015821456909179688, 0.0020351409912109375, 0.0024881362915039062, 0.002941131591796875, 0.0033941268920898438, 0.0038471221923828125, 0.004300117492675781, 0.00475311279296875, 0.005206108093261719, 0.0056591033935546875, 0.006112098693847656, 0.006565093994140625, 0.007018089294433594, 0.0074710845947265625, 0.007924079895019531, 0.0083770751953125, 0.008830070495605469, 0.009283065795898438, 0.009736061096191406, 0.010189056396484375, 0.010642051696777344, 0.011095046997070312, 0.011548042297363281, 0.01200103759765625, 0.012454032897949219, 0.012907028198242188, 0.013360023498535156, 0.013813018798828125, 0.014266014099121094, 0.014719009399414062, 0.015172004699707031, 0.015625]}, "gradients/model.layers.6.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 1.0, 2.0, 6.0, 2.0, 0.0, 0.0, 3.0, 5.0, 2.0, 6.0, 13.0, 9.0, 12.0, 8.0, 4.0, 17.0, 19.0, 24.0, 36.0, 32.0, 60.0, 79.0, 91.0, 143.0, 264.0, 520.0, 1365.0, 4208.0, 22917.0, 3188748.0, 954226.0, 15672.0, 3454.0, 1137.0, 471.0, 223.0, 134.0, 85.0, 69.0, 46.0, 48.0, 17.0, 26.0, 13.0, 14.0, 8.0, 11.0, 9.0, 10.0, 2.0, 5.0, 4.0, 1.0, 7.0, 2.0, 4.0, 1.0, 0.0, 2.0, 3.0, 0.0, 2.0], "bins": [-0.0027313232421875, -0.002645730972290039, -0.002560138702392578, -0.002474546432495117, -0.0023889541625976562, -0.0023033618927001953, -0.0022177696228027344, -0.0021321773529052734, -0.0020465850830078125, -0.0019609928131103516, -0.0018754005432128906, -0.0017898082733154297, -0.0017042160034179688, -0.0016186237335205078, -0.0015330314636230469, -0.001447439193725586, -0.001361846923828125, -0.001276254653930664, -0.0011906623840332031, -0.0011050701141357422, -0.0010194778442382812, -0.0009338855743408203, -0.0008482933044433594, -0.0007627010345458984, -0.0006771087646484375, -0.0005915164947509766, -0.0005059242248535156, -0.0004203319549560547, -0.00033473968505859375, -0.0002491474151611328, -0.00016355514526367188, -7.796287536621094e-05, 7.62939453125e-06, 9.322166442871094e-05, 0.00017881393432617188, 0.0002644062042236328, 0.00034999847412109375, 0.0004355907440185547, 0.0005211830139160156, 0.0006067752838134766, 0.0006923675537109375, 0.0007779598236083984, 0.0008635520935058594, 0.0009491443634033203, 0.0010347366333007812, 0.0011203289031982422, 0.0012059211730957031, 0.001291513442993164, 0.001377105712890625, 0.001462697982788086, 0.0015482902526855469, 0.0016338825225830078, 0.0017194747924804688, 0.0018050670623779297, 0.0018906593322753906, 0.0019762516021728516, 0.0020618438720703125, 0.0021474361419677734, 0.0022330284118652344, 0.0023186206817626953, 0.0024042129516601562, 0.002489805221557617, 0.002575397491455078, 0.002660989761352539, 0.00274658203125]}, "gradients/model.layers.6.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 0.0, 1.0, 2.0, 0.0, 4.0, 5.0, 6.0, 6.0, 4.0, 8.0, 6.0, 22.0, 26.0, 30.0, 45.0, 66.0, 100.0, 138.0, 225.0, 454.0, 928.0, 2283.0, 7825.0, 50889.0, 14344318.0, 2327942.0, 32325.0, 5943.0, 1921.0, 739.0, 359.0, 181.0, 114.0, 92.0, 46.0, 33.0, 25.0, 22.0, 17.0, 10.0, 13.0, 11.0, 2.0, 6.0, 5.0, 5.0, 4.0, 1.0, 0.0, 1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.003204345703125, -0.0031006336212158203, -0.0029969215393066406, -0.002893209457397461, -0.0027894973754882812, -0.0026857852935791016, -0.002582073211669922, -0.002478361129760742, -0.0023746490478515625, -0.002270936965942383, -0.002167224884033203, -0.0020635128021240234, -0.0019598007202148438, -0.001856088638305664, -0.0017523765563964844, -0.0016486644744873047, -0.001544952392578125, -0.0014412403106689453, -0.0013375282287597656, -0.001233816146850586, -0.0011301040649414062, -0.0010263919830322266, -0.0009226799011230469, -0.0008189678192138672, -0.0007152557373046875, -0.0006115436553955078, -0.0005078315734863281, -0.00040411949157714844, -0.00030040740966796875, -0.00019669532775878906, -9.298324584960938e-05, 1.0728836059570312e-05, 0.00011444091796875, 0.0002181529998779297, 0.0003218650817871094, 0.00042557716369628906, 0.0005292892456054688, 0.0006330013275146484, 0.0007367134094238281, 0.0008404254913330078, 0.0009441375732421875, 0.0010478496551513672, 0.0011515617370605469, 0.0012552738189697266, 0.0013589859008789062, 0.001462697982788086, 0.0015664100646972656, 0.0016701221466064453, 0.001773834228515625, 0.0018775463104248047, 0.0019812583923339844, 0.002084970474243164, 0.0021886825561523438, 0.0022923946380615234, 0.002396106719970703, 0.002499818801879883, 0.0026035308837890625, 0.002707242965698242, 0.002810955047607422, 0.0029146671295166016, 0.0030183792114257812, 0.003122091293334961, 0.0032258033752441406, 0.0033295154571533203, 0.0034332275390625]}, "gradients/model.layers.6.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 3.0, 4.0, 25.0, 519.0, 3457.0, 80.0, 3.0], "bins": [-0.01300048828125, -0.012788653373718262, -0.012576818466186523, -0.012364983558654785, -0.012153148651123047, -0.011941313743591309, -0.01172947883605957, -0.011517643928527832, -0.011305809020996094, -0.011093974113464355, -0.010882139205932617, -0.010670304298400879, -0.01045846939086914, -0.010246634483337402, -0.010034799575805664, -0.009822964668273926, -0.009611129760742188, -0.00939929485321045, -0.009187459945678711, -0.008975625038146973, -0.008763790130615234, -0.008551955223083496, -0.008340120315551758, -0.00812828540802002, -0.007916450500488281, -0.007704615592956543, -0.007492780685424805, -0.007280945777893066, -0.007069110870361328, -0.00685727596282959, -0.0066454410552978516, -0.006433606147766113, -0.006221771240234375, -0.006009936332702637, -0.0057981014251708984, -0.00558626651763916, -0.005374431610107422, -0.005162596702575684, -0.004950761795043945, -0.004738926887512207, -0.004527091979980469, -0.0043152570724487305, -0.004103422164916992, -0.003891587257385254, -0.0036797523498535156, -0.0034679174423217773, -0.003256082534790039, -0.0030442476272583008, -0.0028324127197265625, -0.0026205778121948242, -0.002408742904663086, -0.0021969079971313477, -0.0019850730895996094, -0.001773238182067871, -0.0015614032745361328, -0.0013495683670043945, -0.0011377334594726562, -0.000925898551940918, -0.0007140636444091797, -0.0005022287368774414, -0.0002903938293457031, -7.855892181396484e-05, 0.00013327598571777344, 0.0003451108932495117, 0.00055694580078125]}, "gradients/model.layers.5.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0, 3.0, 6.0, 13.0, 14.0, 14.0, 20.0, 21.0, 43.0, 64.0, 87.0, 134.0, 155.0, 161.0, 254.0, 446.0, 1485.0, 16111.0, 3482967.0, 16777216.0, 210223.0, 5301.0, 832.0, 315.0, 233.0, 153.0, 143.0, 101.0, 81.0, 59.0, 57.0, 25.0, 21.0, 9.0, 7.0, 4.0, 5.0, 2.0, 1.0, 5.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.004150390625, -0.003986358642578125, -0.00382232666015625, -0.003658294677734375, -0.0034942626953125, -0.003330230712890625, -0.00316619873046875, -0.003002166748046875, -0.002838134765625, -0.002674102783203125, -0.00251007080078125, -0.002346038818359375, -0.0021820068359375, -0.002017974853515625, -0.00185394287109375, -0.001689910888671875, -0.00152587890625, -0.001361846923828125, -0.00119781494140625, -0.001033782958984375, -0.0008697509765625, -0.000705718994140625, -0.00054168701171875, -0.000377655029296875, -0.000213623046875, -4.9591064453125e-05, 0.00011444091796875, 0.000278472900390625, 0.0004425048828125, 0.000606536865234375, 0.00077056884765625, 0.000934600830078125, 0.0010986328125, 0.001262664794921875, 0.00142669677734375, 0.001590728759765625, 0.0017547607421875, 0.001918792724609375, 0.00208282470703125, 0.002246856689453125, 0.002410888671875, 0.002574920654296875, 0.00273895263671875, 0.002902984619140625, 0.0030670166015625, 0.003231048583984375, 0.00339508056640625, 0.003559112548828125, 0.00372314453125, 0.003887176513671875, 0.00405120849609375, 0.004215240478515625, 0.0043792724609375, 0.004543304443359375, 0.00470733642578125, 0.004871368408203125, 0.005035400390625, 0.005199432373046875, 0.00536346435546875, 0.005527496337890625, 0.0056915283203125, 0.005855560302734375, 0.00601959228515625, 0.006183624267578125, 0.00634765625]}, "gradients/model.layers.5.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 3.0, 0.0, 3.0, 2.0, 0.0, 4.0, 3.0, 3.0, 1.0, 8.0, 13.0, 21.0, 37.0, 62.0, 101.0, 264.0, 521.0, 1175.0, 3223.0, 9945.0, 42916.0, 504454.0, 16777216.0, 16777216.0, 1749686.0, 81981.0, 15164.0, 4739.0, 1667.0, 764.0, 338.0, 156.0, 56.0, 41.0, 22.0, 9.0, 16.0, 7.0, 2.0, 5.0, 5.0, 4.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.00201416015625, -0.0019426345825195312, -0.0018711090087890625, -0.0017995834350585938, -0.001728057861328125, -0.0016565322875976562, -0.0015850067138671875, -0.0015134811401367188, -0.00144195556640625, -0.0013704299926757812, -0.0012989044189453125, -0.0012273788452148438, -0.001155853271484375, -0.0010843276977539062, -0.0010128021240234375, -0.0009412765502929688, -0.0008697509765625, -0.0007982254028320312, -0.0007266998291015625, -0.0006551742553710938, -0.000583648681640625, -0.0005121231079101562, -0.0004405975341796875, -0.00036907196044921875, -0.00029754638671875, -0.00022602081298828125, -0.0001544952392578125, -8.296966552734375e-05, -1.1444091796875e-05, 6.008148193359375e-05, 0.0001316070556640625, 0.00020313262939453125, 0.000274658203125, 0.00034618377685546875, 0.0004177093505859375, 0.0004892349243164062, 0.000560760498046875, 0.0006322860717773438, 0.0007038116455078125, 0.0007753372192382812, 0.00084686279296875, 0.0009183883666992188, 0.0009899139404296875, 0.0010614395141601562, 0.001132965087890625, 0.0012044906616210938, 0.0012760162353515625, 0.0013475418090820312, 0.0014190673828125, 0.0014905929565429688, 0.0015621185302734375, 0.0016336441040039062, 0.001705169677734375, 0.0017766952514648438, 0.0018482208251953125, 0.0019197463989257812, 0.00199127197265625, 0.0020627975463867188, 0.0021343231201171875, 0.0022058486938476562, 0.002277374267578125, 0.0023488998413085938, 0.0024204254150390625, 0.0024919509887695312, 0.0025634765625]}, "gradients/model.layers.5.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 0.0, 1.0, 0.0, 1.0, 5.0, 4.0, 9.0, 13.0, 22.0, 53.0, 90.0, 208.0, 344.0, 522.0, 612.0, 656.0, 538.0, 443.0, 278.0, 138.0, 66.0, 30.0, 24.0, 11.0, 6.0, 3.0, 3.0, 1.0, 2.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0022125244140625, -0.0021497011184692383, -0.0020868778228759766, -0.002024054527282715, -0.001961231231689453, -0.0018984079360961914, -0.0018355846405029297, -0.001772761344909668, -0.0017099380493164062, -0.0016471147537231445, -0.0015842914581298828, -0.001521468162536621, -0.0014586448669433594, -0.0013958215713500977, -0.001332998275756836, -0.0012701749801635742, -0.0012073516845703125, -0.0011445283889770508, -0.001081705093383789, -0.0010188817977905273, -0.0009560585021972656, -0.0008932352066040039, -0.0008304119110107422, -0.0007675886154174805, -0.0007047653198242188, -0.000641942024230957, -0.0005791187286376953, -0.0005162954330444336, -0.0004534721374511719, -0.00039064884185791016, -0.00032782554626464844, -0.0002650022506713867, -0.000202178955078125, -0.00013935565948486328, -7.653236389160156e-05, -1.3709068298339844e-05, 4.9114227294921875e-05, 0.0001119375228881836, 0.0001747608184814453, 0.00023758411407470703, 0.00030040740966796875, 0.00036323070526123047, 0.0004260540008544922, 0.0004888772964477539, 0.0005517005920410156, 0.0006145238876342773, 0.0006773471832275391, 0.0007401704788208008, 0.0008029937744140625, 0.0008658170700073242, 0.0009286403656005859, 0.0009914636611938477, 0.0010542869567871094, 0.001117110252380371, 0.0011799335479736328, 0.0012427568435668945, 0.0013055801391601562, 0.001368403434753418, 0.0014312267303466797, 0.0014940500259399414, 0.0015568733215332031, 0.0016196966171264648, 0.0016825199127197266, 0.0017453432083129883, 0.00180816650390625]}, "gradients/model.layers.5.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 0.0, 0.0, 0.0, 1.0, 3.0, 5.0, 4.0, 3.0, 8.0, 12.0, 21.0, 32.0, 57.0, 98.0, 166.0, 313.0, 542.0, 1012.0, 1774.0, 3595.0, 7720.0, 20864.0, 80495.0, 658834.0, 7794923.0, 7466624.0, 625705.0, 78529.0, 20366.0, 8017.0, 3534.0, 1759.0, 932.0, 548.0, 279.0, 177.0, 103.0, 65.0, 27.0, 23.0, 10.0, 14.0, 4.0, 4.0, 2.0, 0.0, 0.0, 1.0, 1.0, 3.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00250244140625, -0.002413034439086914, -0.002323627471923828, -0.002234220504760742, -0.0021448135375976562, -0.0020554065704345703, -0.0019659996032714844, -0.0018765926361083984, -0.0017871856689453125, -0.0016977787017822266, -0.0016083717346191406, -0.0015189647674560547, -0.0014295578002929688, -0.0013401508331298828, -0.0012507438659667969, -0.001161336898803711, -0.001071929931640625, -0.000982522964477539, -0.0008931159973144531, -0.0008037090301513672, -0.0007143020629882812, -0.0006248950958251953, -0.0005354881286621094, -0.00044608116149902344, -0.0003566741943359375, -0.00026726722717285156, -0.00017786026000976562, -8.845329284667969e-05, 9.5367431640625e-07, 9.036064147949219e-05, 0.00017976760864257812, 0.00026917457580566406, 0.00035858154296875, 0.00044798851013183594, 0.0005373954772949219, 0.0006268024444580078, 0.0007162094116210938, 0.0008056163787841797, 0.0008950233459472656, 0.0009844303131103516, 0.0010738372802734375, 0.0011632442474365234, 0.0012526512145996094, 0.0013420581817626953, 0.0014314651489257812, 0.0015208721160888672, 0.0016102790832519531, 0.001699686050415039, 0.001789093017578125, 0.001878499984741211, 0.001967906951904297, 0.002057313919067383, 0.0021467208862304688, 0.0022361278533935547, 0.0023255348205566406, 0.0024149417877197266, 0.0025043487548828125, 0.0025937557220458984, 0.0026831626892089844, 0.0027725696563720703, 0.0028619766235351562, 0.002951383590698242, 0.003040790557861328, 0.003130197525024414, 0.0032196044921875]}, "gradients/model.layers.5.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 2.0, 2.0, 3.0, 1.0, 3.0, 1.0, 4.0, 6.0, 3.0, 6.0, 5.0, 9.0, 12.0, 14.0, 20.0, 32.0, 17.0, 34.0, 30.0, 41.0, 54.0, 73.0, 77.0, 91.0, 104.0, 165.0, 288.0, 731.0, 2261.0, 10792.0, 755822.0, 3399148.0, 19069.0, 3186.0, 927.0, 387.0, 204.0, 130.0, 84.0, 80.0, 51.0, 39.0, 58.0, 54.0, 46.0, 18.0, 19.0, 21.0, 9.0, 13.0, 10.0, 5.0, 10.0, 11.0, 5.0, 6.0, 4.0, 2.0, 1.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.01495361328125, -0.01448822021484375, -0.0140228271484375, -0.01355743408203125, -0.013092041015625, -0.01262664794921875, -0.0121612548828125, -0.01169586181640625, -0.01123046875, -0.01076507568359375, -0.0102996826171875, -0.00983428955078125, -0.009368896484375, -0.00890350341796875, -0.0084381103515625, -0.00797271728515625, -0.00750732421875, -0.00704193115234375, -0.0065765380859375, -0.00611114501953125, -0.005645751953125, -0.00518035888671875, -0.0047149658203125, -0.00424957275390625, -0.0037841796875, -0.00331878662109375, -0.0028533935546875, -0.00238800048828125, -0.001922607421875, -0.00145721435546875, -0.0009918212890625, -0.00052642822265625, -6.103515625e-05, 0.00040435791015625, 0.0008697509765625, 0.00133514404296875, 0.001800537109375, 0.00226593017578125, 0.0027313232421875, 0.00319671630859375, 0.003662109375, 0.00412750244140625, 0.0045928955078125, 0.00505828857421875, 0.005523681640625, 0.00598907470703125, 0.0064544677734375, 0.00691986083984375, 0.00738525390625, 0.00785064697265625, 0.0083160400390625, 0.00878143310546875, 0.009246826171875, 0.00971221923828125, 0.0101776123046875, 0.01064300537109375, 0.0111083984375, 0.01157379150390625, 0.0120391845703125, 0.01250457763671875, 0.012969970703125, 0.01343536376953125, 0.0139007568359375, 0.01436614990234375, 0.01483154296875]}, "gradients/model.layers.5.self_attn.k_proj.weight": {"_type": "histogram", "values": [3.0, 0.0, 0.0, 1.0, 1.0, 1.0, 2.0, 1.0, 5.0, 4.0, 4.0, 8.0, 16.0, 20.0, 19.0, 21.0, 35.0, 58.0, 64.0, 95.0, 117.0, 212.0, 315.0, 570.0, 1376.0, 3185.0, 10079.0, 51549.0, 2946449.0, 1133464.0, 33497.0, 7826.0, 2796.0, 1099.0, 486.0, 312.0, 194.0, 101.0, 89.0, 54.0, 38.0, 31.0, 21.0, 14.0, 13.0, 12.0, 13.0, 7.0, 5.0, 1.0, 2.0, 6.0, 4.0, 2.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0], "bins": [-0.0014190673828125, -0.0013699531555175781, -0.0013208389282226562, -0.0012717247009277344, -0.0012226104736328125, -0.0011734962463378906, -0.0011243820190429688, -0.0010752677917480469, -0.001026153564453125, -0.0009770393371582031, -0.0009279251098632812, -0.0008788108825683594, -0.0008296966552734375, -0.0007805824279785156, -0.0007314682006835938, -0.0006823539733886719, -0.00063323974609375, -0.0005841255187988281, -0.0005350112915039062, -0.0004858970642089844, -0.0004367828369140625, -0.0003876686096191406, -0.00033855438232421875, -0.0002894401550292969, -0.000240325927734375, -0.00019121170043945312, -0.00014209747314453125, -9.298324584960938e-05, -4.38690185546875e-05, 5.245208740234375e-06, 5.435943603515625e-05, 0.00010347366333007812, 0.000152587890625, 0.00020170211791992188, 0.00025081634521484375, 0.0002999305725097656, 0.0003490447998046875, 0.0003981590270996094, 0.00044727325439453125, 0.0004963874816894531, 0.000545501708984375, 0.0005946159362792969, 0.0006437301635742188, 0.0006928443908691406, 0.0007419586181640625, 0.0007910728454589844, 0.0008401870727539062, 0.0008893013000488281, 0.00093841552734375, 0.0009875297546386719, 0.0010366439819335938, 0.0010857582092285156, 0.0011348724365234375, 0.0011839866638183594, 0.0012331008911132812, 0.0012822151184082031, 0.001331329345703125, 0.0013804435729980469, 0.0014295578002929688, 0.0014786720275878906, 0.0015277862548828125, 0.0015769004821777344, 0.0016260147094726562, 0.0016751289367675781, 0.0017242431640625]}, "gradients/model.layers.5.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 4.0, 0.0, 6.0, 6.0, 11.0, 5.0, 9.0, 13.0, 12.0, 19.0, 31.0, 37.0, 57.0, 55.0, 96.0, 104.0, 153.0, 234.0, 348.0, 622.0, 1007.0, 2038.0, 4488.0, 13041.0, 56358.0, 852468.0, 15578308.0, 220881.0, 31157.0, 8503.0, 3268.0, 1486.0, 807.0, 473.0, 313.0, 224.0, 142.0, 98.0, 71.0, 67.0, 54.0, 33.0, 26.0, 13.0, 7.0, 14.0, 14.0, 4.0, 10.0, 3.0, 4.0, 4.0, 0.0, 1.0, 0.0, 2.0, 0.0, 2.0], "bins": [-0.0019378662109375, -0.001877903938293457, -0.001817941665649414, -0.001757979393005371, -0.0016980171203613281, -0.0016380548477172852, -0.0015780925750732422, -0.0015181303024291992, -0.0014581680297851562, -0.0013982057571411133, -0.0013382434844970703, -0.0012782812118530273, -0.0012183189392089844, -0.0011583566665649414, -0.0010983943939208984, -0.0010384321212768555, -0.0009784698486328125, -0.0009185075759887695, -0.0008585453033447266, -0.0007985830307006836, -0.0007386207580566406, -0.0006786584854125977, -0.0006186962127685547, -0.0005587339401245117, -0.0004987716674804688, -0.0004388093948364258, -0.0003788471221923828, -0.00031888484954833984, -0.0002589225769042969, -0.0001989603042602539, -0.00013899803161621094, -7.903575897216797e-05, -1.9073486328125e-05, 4.088878631591797e-05, 0.00010085105895996094, 0.0001608133316040039, 0.00022077560424804688, 0.00028073787689208984, 0.0003407001495361328, 0.0004006624221801758, 0.00046062469482421875, 0.0005205869674682617, 0.0005805492401123047, 0.0006405115127563477, 0.0007004737854003906, 0.0007604360580444336, 0.0008203983306884766, 0.0008803606033325195, 0.0009403228759765625, 0.0010002851486206055, 0.0010602474212646484, 0.0011202096939086914, 0.0011801719665527344, 0.0012401342391967773, 0.0013000965118408203, 0.0013600587844848633, 0.0014200210571289062, 0.0014799833297729492, 0.0015399456024169922, 0.0015999078750610352, 0.0016598701477050781, 0.001719832420349121, 0.001779794692993164, 0.001839756965637207, 0.00189971923828125]}, "gradients/model.layers.5.input_layernorm.weight": {"_type": "histogram", "values": [2.0, 1.0, 3.0, 36.0, 1613.0, 2351.0, 57.0, 9.0, 11.0, 4.0, 2.0, 0.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.000812530517578125, -0.0006529688835144043, -0.0004934072494506836, -0.0003338456153869629, -0.0001742839813232422, -1.4722347259521484e-05, 0.00014483928680419922, 0.0003044009208679199, 0.0004639625549316406, 0.0006235241889953613, 0.000783085823059082, 0.0009426474571228027, 0.0011022090911865234, 0.0012617707252502441, 0.0014213323593139648, 0.0015808939933776855, 0.0017404556274414062, 0.001900017261505127, 0.0020595788955688477, 0.0022191405296325684, 0.002378702163696289, 0.0025382637977600098, 0.0026978254318237305, 0.002857387065887451, 0.003016948699951172, 0.0031765103340148926, 0.0033360719680786133, 0.003495633602142334, 0.0036551952362060547, 0.0038147568702697754, 0.003974318504333496, 0.004133880138397217, 0.0042934417724609375, 0.004453003406524658, 0.004612565040588379, 0.0047721266746521, 0.00493168830871582, 0.005091249942779541, 0.005250811576843262, 0.005410373210906982, 0.005569934844970703, 0.005729496479034424, 0.0058890581130981445, 0.006048619747161865, 0.006208181381225586, 0.006367743015289307, 0.006527304649353027, 0.006686866283416748, 0.006846427917480469, 0.0070059895515441895, 0.00716555118560791, 0.007325112819671631, 0.0074846744537353516, 0.007644236087799072, 0.007803797721862793, 0.007963359355926514, 0.008122920989990234, 0.008282482624053955, 0.008442044258117676, 0.008601605892181396, 0.008761167526245117, 0.008920729160308838, 0.009080290794372559, 0.00923985242843628, 0.0093994140625]}, "gradients/model.layers.4.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 2.0, 2.0, 2.0, 0.0, 2.0, 3.0, 6.0, 7.0, 15.0, 15.0, 22.0, 28.0, 46.0, 86.0, 80.0, 109.0, 153.0, 239.0, 348.0, 671.0, 1440.0, 4260.0, 17146.0, 112832.0, 2306336.0, 16777216.0, 15037960.0, 444287.0, 42802.0, 8428.0, 2385.0, 935.0, 483.0, 328.0, 193.0, 160.0, 96.0, 67.0, 47.0, 35.0, 33.0, 28.0, 18.0, 14.0, 2.0, 10.0, 3.0, 0.0, 3.0, 1.0, 2.0, 2.0, 3.0], "bins": [-0.002593994140625, -0.0025234222412109375, -0.002452850341796875, -0.0023822784423828125, -0.00231170654296875, -0.0022411346435546875, -0.002170562744140625, -0.0020999908447265625, -0.0020294189453125, -0.0019588470458984375, -0.001888275146484375, -0.0018177032470703125, -0.00174713134765625, -0.0016765594482421875, -0.001605987548828125, -0.0015354156494140625, -0.00146484375, -0.0013942718505859375, -0.001323699951171875, -0.0012531280517578125, -0.00118255615234375, -0.0011119842529296875, -0.001041412353515625, -0.0009708404541015625, -0.0009002685546875, -0.0008296966552734375, -0.000759124755859375, -0.0006885528564453125, -0.00061798095703125, -0.0005474090576171875, -0.000476837158203125, -0.0004062652587890625, -0.000335693359375, -0.0002651214599609375, -0.000194549560546875, -0.0001239776611328125, -5.340576171875e-05, 1.71661376953125e-05, 8.7738037109375e-05, 0.0001583099365234375, 0.0002288818359375, 0.0002994537353515625, 0.000370025634765625, 0.0004405975341796875, 0.00051116943359375, 0.0005817413330078125, 0.000652313232421875, 0.0007228851318359375, 0.00079345703125, 0.0008640289306640625, 0.000934600830078125, 0.0010051727294921875, 0.00107574462890625, 0.0011463165283203125, 0.001216888427734375, 0.0012874603271484375, 0.0013580322265625, 0.0014286041259765625, 0.001499176025390625, 0.0015697479248046875, 0.00164031982421875, 0.0017108917236328125, 0.001781463623046875, 0.0018520355224609375, 0.001922607421875]}, "gradients/model.layers.4.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 3.0, 2.0, 3.0, 8.0, 3.0, 6.0, 8.0, 8.0, 22.0, 39.0, 48.0, 73.0, 107.0, 212.0, 376.0, 711.0, 1377.0, 2813.0, 6629.0, 17589.0, 62922.0, 374666.0, 5232945.0, 16777216.0, 16777216.0, 2334706.0, 210334.0, 42144.0, 13647.0, 5189.0, 2247.0, 1144.0, 621.0, 311.0, 187.0, 88.0, 54.0, 42.0, 21.0, 23.0, 7.0, 12.0, 3.0, 3.0, 4.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00159454345703125, -0.0015500783920288086, -0.0015056133270263672, -0.0014611482620239258, -0.0014166831970214844, -0.001372218132019043, -0.0013277530670166016, -0.0012832880020141602, -0.0012388229370117188, -0.0011943578720092773, -0.001149892807006836, -0.0011054277420043945, -0.0010609626770019531, -0.0010164976119995117, -0.0009720325469970703, -0.0009275674819946289, -0.0008831024169921875, -0.0008386373519897461, -0.0007941722869873047, -0.0007497072219848633, -0.0007052421569824219, -0.0006607770919799805, -0.0006163120269775391, -0.0005718469619750977, -0.0005273818969726562, -0.00048291683197021484, -0.00043845176696777344, -0.00039398670196533203, -0.0003495216369628906, -0.0003050565719604492, -0.0002605915069580078, -0.0002161264419555664, -0.000171661376953125, -0.0001271963119506836, -8.273124694824219e-05, -3.826618194580078e-05, 6.198883056640625e-06, 5.066394805908203e-05, 9.512901306152344e-05, 0.00013959407806396484, 0.00018405914306640625, 0.00022852420806884766, 0.00027298927307128906, 0.00031745433807373047, 0.0003619194030761719, 0.0004063844680786133, 0.0004508495330810547, 0.0004953145980834961, 0.0005397796630859375, 0.0005842447280883789, 0.0006287097930908203, 0.0006731748580932617, 0.0007176399230957031, 0.0007621049880981445, 0.0008065700531005859, 0.0008510351181030273, 0.0008955001831054688, 0.0009399652481079102, 0.0009844303131103516, 0.001028895378112793, 0.0010733604431152344, 0.0011178255081176758, 0.0011622905731201172, 0.0012067556381225586, 0.001251220703125]}, "gradients/model.layers.4.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 0.0, 3.0, 0.0, 2.0, 1.0, 1.0, 2.0, 1.0, 2.0, 3.0, 3.0, 3.0, 6.0, 1.0, 11.0, 11.0, 16.0, 37.0, 65.0, 125.0, 225.0, 307.0, 403.0, 541.0, 532.0, 465.0, 457.0, 298.0, 235.0, 128.0, 82.0, 39.0, 33.0, 17.0, 8.0, 5.0, 2.0, 3.0, 1.0, 2.0, 3.0, 0.0, 2.0, 0.0, 1.0, 1.0, 3.0, 1.0, 1.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00148773193359375, -0.0014356374740600586, -0.0013835430145263672, -0.0013314485549926758, -0.0012793540954589844, -0.001227259635925293, -0.0011751651763916016, -0.0011230707168579102, -0.0010709762573242188, -0.0010188817977905273, -0.0009667873382568359, -0.0009146928787231445, -0.0008625984191894531, -0.0008105039596557617, -0.0007584095001220703, -0.0007063150405883789, -0.0006542205810546875, -0.0006021261215209961, -0.0005500316619873047, -0.0004979372024536133, -0.0004458427429199219, -0.00039374828338623047, -0.00034165382385253906, -0.00028955936431884766, -0.00023746490478515625, -0.00018537044525146484, -0.00013327598571777344, -8.118152618408203e-05, -2.9087066650390625e-05, 2.300739288330078e-05, 7.510185241699219e-05, 0.0001271963119506836, 0.000179290771484375, 0.0002313852310180664, 0.0002834796905517578, 0.0003355741500854492, 0.0003876686096191406, 0.00043976306915283203, 0.0004918575286865234, 0.0005439519882202148, 0.0005960464477539062, 0.0006481409072875977, 0.0007002353668212891, 0.0007523298263549805, 0.0008044242858886719, 0.0008565187454223633, 0.0009086132049560547, 0.0009607076644897461, 0.0010128021240234375, 0.001064896583557129, 0.0011169910430908203, 0.0011690855026245117, 0.0012211799621582031, 0.0012732744216918945, 0.001325368881225586, 0.0013774633407592773, 0.0014295578002929688, 0.0014816522598266602, 0.0015337467193603516, 0.001585841178894043, 0.0016379356384277344, 0.0016900300979614258, 0.0017421245574951172, 0.0017942190170288086, 0.0018463134765625]}, "gradients/model.layers.4.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 2.0, 0.0, 4.0, 2.0, 0.0, 7.0, 10.0, 8.0, 20.0, 40.0, 64.0, 105.0, 199.0, 372.0, 649.0, 1202.0, 2488.0, 5594.0, 16065.0, 95348.0, 1625840.0, 12564287.0, 2302823.0, 130797.0, 19347.0, 6223.0, 2786.0, 1233.0, 766.0, 410.0, 204.0, 121.0, 77.0, 45.0, 32.0, 12.0, 11.0, 7.0, 6.0, 1.0, 1.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.0034027099609375, -0.0033121109008789062, -0.0032215118408203125, -0.0031309127807617188, -0.003040313720703125, -0.0029497146606445312, -0.0028591156005859375, -0.0027685165405273438, -0.00267791748046875, -0.0025873184204101562, -0.0024967193603515625, -0.0024061203002929688, -0.002315521240234375, -0.0022249221801757812, -0.0021343231201171875, -0.0020437240600585938, -0.001953125, -0.0018625259399414062, -0.0017719268798828125, -0.0016813278198242188, -0.001590728759765625, -0.0015001296997070312, -0.0014095306396484375, -0.0013189315795898438, -0.00122833251953125, -0.0011377334594726562, -0.0010471343994140625, -0.0009565353393554688, -0.000865936279296875, -0.0007753372192382812, -0.0006847381591796875, -0.0005941390991210938, -0.0005035400390625, -0.00041294097900390625, -0.0003223419189453125, -0.00023174285888671875, -0.000141143798828125, -5.054473876953125e-05, 4.00543212890625e-05, 0.00013065338134765625, 0.00022125244140625, 0.00031185150146484375, 0.0004024505615234375, 0.0004930496215820312, 0.000583648681640625, 0.0006742477416992188, 0.0007648468017578125, 0.0008554458618164062, 0.000946044921875, 0.0010366439819335938, 0.0011272430419921875, 0.0012178421020507812, 0.001308441162109375, 0.0013990402221679688, 0.0014896392822265625, 0.0015802383422851562, 0.00167083740234375, 0.0017614364624023438, 0.0018520355224609375, 0.0019426345825195312, 0.002033233642578125, 0.0021238327026367188, 0.0022144317626953125, 0.0023050308227539062, 0.0023956298828125]}, "gradients/model.layers.4.self_attn.v_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 1.0, 2.0, 1.0, 3.0, 2.0, 2.0, 2.0, 2.0, 5.0, 8.0, 9.0, 10.0, 6.0, 8.0, 20.0, 27.0, 43.0, 46.0, 52.0, 50.0, 69.0, 68.0, 77.0, 93.0, 157.0, 656.0, 5767.0, 2090826.0, 2089046.0, 5785.0, 672.0, 169.0, 98.0, 91.0, 67.0, 59.0, 50.0, 55.0, 43.0, 41.0, 28.0, 20.0, 17.0, 9.0, 8.0, 3.0, 4.0, 3.0, 6.0, 4.0, 5.0, 2.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.022705078125, -0.02197265625, -0.021240234375, -0.0205078125, -0.019775390625, -0.01904296875, -0.018310546875, -0.017578125, -0.016845703125, -0.01611328125, -0.015380859375, -0.0146484375, -0.013916015625, -0.01318359375, -0.012451171875, -0.01171875, -0.010986328125, -0.01025390625, -0.009521484375, -0.0087890625, -0.008056640625, -0.00732421875, -0.006591796875, -0.005859375, -0.005126953125, -0.00439453125, -0.003662109375, -0.0029296875, -0.002197265625, -0.00146484375, -0.000732421875, 0.0, 0.000732421875, 0.00146484375, 0.002197265625, 0.0029296875, 0.003662109375, 0.00439453125, 0.005126953125, 0.005859375, 0.006591796875, 0.00732421875, 0.008056640625, 0.0087890625, 0.009521484375, 0.01025390625, 0.010986328125, 0.01171875, 0.012451171875, 0.01318359375, 0.013916015625, 0.0146484375, 0.015380859375, 0.01611328125, 0.016845703125, 0.017578125, 0.018310546875, 0.01904296875, 0.019775390625, 0.0205078125, 0.021240234375, 0.02197265625, 0.022705078125, 0.0234375, 0.024169921875]}, "gradients/model.layers.4.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 3.0, 1.0, 1.0, 3.0, 0.0, 4.0, 0.0, 4.0, 6.0, 7.0, 6.0, 7.0, 11.0, 14.0, 26.0, 44.0, 37.0, 48.0, 60.0, 128.0, 182.0, 342.0, 701.0, 2019.0, 7420.0, 63243.0, 4013297.0, 93734.0, 8856.0, 2266.0, 815.0, 384.0, 228.0, 119.0, 70.0, 60.0, 35.0, 30.0, 18.0, 9.0, 14.0, 9.0, 10.0, 9.0, 7.0, 2.0, 1.0, 4.0, 0.0, 1.0, 0.0, 0.0, 1.0, 1.0, 3.0], "bins": [-0.0025177001953125, -0.002446889877319336, -0.002376079559326172, -0.002305269241333008, -0.0022344589233398438, -0.0021636486053466797, -0.0020928382873535156, -0.0020220279693603516, -0.0019512176513671875, -0.0018804073333740234, -0.0018095970153808594, -0.0017387866973876953, -0.0016679763793945312, -0.0015971660614013672, -0.0015263557434082031, -0.001455545425415039, -0.001384735107421875, -0.001313924789428711, -0.0012431144714355469, -0.0011723041534423828, -0.0011014938354492188, -0.0010306835174560547, -0.0009598731994628906, -0.0008890628814697266, -0.0008182525634765625, -0.0007474422454833984, -0.0006766319274902344, -0.0006058216094970703, -0.0005350112915039062, -0.0004642009735107422, -0.0003933906555175781, -0.00032258033752441406, -0.00025177001953125, -0.00018095970153808594, -0.00011014938354492188, -3.933906555175781e-05, 3.147125244140625e-05, 0.00010228157043457031, 0.00017309188842773438, 0.00024390220642089844, 0.0003147125244140625, 0.00038552284240722656, 0.0004563331604003906, 0.0005271434783935547, 0.0005979537963867188, 0.0006687641143798828, 0.0007395744323730469, 0.0008103847503662109, 0.000881195068359375, 0.0009520053863525391, 0.0010228157043457031, 0.0010936260223388672, 0.0011644363403320312, 0.0012352466583251953, 0.0013060569763183594, 0.0013768672943115234, 0.0014476776123046875, 0.0015184879302978516, 0.0015892982482910156, 0.0016601085662841797, 0.0017309188842773438, 0.0018017292022705078, 0.0018725395202636719, 0.001943349838256836, 0.00201416015625]}, "gradients/model.layers.4.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 1.0, 3.0, 1.0, 2.0, 1.0, 3.0, 7.0, 8.0, 10.0, 15.0, 15.0, 23.0, 35.0, 40.0, 58.0, 64.0, 121.0, 203.0, 306.0, 505.0, 988.0, 2151.0, 6324.0, 28786.0, 472738.0, 16132226.0, 110252.0, 15015.0, 4037.0, 1456.0, 728.0, 384.0, 231.0, 136.0, 95.0, 55.0, 52.0, 35.0, 26.0, 23.0, 10.0, 6.0, 10.0, 5.0, 4.0, 6.0, 4.0, 0.0, 5.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0023956298828125, -0.0023190975189208984, -0.002242565155029297, -0.0021660327911376953, -0.0020895004272460938, -0.002012968063354492, -0.0019364356994628906, -0.001859903335571289, -0.0017833709716796875, -0.001706838607788086, -0.0016303062438964844, -0.0015537738800048828, -0.0014772415161132812, -0.0014007091522216797, -0.0013241767883300781, -0.0012476444244384766, -0.001171112060546875, -0.0010945796966552734, -0.0010180473327636719, -0.0009415149688720703, -0.0008649826049804688, -0.0007884502410888672, -0.0007119178771972656, -0.0006353855133056641, -0.0005588531494140625, -0.00048232078552246094, -0.0004057884216308594, -0.0003292560577392578, -0.00025272369384765625, -0.0001761913299560547, -9.965896606445312e-05, -2.3126602172851562e-05, 5.340576171875e-05, 0.00012993812561035156, 0.00020647048950195312, 0.0002830028533935547, 0.00035953521728515625, 0.0004360675811767578, 0.0005125999450683594, 0.0005891323089599609, 0.0006656646728515625, 0.0007421970367431641, 0.0008187294006347656, 0.0008952617645263672, 0.0009717941284179688, 0.0010483264923095703, 0.0011248588562011719, 0.0012013912200927734, 0.001277923583984375, 0.0013544559478759766, 0.0014309883117675781, 0.0015075206756591797, 0.0015840530395507812, 0.0016605854034423828, 0.0017371177673339844, 0.001813650131225586, 0.0018901824951171875, 0.001966714859008789, 0.0020432472229003906, 0.002119779586791992, 0.0021963119506835938, 0.0022728443145751953, 0.002349376678466797, 0.0024259090423583984, 0.00250244140625]}, "gradients/model.layers.4.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 1.0, 0.0, 4.0, 2.0, 4.0, 2.0, 7.0, 15.0, 11.0, 19.0, 56.0, 252.0, 783.0, 1315.0, 1056.0, 405.0, 106.0, 28.0, 11.0, 3.0, 1.0, 3.0, 2.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00131988525390625, -0.0012685060501098633, -0.0012171268463134766, -0.0011657476425170898, -0.0011143684387207031, -0.0010629892349243164, -0.0010116100311279297, -0.000960230827331543, -0.0009088516235351562, -0.0008574724197387695, -0.0008060932159423828, -0.0007547140121459961, -0.0007033348083496094, -0.0006519556045532227, -0.0006005764007568359, -0.0005491971969604492, -0.0004978179931640625, -0.0004464387893676758, -0.00039505958557128906, -0.00034368038177490234, -0.0002923011779785156, -0.0002409219741821289, -0.0001895427703857422, -0.00013816356658935547, -8.678436279296875e-05, -3.540515899658203e-05, 1.5974044799804688e-05, 6.73532485961914e-05, 0.00011873245239257812, 0.00017011165618896484, 0.00022149085998535156, 0.0002728700637817383, 0.000324249267578125, 0.0003756284713745117, 0.00042700767517089844, 0.00047838687896728516, 0.0005297660827636719, 0.0005811452865600586, 0.0006325244903564453, 0.000683903694152832, 0.0007352828979492188, 0.0007866621017456055, 0.0008380413055419922, 0.0008894205093383789, 0.0009407997131347656, 0.0009921789169311523, 0.001043558120727539, 0.0010949373245239258, 0.0011463165283203125, 0.0011976957321166992, 0.001249074935913086, 0.0013004541397094727, 0.0013518333435058594, 0.001403212547302246, 0.0014545917510986328, 0.0015059709548950195, 0.0015573501586914062, 0.001608729362487793, 0.0016601085662841797, 0.0017114877700805664, 0.0017628669738769531, 0.0018142461776733398, 0.0018656253814697266, 0.0019170045852661133, 0.0019683837890625]}, "gradients/model.layers.3.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 4.0, 2.0, 1.0, 2.0, 5.0, 3.0, 13.0, 13.0, 12.0, 20.0, 27.0, 43.0, 72.0, 71.0, 88.0, 124.0, 172.0, 206.0, 359.0, 842.0, 3290.0, 25169.0, 689019.0, 16777216.0, 8210535.0, 89634.0, 7656.0, 1386.0, 472.0, 234.0, 178.0, 141.0, 113.0, 65.0, 75.0, 51.0, 39.0, 29.0, 19.0, 13.0, 5.0, 7.0, 5.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00274658203125, -0.0026509761810302734, -0.002555370330810547, -0.0024597644805908203, -0.0023641586303710938, -0.002268552780151367, -0.0021729469299316406, -0.002077341079711914, -0.0019817352294921875, -0.001886129379272461, -0.0017905235290527344, -0.0016949176788330078, -0.0015993118286132812, -0.0015037059783935547, -0.0014081001281738281, -0.0013124942779541016, -0.001216888427734375, -0.0011212825775146484, -0.0010256767272949219, -0.0009300708770751953, -0.0008344650268554688, -0.0007388591766357422, -0.0006432533264160156, -0.0005476474761962891, -0.0004520416259765625, -0.00035643577575683594, -0.0002608299255371094, -0.0001652240753173828, -6.961822509765625e-05, 2.5987625122070312e-05, 0.00012159347534179688, 0.00021719932556152344, 0.00031280517578125, 0.00040841102600097656, 0.0005040168762207031, 0.0005996227264404297, 0.0006952285766601562, 0.0007908344268798828, 0.0008864402770996094, 0.000982046127319336, 0.0010776519775390625, 0.001173257827758789, 0.0012688636779785156, 0.0013644695281982422, 0.0014600753784179688, 0.0015556812286376953, 0.0016512870788574219, 0.0017468929290771484, 0.001842498779296875, 0.0019381046295166016, 0.002033710479736328, 0.0021293163299560547, 0.0022249221801757812, 0.002320528030395508, 0.0024161338806152344, 0.002511739730834961, 0.0026073455810546875, 0.002702951431274414, 0.0027985572814941406, 0.002894163131713867, 0.0029897689819335938, 0.0030853748321533203, 0.003180980682373047, 0.0032765865325927734, 0.0033721923828125]}, "gradients/model.layers.3.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 2.0, 4.0, 1.0, 1.0, 1.0, 4.0, 9.0, 3.0, 11.0, 16.0, 16.0, 27.0, 37.0, 62.0, 84.0, 159.0, 274.0, 536.0, 964.0, 1982.0, 5516.0, 21458.0, 161174.0, 5031827.0, 16777216.0, 16777216.0, 393661.0, 37313.0, 8068.0, 2697.0, 1236.0, 639.0, 322.0, 190.0, 113.0, 65.0, 41.0, 28.0, 16.0, 20.0, 5.0, 7.0, 8.0, 1.0, 3.0, 1.0, 0.0, 3.0, 2.0, 0.0, 1.0, 0.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0013885498046875, -0.0013344287872314453, -0.0012803077697753906, -0.001226186752319336, -0.0011720657348632812, -0.0011179447174072266, -0.0010638236999511719, -0.0010097026824951172, -0.0009555816650390625, -0.0009014606475830078, -0.0008473396301269531, -0.0007932186126708984, -0.0007390975952148438, -0.0006849765777587891, -0.0006308555603027344, -0.0005767345428466797, -0.000522613525390625, -0.0004684925079345703, -0.0004143714904785156, -0.00036025047302246094, -0.00030612945556640625, -0.00025200843811035156, -0.00019788742065429688, -0.0001437664031982422, -8.96453857421875e-05, -3.552436828613281e-05, 1.8596649169921875e-05, 7.271766662597656e-05, 0.00012683868408203125, 0.00018095970153808594, 0.00023508071899414062, 0.0002892017364501953, 0.00034332275390625, 0.0003974437713623047, 0.0004515647888183594, 0.0005056858062744141, 0.0005598068237304688, 0.0006139278411865234, 0.0006680488586425781, 0.0007221698760986328, 0.0007762908935546875, 0.0008304119110107422, 0.0008845329284667969, 0.0009386539459228516, 0.0009927749633789062, 0.001046895980834961, 0.0011010169982910156, 0.0011551380157470703, 0.001209259033203125, 0.0012633800506591797, 0.0013175010681152344, 0.001371622085571289, 0.0014257431030273438, 0.0014798641204833984, 0.0015339851379394531, 0.0015881061553955078, 0.0016422271728515625, 0.0016963481903076172, 0.0017504692077636719, 0.0018045902252197266, 0.0018587112426757812, 0.001912832260131836, 0.0019669532775878906, 0.0020210742950439453, 0.0020751953125]}, "gradients/model.layers.3.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0, 3.0, 1.0, 2.0, 1.0, 3.0, 3.0, 3.0, 0.0, 3.0, 9.0, 14.0, 14.0, 36.0, 98.0, 163.0, 268.0, 445.0, 513.0, 638.0, 558.0, 489.0, 348.0, 229.0, 126.0, 58.0, 26.0, 12.0, 9.0, 4.0, 5.0, 1.0, 1.0, 1.0, 2.0, 2.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.00274658203125, -0.00268399715423584, -0.0026214122772216797, -0.0025588274002075195, -0.0024962425231933594, -0.0024336576461791992, -0.002371072769165039, -0.002308487892150879, -0.0022459030151367188, -0.0021833181381225586, -0.0021207332611083984, -0.0020581483840942383, -0.001995563507080078, -0.001932978630065918, -0.0018703937530517578, -0.0018078088760375977, -0.0017452239990234375, -0.0016826391220092773, -0.0016200542449951172, -0.001557469367980957, -0.0014948844909667969, -0.0014322996139526367, -0.0013697147369384766, -0.0013071298599243164, -0.0012445449829101562, -0.001181960105895996, -0.001119375228881836, -0.0010567903518676758, -0.0009942054748535156, -0.0009316205978393555, -0.0008690357208251953, -0.0008064508438110352, -0.000743865966796875, -0.0006812810897827148, -0.0006186962127685547, -0.0005561113357543945, -0.0004935264587402344, -0.0004309415817260742, -0.00036835670471191406, -0.0003057718276977539, -0.00024318695068359375, -0.0001806020736694336, -0.00011801719665527344, -5.543231964111328e-05, 7.152557373046875e-06, 6.973743438720703e-05, 0.0001323223114013672, 0.00019490718841552734, 0.0002574920654296875, 0.00032007694244384766, 0.0003826618194580078, 0.00044524669647216797, 0.0005078315734863281, 0.0005704164505004883, 0.0006330013275146484, 0.0006955862045288086, 0.0007581710815429688, 0.0008207559585571289, 0.0008833408355712891, 0.0009459257125854492, 0.0010085105895996094, 0.0010710954666137695, 0.0011336803436279297, 0.0011962652206420898, 0.00125885009765625]}, "gradients/model.layers.3.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 4.0, 1.0, 1.0, 2.0, 5.0, 6.0, 17.0, 20.0, 44.0, 64.0, 133.0, 197.0, 405.0, 835.0, 1698.0, 3658.0, 9725.0, 30416.0, 175398.0, 1696347.0, 9596886.0, 4646752.0, 523317.0, 64563.0, 16084.0, 5707.0, 2516.0, 1184.0, 584.0, 288.0, 163.0, 75.0, 54.0, 25.0, 13.0, 11.0, 4.0, 9.0, 1.0, 1.0, 0.0, 1.0], "bins": [-0.0029144287109375, -0.002846240997314453, -0.0027780532836914062, -0.0027098655700683594, -0.0026416778564453125, -0.0025734901428222656, -0.0025053024291992188, -0.002437114715576172, -0.002368927001953125, -0.002300739288330078, -0.0022325515747070312, -0.0021643638610839844, -0.0020961761474609375, -0.0020279884338378906, -0.0019598007202148438, -0.0018916130065917969, -0.00182342529296875, -0.0017552375793457031, -0.0016870498657226562, -0.0016188621520996094, -0.0015506744384765625, -0.0014824867248535156, -0.0014142990112304688, -0.0013461112976074219, -0.001277923583984375, -0.0012097358703613281, -0.0011415481567382812, -0.0010733604431152344, -0.0010051727294921875, -0.0009369850158691406, -0.0008687973022460938, -0.0008006095886230469, -0.000732421875, -0.0006642341613769531, -0.0005960464477539062, -0.0005278587341308594, -0.0004596710205078125, -0.0003914833068847656, -0.00032329559326171875, -0.0002551078796386719, -0.000186920166015625, -0.00011873245239257812, -5.054473876953125e-05, 1.7642974853515625e-05, 8.58306884765625e-05, 0.00015401840209960938, 0.00022220611572265625, 0.0002903938293457031, 0.00035858154296875, 0.0004267692565917969, 0.0004949569702148438, 0.0005631446838378906, 0.0006313323974609375, 0.0006995201110839844, 0.0007677078247070312, 0.0008358955383300781, 0.000904083251953125, 0.0009722709655761719, 0.0010404586791992188, 0.0011086463928222656, 0.0011768341064453125, 0.0012450218200683594, 0.0013132095336914062, 0.0013813972473144531, 0.0014495849609375]}, "gradients/model.layers.3.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 2.0, 1.0, 1.0, 2.0, 4.0, 3.0, 8.0, 5.0, 9.0, 10.0, 9.0, 19.0, 21.0, 19.0, 19.0, 28.0, 36.0, 39.0, 51.0, 53.0, 60.0, 70.0, 90.0, 98.0, 215.0, 648.0, 3986.0, 308209.0, 3870850.0, 7785.0, 1000.0, 243.0, 123.0, 90.0, 58.0, 62.0, 67.0, 41.0, 37.0, 40.0, 38.0, 20.0, 22.0, 18.0, 21.0, 11.0, 16.0, 3.0, 9.0, 6.0, 5.0, 7.0, 4.0, 2.0, 3.0, 2.0, 1.0], "bins": [-0.026611328125, -0.025854110717773438, -0.025096893310546875, -0.024339675903320312, -0.02358245849609375, -0.022825241088867188, -0.022068023681640625, -0.021310806274414062, -0.0205535888671875, -0.019796371459960938, -0.019039154052734375, -0.018281936645507812, -0.01752471923828125, -0.016767501831054688, -0.016010284423828125, -0.015253067016601562, -0.014495849609375, -0.013738632202148438, -0.012981414794921875, -0.012224197387695312, -0.01146697998046875, -0.010709762573242188, -0.009952545166015625, -0.009195327758789062, -0.0084381103515625, -0.0076808929443359375, -0.006923675537109375, -0.0061664581298828125, -0.00540924072265625, -0.0046520233154296875, -0.003894805908203125, -0.0031375885009765625, -0.00238037109375, -0.0016231536865234375, -0.000865936279296875, -0.0001087188720703125, 0.00064849853515625, 0.0014057159423828125, 0.002162933349609375, 0.0029201507568359375, 0.0036773681640625, 0.0044345855712890625, 0.005191802978515625, 0.0059490203857421875, 0.00670623779296875, 0.0074634552001953125, 0.008220672607421875, 0.008977890014648438, 0.009735107421875, 0.010492324829101562, 0.011249542236328125, 0.012006759643554688, 0.01276397705078125, 0.013521194458007812, 0.014278411865234375, 0.015035629272460938, 0.0157928466796875, 0.016550064086914062, 0.017307281494140625, 0.018064498901367188, 0.01882171630859375, 0.019578933715820312, 0.020336151123046875, 0.021093368530273438, 0.0218505859375]}, "gradients/model.layers.3.self_attn.k_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 2.0, 2.0, 0.0, 2.0, 2.0, 3.0, 2.0, 2.0, 5.0, 6.0, 5.0, 5.0, 7.0, 20.0, 17.0, 22.0, 30.0, 41.0, 56.0, 102.0, 174.0, 289.0, 557.0, 1153.0, 2755.0, 6731.0, 24837.0, 260677.0, 3695140.0, 169912.0, 20817.0, 6183.0, 2420.0, 1083.0, 504.0, 265.0, 164.0, 77.0, 72.0, 36.0, 25.0, 20.0, 24.0, 15.0, 8.0, 6.0, 4.0, 4.0, 3.0, 3.0, 3.0, 1.0, 6.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.00141143798828125, -0.0013679265975952148, -0.0013244152069091797, -0.0012809038162231445, -0.0012373924255371094, -0.0011938810348510742, -0.001150369644165039, -0.001106858253479004, -0.0010633468627929688, -0.0010198354721069336, -0.0009763240814208984, -0.0009328126907348633, -0.0008893013000488281, -0.000845789909362793, -0.0008022785186767578, -0.0007587671279907227, -0.0007152557373046875, -0.0006717443466186523, -0.0006282329559326172, -0.000584721565246582, -0.0005412101745605469, -0.0004976987838745117, -0.00045418739318847656, -0.0004106760025024414, -0.00036716461181640625, -0.0003236532211303711, -0.00028014183044433594, -0.00023663043975830078, -0.00019311904907226562, -0.00014960765838623047, -0.00010609626770019531, -6.258487701416016e-05, -1.9073486328125e-05, 2.4437904357910156e-05, 6.794929504394531e-05, 0.00011146068572998047, 0.00015497207641601562, 0.00019848346710205078, 0.00024199485778808594, 0.0002855062484741211, 0.00032901763916015625, 0.0003725290298461914, 0.00041604042053222656, 0.0004595518112182617, 0.0005030632019042969, 0.000546574592590332, 0.0005900859832763672, 0.0006335973739624023, 0.0006771087646484375, 0.0007206201553344727, 0.0007641315460205078, 0.000807642936706543, 0.0008511543273925781, 0.0008946657180786133, 0.0009381771087646484, 0.0009816884994506836, 0.0010251998901367188, 0.001068711280822754, 0.001112222671508789, 0.0011557340621948242, 0.0011992454528808594, 0.0012427568435668945, 0.0012862682342529297, 0.0013297796249389648, 0.001373291015625]}, "gradients/model.layers.3.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 3.0, 2.0, 1.0, 3.0, 7.0, 2.0, 9.0, 8.0, 11.0, 9.0, 12.0, 16.0, 25.0, 33.0, 42.0, 60.0, 84.0, 113.0, 199.0, 369.0, 648.0, 1421.0, 3560.0, 10247.0, 48116.0, 2252508.0, 14341672.0, 94365.0, 15106.0, 4749.0, 1835.0, 854.0, 387.0, 229.0, 135.0, 99.0, 65.0, 37.0, 38.0, 32.0, 24.0, 18.0, 13.0, 9.0, 4.0, 8.0, 9.0, 2.0, 3.0, 2.0, 4.0, 1.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.0021209716796875, -0.002058863639831543, -0.001996755599975586, -0.001934647560119629, -0.0018725395202636719, -0.0018104314804077148, -0.0017483234405517578, -0.0016862154006958008, -0.0016241073608398438, -0.0015619993209838867, -0.0014998912811279297, -0.0014377832412719727, -0.0013756752014160156, -0.0013135671615600586, -0.0012514591217041016, -0.0011893510818481445, -0.0011272430419921875, -0.0010651350021362305, -0.0010030269622802734, -0.0009409189224243164, -0.0008788108825683594, -0.0008167028427124023, -0.0007545948028564453, -0.0006924867630004883, -0.0006303787231445312, -0.0005682706832885742, -0.0005061626434326172, -0.00044405460357666016, -0.0003819465637207031, -0.0003198385238647461, -0.00025773048400878906, -0.00019562244415283203, -0.000133514404296875, -7.140636444091797e-05, -9.298324584960938e-06, 5.2809715270996094e-05, 0.00011491775512695312, 0.00017702579498291016, 0.0002391338348388672, 0.0003012418746948242, 0.00036334991455078125, 0.0004254579544067383, 0.0004875659942626953, 0.0005496740341186523, 0.0006117820739746094, 0.0006738901138305664, 0.0007359981536865234, 0.0007981061935424805, 0.0008602142333984375, 0.0009223222732543945, 0.0009844303131103516, 0.0010465383529663086, 0.0011086463928222656, 0.0011707544326782227, 0.0012328624725341797, 0.0012949705123901367, 0.0013570785522460938, 0.0014191865921020508, 0.0014812946319580078, 0.0015434026718139648, 0.0016055107116699219, 0.001667618751525879, 0.001729726791381836, 0.001791834831237793, 0.00185394287109375]}, "gradients/model.layers.3.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 1.0, 2.0, 3.0, 5.0, 6.0, 16.0, 118.0, 1417.0, 2174.0, 314.0, 13.0, 10.0, 6.0, 3.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.0009918212890625, -0.0008838176727294922, -0.0007758140563964844, -0.0006678104400634766, -0.0005598068237304688, -0.00045180320739746094, -0.0003437995910644531, -0.0002357959747314453, -0.0001277923583984375, -1.9788742065429688e-05, 8.821487426757812e-05, 0.00019621849060058594, 0.00030422210693359375, 0.00041222572326660156, 0.0005202293395996094, 0.0006282329559326172, 0.000736236572265625, 0.0008442401885986328, 0.0009522438049316406, 0.0010602474212646484, 0.0011682510375976562, 0.001276254653930664, 0.0013842582702636719, 0.0014922618865966797, 0.0016002655029296875, 0.0017082691192626953, 0.0018162727355957031, 0.001924276351928711, 0.0020322799682617188, 0.0021402835845947266, 0.0022482872009277344, 0.002356290817260742, 0.00246429443359375, 0.002572298049926758, 0.0026803016662597656, 0.0027883052825927734, 0.0028963088989257812, 0.003004312515258789, 0.003112316131591797, 0.0032203197479248047, 0.0033283233642578125, 0.0034363269805908203, 0.003544330596923828, 0.003652334213256836, 0.0037603378295898438, 0.0038683414459228516, 0.003976345062255859, 0.004084348678588867, 0.004192352294921875, 0.004300355911254883, 0.004408359527587891, 0.0045163631439208984, 0.004624366760253906, 0.004732370376586914, 0.004840373992919922, 0.00494837760925293, 0.0050563812255859375, 0.005164384841918945, 0.005272388458251953, 0.005380392074584961, 0.005488395690917969, 0.0055963993072509766, 0.005704402923583984, 0.005812406539916992, 0.00592041015625]}, "gradients/model.layers.2.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 3.0, 4.0, 5.0, 6.0, 6.0, 7.0, 18.0, 20.0, 34.0, 32.0, 42.0, 50.0, 65.0, 77.0, 104.0, 118.0, 128.0, 186.0, 251.0, 450.0, 1329.0, 8000.0, 128507.0, 16777216.0, 16777216.0, 128003.0, 8100.0, 1292.0, 478.0, 288.0, 183.0, 121.0, 118.0, 82.0, 75.0, 74.0, 49.0, 41.0, 26.0, 22.0, 15.0, 10.0, 17.0, 11.0, 3.0, 4.0, 6.0, 3.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0035400390625, -0.0034258365631103516, -0.003311634063720703, -0.0031974315643310547, -0.0030832290649414062, -0.002969026565551758, -0.0028548240661621094, -0.002740621566772461, -0.0026264190673828125, -0.002512216567993164, -0.0023980140686035156, -0.002283811569213867, -0.0021696090698242188, -0.0020554065704345703, -0.0019412040710449219, -0.0018270015716552734, -0.001712799072265625, -0.0015985965728759766, -0.0014843940734863281, -0.0013701915740966797, -0.0012559890747070312, -0.0011417865753173828, -0.0010275840759277344, -0.0009133815765380859, -0.0007991790771484375, -0.0006849765777587891, -0.0005707740783691406, -0.0004565715789794922, -0.00034236907958984375, -0.0002281665802001953, -0.00011396408081054688, 2.384185791015625e-07, 0.00011444091796875, 0.00022864341735839844, 0.0003428459167480469, 0.0004570484161376953, 0.0005712509155273438, 0.0006854534149169922, 0.0007996559143066406, 0.0009138584136962891, 0.0010280609130859375, 0.001142263412475586, 0.0012564659118652344, 0.0013706684112548828, 0.0014848709106445312, 0.0015990734100341797, 0.0017132759094238281, 0.0018274784088134766, 0.001941680908203125, 0.0020558834075927734, 0.002170085906982422, 0.0022842884063720703, 0.0023984909057617188, 0.002512693405151367, 0.0026268959045410156, 0.002741098403930664, 0.0028553009033203125, 0.002969503402709961, 0.0030837059020996094, 0.003197908401489258, 0.0033121109008789062, 0.0034263134002685547, 0.003540515899658203, 0.0036547183990478516, 0.0037689208984375]}, "gradients/model.layers.2.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 2.0, 2.0, 5.0, 3.0, 5.0, 8.0, 7.0, 14.0, 14.0, 20.0, 27.0, 56.0, 54.0, 83.0, 131.0, 205.0, 288.0, 484.0, 977.0, 2267.0, 6235.0, 21265.0, 109038.0, 1183547.0, 16777216.0, 16777216.0, 16777216.0, 958914.0, 94118.0, 19196.0, 5592.0, 2082.0, 942.0, 518.0, 292.0, 168.0, 99.0, 89.0, 59.0, 33.0, 39.0, 21.0, 15.0, 6.0, 8.0, 8.0, 3.0, 1.0, 1.0, 3.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0], "bins": [-0.0011138916015625, -0.001078486442565918, -0.001043081283569336, -0.001007676124572754, -0.0009722709655761719, -0.0009368658065795898, -0.0009014606475830078, -0.0008660554885864258, -0.0008306503295898438, -0.0007952451705932617, -0.0007598400115966797, -0.0007244348526000977, -0.0006890296936035156, -0.0006536245346069336, -0.0006182193756103516, -0.0005828142166137695, -0.0005474090576171875, -0.0005120038986206055, -0.00047659873962402344, -0.0004411935806274414, -0.0004057884216308594, -0.00037038326263427734, -0.0003349781036376953, -0.0002995729446411133, -0.00026416778564453125, -0.00022876262664794922, -0.0001933574676513672, -0.00015795230865478516, -0.00012254714965820312, -8.71419906616211e-05, -5.173683166503906e-05, -1.633167266845703e-05, 1.9073486328125e-05, 5.447864532470703e-05, 8.988380432128906e-05, 0.0001252889633178711, 0.00016069412231445312, 0.00019609928131103516, 0.0002315044403076172, 0.0002669095993041992, 0.00030231475830078125, 0.0003377199172973633, 0.0003731250762939453, 0.00040853023529052734, 0.0004439353942871094, 0.0004793405532836914, 0.0005147457122802734, 0.0005501508712768555, 0.0005855560302734375, 0.0006209611892700195, 0.0006563663482666016, 0.0006917715072631836, 0.0007271766662597656, 0.0007625818252563477, 0.0007979869842529297, 0.0008333921432495117, 0.0008687973022460938, 0.0009042024612426758, 0.0009396076202392578, 0.0009750127792358398, 0.0010104179382324219, 0.001045823097229004, 0.001081228256225586, 0.001116633415222168, 0.00115203857421875]}, "gradients/model.layers.2.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 3.0, 0.0, 0.0, 2.0, 1.0, 2.0, 1.0, 2.0, 3.0, 4.0, 4.0, 6.0, 8.0, 5.0, 18.0, 15.0, 40.0, 87.0, 169.0, 275.0, 441.0, 566.0, 593.0, 625.0, 488.0, 313.0, 196.0, 126.0, 42.0, 20.0, 9.0, 8.0, 5.0, 4.0, 4.0, 1.0, 2.0, 0.0, 0.0, 2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0], "bins": [-0.0028228759765625, -0.0027533769607543945, -0.002683877944946289, -0.0026143789291381836, -0.002544879913330078, -0.0024753808975219727, -0.002405881881713867, -0.0023363828659057617, -0.0022668838500976562, -0.0021973848342895508, -0.0021278858184814453, -0.00205838680267334, -0.0019888877868652344, -0.001919388771057129, -0.0018498897552490234, -0.001780390739440918, -0.0017108917236328125, -0.001641392707824707, -0.0015718936920166016, -0.001502394676208496, -0.0014328956604003906, -0.0013633966445922852, -0.0012938976287841797, -0.0012243986129760742, -0.0011548995971679688, -0.0010854005813598633, -0.0010159015655517578, -0.0009464025497436523, -0.0008769035339355469, -0.0008074045181274414, -0.0007379055023193359, -0.0006684064865112305, -0.000598907470703125, -0.0005294084548950195, -0.00045990943908691406, -0.0003904104232788086, -0.0003209114074707031, -0.00025141239166259766, -0.0001819133758544922, -0.00011241436004638672, -4.291534423828125e-05, 2.658367156982422e-05, 9.608268737792969e-05, 0.00016558170318603516, 0.00023508071899414062, 0.0003045797348022461, 0.00037407875061035156, 0.00044357776641845703, 0.0005130767822265625, 0.000582575798034668, 0.0006520748138427734, 0.0007215738296508789, 0.0007910728454589844, 0.0008605718612670898, 0.0009300708770751953, 0.0009995698928833008, 0.0010690689086914062, 0.0011385679244995117, 0.0012080669403076172, 0.0012775659561157227, 0.0013470649719238281, 0.0014165639877319336, 0.001486063003540039, 0.0015555620193481445, 0.00162506103515625]}, "gradients/model.layers.2.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 3.0, 2.0, 0.0, 2.0, 5.0, 6.0, 9.0, 13.0, 25.0, 24.0, 51.0, 71.0, 124.0, 152.0, 233.0, 371.0, 595.0, 827.0, 1362.0, 1997.0, 3006.0, 4629.0, 7798.0, 14801.0, 35323.0, 130843.0, 648454.0, 3529230.0, 8675547.0, 2996594.0, 546704.0, 112901.0, 31492.0, 13894.0, 7206.0, 4470.0, 2830.0, 1857.0, 1290.0, 895.0, 538.0, 344.0, 261.0, 142.0, 102.0, 60.0, 50.0, 26.0, 23.0, 10.0, 5.0, 4.0, 3.0, 3.0, 4.0, 1.0, 2.0], "bins": [-0.00168609619140625, -0.0016385316848754883, -0.0015909671783447266, -0.0015434026718139648, -0.0014958381652832031, -0.0014482736587524414, -0.0014007091522216797, -0.001353144645690918, -0.0013055801391601562, -0.0012580156326293945, -0.0012104511260986328, -0.001162886619567871, -0.0011153221130371094, -0.0010677576065063477, -0.001020193099975586, -0.0009726285934448242, -0.0009250640869140625, -0.0008774995803833008, -0.0008299350738525391, -0.0007823705673217773, -0.0007348060607910156, -0.0006872415542602539, -0.0006396770477294922, -0.0005921125411987305, -0.0005445480346679688, -0.000496983528137207, -0.0004494190216064453, -0.0004018545150756836, -0.0003542900085449219, -0.00030672550201416016, -0.00025916099548339844, -0.00021159648895263672, -0.000164031982421875, -0.00011646747589111328, -6.890296936035156e-05, -2.1338462829589844e-05, 2.6226043701171875e-05, 7.37905502319336e-05, 0.00012135505676269531, 0.00016891956329345703, 0.00021648406982421875, 0.00026404857635498047, 0.0003116130828857422, 0.0003591775894165039, 0.0004067420959472656, 0.00045430660247802734, 0.0005018711090087891, 0.0005494356155395508, 0.0005970001220703125, 0.0006445646286010742, 0.0006921291351318359, 0.0007396936416625977, 0.0007872581481933594, 0.0008348226547241211, 0.0008823871612548828, 0.0009299516677856445, 0.0009775161743164062, 0.001025080680847168, 0.0010726451873779297, 0.0011202096939086914, 0.0011677742004394531, 0.0012153387069702148, 0.0012629032135009766, 0.0013104677200317383, 0.0013580322265625]}, "gradients/model.layers.2.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 1.0, 2.0, 4.0, 2.0, 1.0, 4.0, 6.0, 10.0, 3.0, 10.0, 11.0, 10.0, 10.0, 13.0, 16.0, 16.0, 17.0, 16.0, 35.0, 28.0, 41.0, 29.0, 31.0, 27.0, 36.0, 48.0, 210.0, 7615.0, 4144101.0, 40933.0, 624.0, 56.0, 39.0, 30.0, 30.0, 34.0, 26.0, 24.0, 19.0, 23.0, 16.0, 14.0, 14.0, 10.0, 10.0, 7.0, 4.0, 9.0, 5.0, 4.0, 3.0, 5.0, 2.0, 0.0, 2.0, 1.0, 2.0, 1.0, 0.0, 2.0], "bins": [-0.035400390625, -0.0343170166015625, -0.033233642578125, -0.0321502685546875, -0.03106689453125, -0.0299835205078125, -0.028900146484375, -0.0278167724609375, -0.0267333984375, -0.0256500244140625, -0.024566650390625, -0.0234832763671875, -0.02239990234375, -0.0213165283203125, -0.020233154296875, -0.0191497802734375, -0.01806640625, -0.0169830322265625, -0.015899658203125, -0.0148162841796875, -0.01373291015625, -0.0126495361328125, -0.011566162109375, -0.0104827880859375, -0.0093994140625, -0.0083160400390625, -0.007232666015625, -0.0061492919921875, -0.00506591796875, -0.0039825439453125, -0.002899169921875, -0.0018157958984375, -0.000732421875, 0.0003509521484375, 0.001434326171875, 0.0025177001953125, 0.00360107421875, 0.0046844482421875, 0.005767822265625, 0.0068511962890625, 0.0079345703125, 0.0090179443359375, 0.010101318359375, 0.0111846923828125, 0.01226806640625, 0.0133514404296875, 0.014434814453125, 0.0155181884765625, 0.0166015625, 0.0176849365234375, 0.018768310546875, 0.0198516845703125, 0.02093505859375, 0.0220184326171875, 0.023101806640625, 0.0241851806640625, 0.0252685546875, 0.0263519287109375, 0.027435302734375, 0.0285186767578125, 0.02960205078125, 0.0306854248046875, 0.031768798828125, 0.0328521728515625, 0.033935546875]}, "gradients/model.layers.2.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 2.0, 3.0, 1.0, 2.0, 4.0, 5.0, 6.0, 7.0, 17.0, 9.0, 25.0, 25.0, 34.0, 47.0, 77.0, 107.0, 178.0, 317.0, 569.0, 1109.0, 2708.0, 10181.0, 85647.0, 3951879.0, 122986.0, 12499.0, 3201.0, 1191.0, 566.0, 331.0, 180.0, 108.0, 68.0, 55.0, 34.0, 36.0, 22.0, 18.0, 12.0, 12.0, 3.0, 4.0, 3.0, 1.0, 1.0, 5.0, 1.0, 1.0, 2.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00225830078125, -0.0021843910217285156, -0.0021104812622070312, -0.002036571502685547, -0.0019626617431640625, -0.0018887519836425781, -0.0018148422241210938, -0.0017409324645996094, -0.001667022705078125, -0.0015931129455566406, -0.0015192031860351562, -0.0014452934265136719, -0.0013713836669921875, -0.0012974739074707031, -0.0012235641479492188, -0.0011496543884277344, -0.00107574462890625, -0.0010018348693847656, -0.0009279251098632812, -0.0008540153503417969, -0.0007801055908203125, -0.0007061958312988281, -0.0006322860717773438, -0.0005583763122558594, -0.000484466552734375, -0.0004105567932128906, -0.00033664703369140625, -0.0002627372741699219, -0.0001888275146484375, -0.00011491775512695312, -4.100799560546875e-05, 3.2901763916015625e-05, 0.0001068115234375, 0.00018072128295898438, 0.00025463104248046875, 0.0003285408020019531, 0.0004024505615234375, 0.0004763603210449219, 0.0005502700805664062, 0.0006241798400878906, 0.000698089599609375, 0.0007719993591308594, 0.0008459091186523438, 0.0009198188781738281, 0.0009937286376953125, 0.0010676383972167969, 0.0011415481567382812, 0.0012154579162597656, 0.00128936767578125, 0.0013632774353027344, 0.0014371871948242188, 0.0015110969543457031, 0.0015850067138671875, 0.0016589164733886719, 0.0017328262329101562, 0.0018067359924316406, 0.001880645751953125, 0.0019545555114746094, 0.0020284652709960938, 0.002102375030517578, 0.0021762847900390625, 0.002250194549560547, 0.0023241043090820312, 0.0023980140686035156, 0.002471923828125]}, "gradients/model.layers.2.self_attn.q_proj.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 1.0, 2.0, 0.0, 4.0, 4.0, 7.0, 9.0, 6.0, 16.0, 14.0, 24.0, 16.0, 20.0, 35.0, 45.0, 64.0, 93.0, 130.0, 182.0, 302.0, 510.0, 884.0, 1684.0, 4307.0, 14302.0, 90654.0, 15479770.0, 1127201.0, 41487.0, 9019.0, 3056.0, 1329.0, 708.0, 404.0, 253.0, 173.0, 138.0, 81.0, 62.0, 46.0, 43.0, 23.0, 27.0, 18.0, 10.0, 8.0, 7.0, 6.0, 9.0, 3.0, 3.0, 6.0, 1.0, 3.0, 1.0], "bins": [-0.0030059814453125, -0.00292205810546875, -0.002838134765625, -0.00275421142578125, -0.0026702880859375, -0.00258636474609375, -0.00250244140625, -0.00241851806640625, -0.0023345947265625, -0.00225067138671875, -0.002166748046875, -0.00208282470703125, -0.0019989013671875, -0.00191497802734375, -0.0018310546875, -0.00174713134765625, -0.0016632080078125, -0.00157928466796875, -0.001495361328125, -0.00141143798828125, -0.0013275146484375, -0.00124359130859375, -0.00115966796875, -0.00107574462890625, -0.0009918212890625, -0.00090789794921875, -0.000823974609375, -0.00074005126953125, -0.0006561279296875, -0.00057220458984375, -0.00048828125, -0.00040435791015625, -0.0003204345703125, -0.00023651123046875, -0.000152587890625, -6.866455078125e-05, 1.52587890625e-05, 9.918212890625e-05, 0.00018310546875, 0.00026702880859375, 0.0003509521484375, 0.00043487548828125, 0.000518798828125, 0.00060272216796875, 0.0006866455078125, 0.00077056884765625, 0.0008544921875, 0.00093841552734375, 0.0010223388671875, 0.00110626220703125, 0.001190185546875, 0.00127410888671875, 0.0013580322265625, 0.00144195556640625, 0.00152587890625, 0.00160980224609375, 0.0016937255859375, 0.00177764892578125, 0.001861572265625, 0.00194549560546875, 0.0020294189453125, 0.00211334228515625, 0.002197265625, 0.00228118896484375, 0.0023651123046875]}, "gradients/model.layers.2.input_layernorm.weight": {"_type": "histogram", "values": [117.0, 3918.0, 43.0, 13.0, 2.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00066375732421875, -0.00013840198516845703, 0.00038695335388183594, 0.0009123086929321289, 0.0014376640319824219, 0.001963019371032715, 0.002488374710083008, 0.0030137300491333008, 0.0035390853881835938, 0.004064440727233887, 0.00458979606628418, 0.005115151405334473, 0.005640506744384766, 0.006165862083435059, 0.0066912174224853516, 0.0072165727615356445, 0.0077419281005859375, 0.00826728343963623, 0.008792638778686523, 0.009317994117736816, 0.00984334945678711, 0.010368704795837402, 0.010894060134887695, 0.011419415473937988, 0.011944770812988281, 0.012470126152038574, 0.012995481491088867, 0.01352083683013916, 0.014046192169189453, 0.014571547508239746, 0.015096902847290039, 0.015622258186340332, 0.016147613525390625, 0.016672968864440918, 0.01719832420349121, 0.017723679542541504, 0.018249034881591797, 0.01877439022064209, 0.019299745559692383, 0.019825100898742676, 0.02035045623779297, 0.02087581157684326, 0.021401166915893555, 0.021926522254943848, 0.02245187759399414, 0.022977232933044434, 0.023502588272094727, 0.02402794361114502, 0.024553298950195312, 0.025078654289245605, 0.0256040096282959, 0.02612936496734619, 0.026654720306396484, 0.027180075645446777, 0.02770543098449707, 0.028230786323547363, 0.028756141662597656, 0.02928149700164795, 0.029806852340698242, 0.030332207679748535, 0.030857563018798828, 0.03138291835784912, 0.031908273696899414, 0.03243362903594971, 0.032958984375]}, "gradients/model.layers.1.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [1.0, 1.0, 1.0, 0.0, 2.0, 4.0, 5.0, 3.0, 7.0, 9.0, 16.0, 16.0, 22.0, 22.0, 25.0, 30.0, 45.0, 56.0, 53.0, 79.0, 94.0, 93.0, 126.0, 114.0, 131.0, 142.0, 158.0, 175.0, 290.0, 747.0, 16777216.0, 5569430.0, 718.0, 281.0, 210.0, 178.0, 142.0, 145.0, 131.0, 113.0, 78.0, 86.0, 92.0, 57.0, 46.0, 41.0, 38.0, 26.0, 25.0, 7.0, 6.0, 10.0, 5.0, 4.0, 3.0, 5.0, 3.0, 2.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.1201171875, -0.116241455078125, -0.11236572265625, -0.108489990234375, -0.1046142578125, -0.100738525390625, -0.09686279296875, -0.092987060546875, -0.089111328125, -0.085235595703125, -0.08135986328125, -0.077484130859375, -0.0736083984375, -0.069732666015625, -0.06585693359375, -0.061981201171875, -0.05810546875, -0.054229736328125, -0.05035400390625, -0.046478271484375, -0.0426025390625, -0.038726806640625, -0.03485107421875, -0.030975341796875, -0.027099609375, -0.023223876953125, -0.01934814453125, -0.015472412109375, -0.0115966796875, -0.007720947265625, -0.00384521484375, 3.0517578125e-05, 0.00390625, 0.007781982421875, 0.01165771484375, 0.015533447265625, 0.0194091796875, 0.023284912109375, 0.02716064453125, 0.031036376953125, 0.034912109375, 0.038787841796875, 0.04266357421875, 0.046539306640625, 0.0504150390625, 0.054290771484375, 0.05816650390625, 0.062042236328125, 0.06591796875, 0.069793701171875, 0.07366943359375, 0.077545166015625, 0.0814208984375, 0.085296630859375, 0.08917236328125, 0.093048095703125, 0.096923828125, 0.100799560546875, 0.10467529296875, 0.108551025390625, 0.1124267578125, 0.116302490234375, 0.12017822265625, 0.124053955078125, 0.1279296875]}, "gradients/model.layers.1.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 1.0, 2.0, 2.0, 1.0, 2.0, 3.0, 1.0, 0.0, 1.0, 1.0, 4.0, 5.0, 4.0, 8.0, 26.0, 69.0, 271.0, 1105.0, 6746.0, 85070.0, 7986881.0, 16777216.0, 2627307.0, 50097.0, 4775.0, 897.0, 204.0, 73.0, 18.0, 13.0, 5.0, 3.0, 0.0, 1.0, 0.0, 0.0, 1.0, 2.0, 0.0, 0.0, 2.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 2.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00174713134765625, -0.0016809701919555664, -0.0016148090362548828, -0.0015486478805541992, -0.0014824867248535156, -0.001416325569152832, -0.0013501644134521484, -0.0012840032577514648, -0.0012178421020507812, -0.0011516809463500977, -0.001085519790649414, -0.0010193586349487305, -0.0009531974792480469, -0.0008870363235473633, -0.0008208751678466797, -0.0007547140121459961, -0.0006885528564453125, -0.0006223917007446289, -0.0005562305450439453, -0.0004900693893432617, -0.0004239082336425781, -0.00035774707794189453, -0.00029158592224121094, -0.00022542476654052734, -0.00015926361083984375, -9.310245513916016e-05, -2.6941299438476562e-05, 3.921985626220703e-05, 0.00010538101196289062, 0.00017154216766357422, 0.0002377033233642578, 0.0003038644790649414, 0.000370025634765625, 0.0004361867904663086, 0.0005023479461669922, 0.0005685091018676758, 0.0006346702575683594, 0.000700831413269043, 0.0007669925689697266, 0.0008331537246704102, 0.0008993148803710938, 0.0009654760360717773, 0.001031637191772461, 0.0010977983474731445, 0.0011639595031738281, 0.0012301206588745117, 0.0012962818145751953, 0.001362442970275879, 0.0014286041259765625, 0.001494765281677246, 0.0015609264373779297, 0.0016270875930786133, 0.0016932487487792969, 0.0017594099044799805, 0.001825571060180664, 0.0018917322158813477, 0.0019578933715820312, 0.002024054527282715, 0.0020902156829833984, 0.002156376838684082, 0.0022225379943847656, 0.0022886991500854492, 0.002354860305786133, 0.0024210214614868164, 0.0024871826171875]}, "gradients/model.layers.1.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 4.0, 2.0, 4.0, 2.0, 1.0, 6.0, 2.0, 3.0, 6.0, 8.0, 7.0, 8.0, 24.0, 84.0, 243.0, 465.0, 750.0, 862.0, 704.0, 464.0, 214.0, 99.0, 38.0, 27.0, 13.0, 11.0, 10.0, 6.0, 8.0, 3.0, 3.0, 5.0, 3.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0033416748046875, -0.0032389163970947266, -0.003136157989501953, -0.0030333995819091797, -0.0029306411743164062, -0.002827882766723633, -0.0027251243591308594, -0.002622365951538086, -0.0025196075439453125, -0.002416849136352539, -0.0023140907287597656, -0.002211332321166992, -0.0021085739135742188, -0.0020058155059814453, -0.0019030570983886719, -0.0018002986907958984, -0.001697540283203125, -0.0015947818756103516, -0.0014920234680175781, -0.0013892650604248047, -0.0012865066528320312, -0.0011837482452392578, -0.0010809898376464844, -0.000978231430053711, -0.0008754730224609375, -0.0007727146148681641, -0.0006699562072753906, -0.0005671977996826172, -0.00046443939208984375, -0.0003616809844970703, -0.0002589225769042969, -0.00015616416931152344, -5.340576171875e-05, 4.935264587402344e-05, 0.00015211105346679688, 0.0002548694610595703, 0.00035762786865234375, 0.0004603862762451172, 0.0005631446838378906, 0.0006659030914306641, 0.0007686614990234375, 0.0008714199066162109, 0.0009741783142089844, 0.0010769367218017578, 0.0011796951293945312, 0.0012824535369873047, 0.0013852119445800781, 0.0014879703521728516, 0.001590728759765625, 0.0016934871673583984, 0.0017962455749511719, 0.0018990039825439453, 0.0020017623901367188, 0.002104520797729492, 0.0022072792053222656, 0.002310037612915039, 0.0024127960205078125, 0.002515554428100586, 0.0026183128356933594, 0.002721071243286133, 0.0028238296508789062, 0.0029265880584716797, 0.003029346466064453, 0.0031321048736572266, 0.00323486328125]}, "gradients/model.layers.1.self_attn.o_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 3.0, 0.0, 1.0, 1.0, 4.0, 2.0, 5.0, 6.0, 11.0, 24.0, 37.0, 63.0, 83.0, 143.0, 184.0, 335.0, 605.0, 1020.0, 1590.0, 3070.0, 5488.0, 11672.0, 25019.0, 68120.0, 210842.0, 765796.0, 3029580.0, 7957381.0, 3442159.0, 884452.0, 239841.0, 76003.0, 27803.0, 12224.0, 5942.0, 3191.0, 1797.0, 1083.0, 620.0, 414.0, 229.0, 128.0, 97.0, 45.0, 35.0, 21.0, 10.0, 13.0, 10.0, 7.0, 2.0, 2.0, 0.0, 0.0, 1.0], "bins": [-0.00165557861328125, -0.0016102790832519531, -0.0015649795532226562, -0.0015196800231933594, -0.0014743804931640625, -0.0014290809631347656, -0.0013837814331054688, -0.0013384819030761719, -0.001293182373046875, -0.0012478828430175781, -0.0012025833129882812, -0.0011572837829589844, -0.0011119842529296875, -0.0010666847229003906, -0.0010213851928710938, -0.0009760856628417969, -0.0009307861328125, -0.0008854866027832031, -0.0008401870727539062, -0.0007948875427246094, -0.0007495880126953125, -0.0007042884826660156, -0.0006589889526367188, -0.0006136894226074219, -0.000568389892578125, -0.0005230903625488281, -0.00047779083251953125, -0.0004324913024902344, -0.0003871917724609375, -0.0003418922424316406, -0.00029659271240234375, -0.0002512931823730469, -0.00020599365234375, -0.00016069412231445312, -0.00011539459228515625, -7.009506225585938e-05, -2.47955322265625e-05, 2.0503997802734375e-05, 6.580352783203125e-05, 0.00011110305786132812, 0.000156402587890625, 0.00020170211791992188, 0.00024700164794921875, 0.0002923011779785156, 0.0003376007080078125, 0.0003829002380371094, 0.00042819976806640625, 0.0004734992980957031, 0.000518798828125, 0.0005640983581542969, 0.0006093978881835938, 0.0006546974182128906, 0.0006999969482421875, 0.0007452964782714844, 0.0007905960083007812, 0.0008358955383300781, 0.000881195068359375, 0.0009264945983886719, 0.0009717941284179688, 0.0010170936584472656, 0.0010623931884765625, 0.0011076927185058594, 0.0011529922485351562, 0.0011982917785644531, 0.00124359130859375]}, "gradients/model.layers.1.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 4.0, 1.0, 3.0, 2.0, 4.0, 6.0, 5.0, 9.0, 5.0, 15.0, 11.0, 22.0, 26.0, 22.0, 31.0, 27.0, 36.0, 37.0, 42.0, 83.0, 255.0, 1659.0, 19129.0, 4134673.0, 34981.0, 2431.0, 356.0, 85.0, 54.0, 49.0, 35.0, 25.0, 28.0, 33.0, 23.0, 27.0, 14.0, 13.0, 9.0, 6.0, 5.0, 3.0, 2.0, 3.0, 2.0, 2.0, 2.0, 0.0, 2.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 1.0], "bins": [-0.039794921875, -0.038494110107421875, -0.03719329833984375, -0.035892486572265625, -0.0345916748046875, -0.033290863037109375, -0.03199005126953125, -0.030689239501953125, -0.029388427734375, -0.028087615966796875, -0.02678680419921875, -0.025485992431640625, -0.0241851806640625, -0.022884368896484375, -0.02158355712890625, -0.020282745361328125, -0.01898193359375, -0.017681121826171875, -0.01638031005859375, -0.015079498291015625, -0.0137786865234375, -0.012477874755859375, -0.01117706298828125, -0.009876251220703125, -0.008575439453125, -0.007274627685546875, -0.00597381591796875, -0.004673004150390625, -0.0033721923828125, -0.002071380615234375, -0.00077056884765625, 0.000530242919921875, 0.0018310546875, 0.003131866455078125, 0.00443267822265625, 0.005733489990234375, 0.0070343017578125, 0.008335113525390625, 0.00963592529296875, 0.010936737060546875, 0.012237548828125, 0.013538360595703125, 0.01483917236328125, 0.016139984130859375, 0.0174407958984375, 0.018741607666015625, 0.02004241943359375, 0.021343231201171875, 0.02264404296875, 0.023944854736328125, 0.02524566650390625, 0.026546478271484375, 0.0278472900390625, 0.029148101806640625, 0.03044891357421875, 0.031749725341796875, 0.033050537109375, 0.034351348876953125, 0.03565216064453125, 0.036952972412109375, 0.0382537841796875, 0.039554595947265625, 0.04085540771484375, 0.042156219482421875, 0.04345703125]}, "gradients/model.layers.1.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 1.0, 0.0, 2.0, 2.0, 3.0, 2.0, 5.0, 10.0, 2.0, 11.0, 16.0, 17.0, 44.0, 58.0, 108.0, 161.0, 286.0, 608.0, 1364.0, 3861.0, 14630.0, 251041.0, 3886474.0, 26307.0, 5643.0, 1991.0, 788.0, 387.0, 180.0, 113.0, 67.0, 38.0, 22.0, 24.0, 6.0, 8.0, 2.0, 5.0, 4.0, 1.0, 3.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.00180816650390625, -0.0017552375793457031, -0.0017023086547851562, -0.0016493797302246094, -0.0015964508056640625, -0.0015435218811035156, -0.0014905929565429688, -0.0014376640319824219, -0.001384735107421875, -0.0013318061828613281, -0.0012788772583007812, -0.0012259483337402344, -0.0011730194091796875, -0.0011200904846191406, -0.0010671615600585938, -0.0010142326354980469, -0.0009613037109375, -0.0009083747863769531, -0.0008554458618164062, -0.0008025169372558594, -0.0007495880126953125, -0.0006966590881347656, -0.0006437301635742188, -0.0005908012390136719, -0.000537872314453125, -0.0004849433898925781, -0.00043201446533203125, -0.0003790855407714844, -0.0003261566162109375, -0.0002732276916503906, -0.00022029876708984375, -0.00016736984252929688, -0.00011444091796875, -6.151199340820312e-05, -8.58306884765625e-06, 4.4345855712890625e-05, 9.72747802734375e-05, 0.00015020370483398438, 0.00020313262939453125, 0.0002560615539550781, 0.000308990478515625, 0.0003619194030761719, 0.00041484832763671875, 0.0004677772521972656, 0.0005207061767578125, 0.0005736351013183594, 0.0006265640258789062, 0.0006794929504394531, 0.000732421875, 0.0007853507995605469, 0.0008382797241210938, 0.0008912086486816406, 0.0009441375732421875, 0.0009970664978027344, 0.0010499954223632812, 0.0011029243469238281, 0.001155853271484375, 0.0012087821960449219, 0.0012617111206054688, 0.0013146400451660156, 0.0013675689697265625, 0.0014204978942871094, 0.0014734268188476562, 0.0015263557434082031, 0.00157928466796875]}, "gradients/model.layers.1.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 2.0, 0.0, 5.0, 1.0, 1.0, 3.0, 5.0, 3.0, 15.0, 5.0, 11.0, 17.0, 21.0, 33.0, 53.0, 88.0, 116.0, 200.0, 382.0, 646.0, 1183.0, 2545.0, 6507.0, 19061.0, 87100.0, 15883226.0, 710090.0, 45361.0, 11930.0, 4354.0, 1984.0, 940.0, 542.0, 266.0, 194.0, 94.0, 57.0, 56.0, 34.0, 21.0, 13.0, 11.0, 11.0, 7.0, 5.0, 4.0, 3.0, 4.0, 1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 1.0], "bins": [-0.0013885498046875, -0.0013462305068969727, -0.0013039112091064453, -0.001261591911315918, -0.0012192726135253906, -0.0011769533157348633, -0.001134634017944336, -0.0010923147201538086, -0.0010499954223632812, -0.001007676124572754, -0.0009653568267822266, -0.0009230375289916992, -0.0008807182312011719, -0.0008383989334106445, -0.0007960796356201172, -0.0007537603378295898, -0.0007114410400390625, -0.0006691217422485352, -0.0006268024444580078, -0.0005844831466674805, -0.0005421638488769531, -0.0004998445510864258, -0.00045752525329589844, -0.0004152059555053711, -0.00037288665771484375, -0.0003305673599243164, -0.00028824806213378906, -0.0002459287643432617, -0.00020360946655273438, -0.00016129016876220703, -0.00011897087097167969, -7.665157318115234e-05, -3.4332275390625e-05, 7.987022399902344e-06, 5.030632019042969e-05, 9.262561798095703e-05, 0.00013494491577148438, 0.00017726421356201172, 0.00021958351135253906, 0.0002619028091430664, 0.00030422210693359375, 0.0003465414047241211, 0.00038886070251464844, 0.0004311800003051758, 0.0004734992980957031, 0.0005158185958862305, 0.0005581378936767578, 0.0006004571914672852, 0.0006427764892578125, 0.0006850957870483398, 0.0007274150848388672, 0.0007697343826293945, 0.0008120536804199219, 0.0008543729782104492, 0.0008966922760009766, 0.0009390115737915039, 0.0009813308715820312, 0.0010236501693725586, 0.001065969467163086, 0.0011082887649536133, 0.0011506080627441406, 0.001192927360534668, 0.0012352466583251953, 0.0012775659561157227, 0.00131988525390625]}, "gradients/model.layers.1.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 4.0, 1.0, 8.0, 8.0, 8.0, 25.0, 36.0, 330.0, 1591.0, 1652.0, 376.0, 37.0, 7.0, 7.0, 2.0, 0.0, 1.0], "bins": [-0.01025390625, -0.01007091999053955, -0.009887933731079102, -0.009704947471618652, -0.009521961212158203, -0.009338974952697754, -0.009155988693237305, -0.008973002433776855, -0.008790016174316406, -0.008607029914855957, -0.008424043655395508, -0.008241057395935059, -0.00805807113647461, -0.00787508487701416, -0.007692098617553711, -0.007509112358093262, -0.0073261260986328125, -0.007143139839172363, -0.006960153579711914, -0.006777167320251465, -0.006594181060791016, -0.006411194801330566, -0.006228208541870117, -0.006045222282409668, -0.005862236022949219, -0.0056792497634887695, -0.00549626350402832, -0.005313277244567871, -0.005130290985107422, -0.004947304725646973, -0.0047643184661865234, -0.004581332206726074, -0.004398345947265625, -0.004215359687805176, -0.0040323734283447266, -0.0038493871688842773, -0.003666400909423828, -0.003483414649963379, -0.0033004283905029297, -0.0031174421310424805, -0.0029344558715820312, -0.002751469612121582, -0.002568483352661133, -0.0023854970932006836, -0.0022025108337402344, -0.002019524574279785, -0.001836538314819336, -0.0016535520553588867, -0.0014705657958984375, -0.0012875795364379883, -0.001104593276977539, -0.0009216070175170898, -0.0007386207580566406, -0.0005556344985961914, -0.0003726482391357422, -0.00018966197967529297, -6.67572021484375e-06, 0.00017631053924560547, 0.0003592967987060547, 0.0005422830581665039, 0.0007252693176269531, 0.0009082555770874023, 0.0010912418365478516, 0.0012742280960083008, 0.00145721435546875]}, "gradients/model.layers.0.mlp.swiglu.w3.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 3.0, 0.0, 1.0, 1.0, 0.0, 1.0, 4.0, 3.0, 1.0, 7.0, 8.0, 9.0, 8.0, 10.0, 21.0, 21.0, 32.0, 29.0, 34.0, 51.0, 70.0, 92.0, 142.0, 202.0, 271.0, 360.0, 577.0, 2034.0, 16777216.0, 278909.0, 1093.0, 505.0, 314.0, 243.0, 173.0, 114.0, 98.0, 71.0, 56.0, 41.0, 28.0, 24.0, 17.0, 14.0, 9.0, 7.0, 6.0, 4.0, 3.0, 4.0, 1.0, 1.0, 2.0, 3.0, 0.0, 1.0, 0.0, 0.0, 0.0, 3.0], "bins": [-0.0211181640625, -0.020475387573242188, -0.019832611083984375, -0.019189834594726562, -0.01854705810546875, -0.017904281616210938, -0.017261505126953125, -0.016618728637695312, -0.0159759521484375, -0.015333175659179688, -0.014690399169921875, -0.014047622680664062, -0.01340484619140625, -0.012762069702148438, -0.012119293212890625, -0.011476516723632812, -0.010833740234375, -0.010190963745117188, -0.009548187255859375, -0.008905410766601562, -0.00826263427734375, -0.0076198577880859375, -0.006977081298828125, -0.0063343048095703125, -0.0056915283203125, -0.0050487518310546875, -0.004405975341796875, -0.0037631988525390625, -0.00312042236328125, -0.0024776458740234375, -0.001834869384765625, -0.0011920928955078125, -0.00054931640625, 9.34600830078125e-05, 0.000736236572265625, 0.0013790130615234375, 0.00202178955078125, 0.0026645660400390625, 0.003307342529296875, 0.0039501190185546875, 0.0045928955078125, 0.0052356719970703125, 0.005878448486328125, 0.0065212249755859375, 0.00716400146484375, 0.0078067779541015625, 0.008449554443359375, 0.009092330932617188, 0.009735107421875, 0.010377883911132812, 0.011020660400390625, 0.011663436889648438, 0.01230621337890625, 0.012948989868164062, 0.013591766357421875, 0.014234542846679688, 0.0148773193359375, 0.015520095825195312, 0.016162872314453125, 0.016805648803710938, 0.01744842529296875, 0.018091201782226562, 0.018733978271484375, 0.019376754760742188, 0.02001953125]}, "gradients/model.layers.0.mlp.swiglu.w12.weight": {"_type": "histogram", "values": [2.0, 0.0, 0.0, 0.0, 2.0, 0.0, 2.0, 1.0, 1.0, 3.0, 7.0, 3.0, 3.0, 6.0, 4.0, 18.0, 22.0, 26.0, 70.0, 103.0, 179.0, 349.0, 576.0, 1001.0, 2314.0, 5700.0, 19456.0, 107244.0, 1481236.0, 16777216.0, 16777216.0, 4888898.0, 235542.0, 32842.0, 8672.0, 3249.0, 1514.0, 792.0, 398.0, 203.0, 115.0, 65.0, 51.0, 14.0, 16.0, 13.0, 10.0, 5.0, 4.0, 3.0, 5.0, 3.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 2.0, 1.0, 1.0], "bins": [-0.001007080078125, -0.0009737014770507812, -0.0009403228759765625, -0.0009069442749023438, -0.000873565673828125, -0.0008401870727539062, -0.0008068084716796875, -0.0007734298706054688, -0.00074005126953125, -0.0007066726684570312, -0.0006732940673828125, -0.0006399154663085938, -0.000606536865234375, -0.0005731582641601562, -0.0005397796630859375, -0.0005064010620117188, -0.0004730224609375, -0.00043964385986328125, -0.0004062652587890625, -0.00037288665771484375, -0.000339508056640625, -0.00030612945556640625, -0.0002727508544921875, -0.00023937225341796875, -0.00020599365234375, -0.00017261505126953125, -0.0001392364501953125, -0.00010585784912109375, -7.2479248046875e-05, -3.910064697265625e-05, -5.7220458984375e-06, 2.765655517578125e-05, 6.103515625e-05, 9.441375732421875e-05, 0.0001277923583984375, 0.00016117095947265625, 0.000194549560546875, 0.00022792816162109375, 0.0002613067626953125, 0.00029468536376953125, 0.00032806396484375, 0.00036144256591796875, 0.0003948211669921875, 0.00042819976806640625, 0.000461578369140625, 0.0004949569702148438, 0.0005283355712890625, 0.0005617141723632812, 0.0005950927734375, 0.0006284713745117188, 0.0006618499755859375, 0.0006952285766601562, 0.000728607177734375, 0.0007619857788085938, 0.0007953643798828125, 0.0008287429809570312, 0.00086212158203125, 0.0008955001831054688, 0.0009288787841796875, 0.0009622573852539062, 0.000995635986328125, 0.0010290145874023438, 0.0010623931884765625, 0.0010957717895507812, 0.001129150390625]}, "gradients/model.layers.0.post_attention_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 2.0, 1.0, 0.0, 2.0, 0.0, 0.0, 1.0, 2.0, 2.0, 3.0, 1.0, 2.0, 3.0, 7.0, 12.0, 35.0, 71.0, 151.0, 279.0, 443.0, 611.0, 688.0, 630.0, 497.0, 329.0, 176.0, 75.0, 43.0, 12.0, 6.0, 5.0, 0.0, 2.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.003936767578125, -0.003850698471069336, -0.003764629364013672, -0.003678560256958008, -0.0035924911499023438, -0.0035064220428466797, -0.0034203529357910156, -0.0033342838287353516, -0.0032482147216796875, -0.0031621456146240234, -0.0030760765075683594, -0.0029900074005126953, -0.0029039382934570312, -0.002817869186401367, -0.002731800079345703, -0.002645730972290039, -0.002559661865234375, -0.002473592758178711, -0.002387523651123047, -0.002301454544067383, -0.0022153854370117188, -0.0021293163299560547, -0.0020432472229003906, -0.0019571781158447266, -0.0018711090087890625, -0.0017850399017333984, -0.0016989707946777344, -0.0016129016876220703, -0.0015268325805664062, -0.0014407634735107422, -0.0013546943664550781, -0.001268625259399414, -0.00118255615234375, -0.001096487045288086, -0.0010104179382324219, -0.0009243488311767578, -0.0008382797241210938, -0.0007522106170654297, -0.0006661415100097656, -0.0005800724029541016, -0.0004940032958984375, -0.00040793418884277344, -0.0003218650817871094, -0.0002357959747314453, -0.00014972686767578125, -6.365776062011719e-05, 2.2411346435546875e-05, 0.00010848045349121094, 0.000194549560546875, 0.00028061866760253906, 0.0003666877746582031, 0.0004527568817138672, 0.0005388259887695312, 0.0006248950958251953, 0.0007109642028808594, 0.0007970333099365234, 0.0008831024169921875, 0.0009691715240478516, 0.0010552406311035156, 0.0011413097381591797, 0.0012273788452148438, 0.0013134479522705078, 0.0013995170593261719, 0.001485586166381836, 0.0015716552734375]}, "gradients/model.layers.0.self_attn.o_proj.weight": {"_type": "histogram", "values": [2.0, 9.0, 5.0, 11.0, 13.0, 26.0, 19.0, 36.0, 55.0, 84.0, 136.0, 155.0, 184.0, 314.0, 384.0, 513.0, 733.0, 1056.0, 1570.0, 2219.0, 3312.0, 5613.0, 9587.0, 17859.0, 39196.0, 92165.0, 270241.0, 1162423.0, 8477581.0, 5608284.0, 742697.0, 197239.0, 74377.0, 31250.0, 15219.0, 7950.0, 4803.0, 3107.0, 1935.0, 1409.0, 968.0, 692.0, 480.0, 375.0, 281.0, 199.0, 139.0, 107.0, 62.0, 42.0, 28.0, 24.0, 14.0, 12.0, 7.0, 2.0, 3.0, 4.0, 1.0, 0.0, 2.0, 2.0, 0.0, 1.0], "bins": [-0.00144195556640625, -0.0013920068740844727, -0.0013420581817626953, -0.001292109489440918, -0.0012421607971191406, -0.0011922121047973633, -0.001142263412475586, -0.0010923147201538086, -0.0010423660278320312, -0.000992417335510254, -0.0009424686431884766, -0.0008925199508666992, -0.0008425712585449219, -0.0007926225662231445, -0.0007426738739013672, -0.0006927251815795898, -0.0006427764892578125, -0.0005928277969360352, -0.0005428791046142578, -0.0004929304122924805, -0.0004429817199707031, -0.0003930330276489258, -0.00034308433532714844, -0.0002931356430053711, -0.00024318695068359375, -0.0001932382583618164, -0.00014328956604003906, -9.334087371826172e-05, -4.3392181396484375e-05, 6.556510925292969e-06, 5.650520324707031e-05, 0.00010645389556884766, 0.000156402587890625, 0.00020635128021240234, 0.0002562999725341797, 0.00030624866485595703, 0.0003561973571777344, 0.0004061460494995117, 0.00045609474182128906, 0.0005060434341430664, 0.0005559921264648438, 0.0006059408187866211, 0.0006558895111083984, 0.0007058382034301758, 0.0007557868957519531, 0.0008057355880737305, 0.0008556842803955078, 0.0009056329727172852, 0.0009555816650390625, 0.0010055303573608398, 0.0010554790496826172, 0.0011054277420043945, 0.0011553764343261719, 0.0012053251266479492, 0.0012552738189697266, 0.001305222511291504, 0.0013551712036132812, 0.0014051198959350586, 0.001455068588256836, 0.0015050172805786133, 0.0015549659729003906, 0.001604914665222168, 0.0016548633575439453, 0.0017048120498657227, 0.0017547607421875]}, "gradients/model.layers.0.self_attn.v_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0, 0.0, 2.0, 0.0, 3.0, 3.0, 1.0, 5.0, 4.0, 8.0, 21.0, 14.0, 26.0, 39.0, 55.0, 83.0, 133.0, 259.0, 606.0, 8025.0, 4180426.0, 3568.0, 443.0, 209.0, 128.0, 75.0, 42.0, 34.0, 24.0, 21.0, 10.0, 14.0, 5.0, 5.0, 3.0, 3.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.1162109375, -0.11333465576171875, -0.1104583740234375, -0.10758209228515625, -0.104705810546875, -0.10182952880859375, -0.0989532470703125, -0.09607696533203125, -0.09320068359375, -0.09032440185546875, -0.0874481201171875, -0.08457183837890625, -0.081695556640625, -0.07881927490234375, -0.0759429931640625, -0.07306671142578125, -0.0701904296875, -0.06731414794921875, -0.0644378662109375, -0.06156158447265625, -0.058685302734375, -0.05580902099609375, -0.0529327392578125, -0.05005645751953125, -0.04718017578125, -0.04430389404296875, -0.0414276123046875, -0.03855133056640625, -0.035675048828125, -0.03279876708984375, -0.0299224853515625, -0.02704620361328125, -0.024169921875, -0.02129364013671875, -0.0184173583984375, -0.01554107666015625, -0.012664794921875, -0.00978851318359375, -0.0069122314453125, -0.00403594970703125, -0.00115966796875, 0.00171661376953125, 0.0045928955078125, 0.00746917724609375, 0.010345458984375, 0.01322174072265625, 0.0160980224609375, 0.01897430419921875, 0.0218505859375, 0.02472686767578125, 0.0276031494140625, 0.03047943115234375, 0.033355712890625, 0.03623199462890625, 0.0391082763671875, 0.04198455810546875, 0.04486083984375, 0.04773712158203125, 0.0506134033203125, 0.05348968505859375, 0.056365966796875, 0.05924224853515625, 0.0621185302734375, 0.06499481201171875, 0.06787109375]}, "gradients/model.layers.0.self_attn.k_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 1.0, 0.0, 0.0, 3.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 2.0, 4.0, 6.0, 16.0, 20.0, 36.0, 68.0, 117.0, 244.0, 773.0, 6375.0, 4174975.0, 10162.0, 951.0, 292.0, 107.0, 67.0, 32.0, 14.0, 11.0, 12.0, 2.0, 2.0, 1.0, 0.0, 2.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0021514892578125, -0.0020673274993896484, -0.001983165740966797, -0.0018990039825439453, -0.0018148422241210938, -0.0017306804656982422, -0.0016465187072753906, -0.001562356948852539, -0.0014781951904296875, -0.001394033432006836, -0.0013098716735839844, -0.0012257099151611328, -0.0011415481567382812, -0.0010573863983154297, -0.0009732246398925781, -0.0008890628814697266, -0.000804901123046875, -0.0007207393646240234, -0.0006365776062011719, -0.0005524158477783203, -0.00046825408935546875, -0.0003840923309326172, -0.0002999305725097656, -0.00021576881408691406, -0.0001316070556640625, -4.744529724121094e-05, 3.6716461181640625e-05, 0.00012087821960449219, 0.00020503997802734375, 0.0002892017364501953, 0.0003733634948730469, 0.00045752525329589844, 0.00054168701171875, 0.0006258487701416016, 0.0007100105285644531, 0.0007941722869873047, 0.0008783340454101562, 0.0009624958038330078, 0.0010466575622558594, 0.001130819320678711, 0.0012149810791015625, 0.001299142837524414, 0.0013833045959472656, 0.0014674663543701172, 0.0015516281127929688, 0.0016357898712158203, 0.0017199516296386719, 0.0018041133880615234, 0.001888275146484375, 0.0019724369049072266, 0.002056598663330078, 0.0021407604217529297, 0.0022249221801757812, 0.002309083938598633, 0.0023932456970214844, 0.002477407455444336, 0.0025615692138671875, 0.002645730972290039, 0.0027298927307128906, 0.002814054489135742, 0.0028982162475585938, 0.0029823780059814453, 0.003066539764404297, 0.0031507015228271484, 0.00323486328125]}, "gradients/model.layers.0.self_attn.q_proj.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0, 0.0, 2.0, 1.0, 0.0, 1.0, 0.0, 0.0, 2.0, 1.0, 1.0, 2.0, 3.0, 3.0, 1.0, 4.0, 4.0, 5.0, 2.0, 6.0, 6.0, 15.0, 17.0, 25.0, 39.0, 51.0, 110.0, 160.0, 333.0, 1016.0, 4585.0, 104235.0, 16639697.0, 23121.0, 2406.0, 697.0, 274.0, 118.0, 88.0, 37.0, 43.0, 22.0, 14.0, 16.0, 12.0, 7.0, 5.0, 7.0, 8.0, 4.0, 3.0, 1.0, 0.0, 2.0], "bins": [-0.001220703125, -0.001191854476928711, -0.0011630058288574219, -0.0011341571807861328, -0.0011053085327148438, -0.0010764598846435547, -0.0010476112365722656, -0.0010187625885009766, -0.0009899139404296875, -0.0009610652923583984, -0.0009322166442871094, -0.0009033679962158203, -0.0008745193481445312, -0.0008456707000732422, -0.0008168220520019531, -0.0007879734039306641, -0.000759124755859375, -0.0007302761077880859, -0.0007014274597167969, -0.0006725788116455078, -0.0006437301635742188, -0.0006148815155029297, -0.0005860328674316406, -0.0005571842193603516, -0.0005283355712890625, -0.0004994869232177734, -0.0004706382751464844, -0.0004417896270751953, -0.00041294097900390625, -0.0003840923309326172, -0.0003552436828613281, -0.00032639503479003906, -0.00029754638671875, -0.00026869773864746094, -0.00023984909057617188, -0.0002110004425048828, -0.00018215179443359375, -0.0001533031463623047, -0.00012445449829101562, -9.560585021972656e-05, -6.67572021484375e-05, -3.790855407714844e-05, -9.059906005859375e-06, 1.9788742065429688e-05, 4.863739013671875e-05, 7.748603820800781e-05, 0.00010633468627929688, 0.00013518333435058594, 0.000164031982421875, 0.00019288063049316406, 0.00022172927856445312, 0.0002505779266357422, 0.00027942657470703125, 0.0003082752227783203, 0.0003371238708496094, 0.00036597251892089844, 0.0003948211669921875, 0.00042366981506347656, 0.0004525184631347656, 0.0004813671112060547, 0.0005102157592773438, 0.0005390644073486328, 0.0005679130554199219, 0.0005967617034912109, 0.0006256103515625]}, "gradients/model.layers.0.input_layernorm.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 3.0, 7.0, 8.0, 21.0, 43.0, 78.0, 125.0, 215.0, 392.0, 670.0, 1029.0, 617.0, 396.0, 201.0, 138.0, 73.0, 48.0, 13.0, 7.0, 3.0, 3.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 1.0], "bins": [-0.0023651123046875, -0.002279043197631836, -0.002192974090576172, -0.002106904983520508, -0.0020208358764648438, -0.0019347667694091797, -0.0018486976623535156, -0.0017626285552978516, -0.0016765594482421875, -0.0015904903411865234, -0.0015044212341308594, -0.0014183521270751953, -0.0013322830200195312, -0.0012462139129638672, -0.0011601448059082031, -0.001074075698852539, -0.000988006591796875, -0.0009019374847412109, -0.0008158683776855469, -0.0007297992706298828, -0.0006437301635742188, -0.0005576610565185547, -0.0004715919494628906, -0.00038552284240722656, -0.0002994537353515625, -0.00021338462829589844, -0.00012731552124023438, -4.124641418457031e-05, 4.482269287109375e-05, 0.0001308917999267578, 0.00021696090698242188, 0.00030303001403808594, 0.00038909912109375, 0.00047516822814941406, 0.0005612373352050781, 0.0006473064422607422, 0.0007333755493164062, 0.0008194446563720703, 0.0009055137634277344, 0.0009915828704833984, 0.0010776519775390625, 0.0011637210845947266, 0.0012497901916503906, 0.0013358592987060547, 0.0014219284057617188, 0.0015079975128173828, 0.0015940666198730469, 0.001680135726928711, 0.001766204833984375, 0.001852273941040039, 0.0019383430480957031, 0.002024412155151367, 0.0021104812622070312, 0.0021965503692626953, 0.0022826194763183594, 0.0023686885833740234, 0.0024547576904296875, 0.0025408267974853516, 0.0026268959045410156, 0.0027129650115966797, 0.0027990341186523438, 0.002885103225708008, 0.002971172332763672, 0.003057241439819336, 0.003143310546875]}, "gradients/model.embed_tokens.weight": {"_type": "histogram", "values": [1.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0, 1.0, 1.0, 0.0, 4.0, 3.0, 2.0, 2.0, 4.0, 6.0, 7.0, 21.0, 27.0, 33.0, 47.0, 61.0, 103.0, 172.0, 253.0, 482.0, 1265.0, 5651.0, 81410.0, 16777216.0, 102858.0, 6306.0, 1389.0, 512.0, 238.0, 179.0, 100.0, 72.0, 48.0, 40.0, 30.0, 12.0, 10.0, 11.0, 6.0, 2.0, 3.0, 1.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 2.0], "bins": [-0.022705078125, -0.022006988525390625, -0.02130889892578125, -0.020610809326171875, -0.0199127197265625, -0.019214630126953125, -0.01851654052734375, -0.017818450927734375, -0.017120361328125, -0.016422271728515625, -0.01572418212890625, -0.015026092529296875, -0.0143280029296875, -0.013629913330078125, -0.01293182373046875, -0.012233734130859375, -0.01153564453125, -0.010837554931640625, -0.01013946533203125, -0.009441375732421875, -0.0087432861328125, -0.008045196533203125, -0.00734710693359375, -0.006649017333984375, -0.005950927734375, -0.005252838134765625, -0.00455474853515625, -0.003856658935546875, -0.0031585693359375, -0.002460479736328125, -0.00176239013671875, -0.001064300537109375, -0.0003662109375, 0.000331878662109375, 0.00102996826171875, 0.001728057861328125, 0.0024261474609375, 0.003124237060546875, 0.00382232666015625, 0.004520416259765625, 0.005218505859375, 0.005916595458984375, 0.00661468505859375, 0.007312774658203125, 0.0080108642578125, 0.008708953857421875, 0.00940704345703125, 0.010105133056640625, 0.01080322265625, 0.011501312255859375, 0.01219940185546875, 0.012897491455078125, 0.0135955810546875, 0.014293670654296875, 0.01499176025390625, 0.015689849853515625, 0.016387939453125, 0.017086029052734375, 0.01778411865234375, 0.018482208251953125, 0.0191802978515625, 0.019878387451171875, 0.02057647705078125, 0.021274566650390625, 0.02197265625]}}
\ No newline at end of file
diff --git a/wandb/run-20240515_132302-ky6h7dv9/logs/debug-internal.log b/wandb/run-20240515_132302-ky6h7dv9/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..f400c66bda9b6fcbf2be4e0161dac0c67193c096
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/logs/debug-internal.log
@@ -0,0 +1,27072 @@
+2024-05-15 13:23:02,552 INFO StreamThr :21069 [internal.py:wandb_internal():85] W&B internal server running at pid: 21069, started at: 2024-05-15 13:23:02.551016
+2024-05-15 13:23:02,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status
+2024-05-15 13:23:02,556 INFO WriterThread:21069 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/run-ky6h7dv9.wandb
+2024-05-15 13:23:02,556 DEBUG SenderThread:21069 [sender.py:send():378] send: header
+2024-05-15 13:23:02,565 DEBUG SenderThread:21069 [sender.py:send():378] send: run
+2024-05-15 13:23:02,937 INFO SenderThread:21069 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files
+2024-05-15 13:23:02,937 INFO SenderThread:21069 [sender.py:_start_run_threads():1123] run started: ky6h7dv9 with start time 1715779382.551269
+2024-05-15 13:23:02,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: check_version
+2024-05-15 13:23:02,948 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: check_version
+2024-05-15 13:23:03,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: run_start
+2024-05-15 13:23:03,060 DEBUG HandlerThread:21069 [system_info.py:__init__():26] System info init
+2024-05-15 13:23:03,060 DEBUG HandlerThread:21069 [system_info.py:__init__():41] System info init done
+2024-05-15 13:23:03,060 INFO HandlerThread:21069 [system_monitor.py:start():194] Starting system monitor
+2024-05-15 13:23:03,060 INFO SystemMonitor:21069 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-15 13:23:03,060 INFO HandlerThread:21069 [system_monitor.py:probe():214] Collecting system info
+2024-05-15 13:23:03,061 INFO SystemMonitor:21069 [interfaces.py:start():188] Started cpu monitoring
+2024-05-15 13:23:03,062 INFO SystemMonitor:21069 [interfaces.py:start():188] Started disk monitoring
+2024-05-15 13:23:03,063 INFO SystemMonitor:21069 [interfaces.py:start():188] Started gpu monitoring
+2024-05-15 13:23:03,065 INFO SystemMonitor:21069 [interfaces.py:start():188] Started memory monitoring
+2024-05-15 13:23:03,066 INFO SystemMonitor:21069 [interfaces.py:start():188] Started network monitoring
+2024-05-15 13:23:03,173 DEBUG HandlerThread:21069 [system_info.py:probe():150] Probing system
+2024-05-15 13:23:03,177 DEBUG HandlerThread:21069 [system_info.py:_probe_git():135] Probing git
+2024-05-15 13:23:03,186 DEBUG HandlerThread:21069 [system_info.py:_probe_git():143] Probing git done
+2024-05-15 13:23:03,187 DEBUG HandlerThread:21069 [system_info.py:probe():198] Probing system done
+2024-05-15 13:23:03,187 DEBUG HandlerThread:21069 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-15T13:23:03.174078', 'startedAt': '2024-05-15T13:23:02.544142', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '81da7d2531f19a6b443b05b627f9d692f7d71546'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1341.2250833333335, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2600.0, 'min': 800.0, 'max': 3500.0}, {'current': 1700.0, 'min': 800.0, 'max': 3500.0}, {'current': 1900.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2100.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.41, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3449.587, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 1600.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3450.866, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2199.998, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.40616607666016}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-15 13:23:03,187 INFO HandlerThread:21069 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-15 13:23:03,187 INFO HandlerThread:21069 [system_monitor.py:probe():227] Publishing system info
+2024-05-15 13:23:03,187 DEBUG HandlerThread:21069 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-15 13:23:03,652 DEBUG HandlerThread:21069 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-15 13:23:03,652 DEBUG HandlerThread:21069 [system_info.py:_save_code():44] Saving code
+2024-05-15 13:23:03,652 WARNING HandlerThread:21069 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-15 13:23:03,652 DEBUG HandlerThread:21069 [system_info.py:_save_patches():82] Saving git patches
+2024-05-15 13:23:03,668 DEBUG HandlerThread:21069 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-15 13:23:03,670 INFO HandlerThread:21069 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-15 13:23:03,677 DEBUG SenderThread:21069 [sender.py:send():378] send: files
+2024-05-15 13:23:03,677 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-15 13:23:03,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-15 13:23:03,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:23:03,892 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: python_packages
+2024-05-15 13:23:03,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:23:03,939 INFO Thread-12 :21069 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/conda-environment.yaml
+2024-05-15 13:23:03,939 INFO Thread-12 :21069 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-metadata.json
+2024-05-15 13:23:03,939 INFO Thread-12 :21069 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/requirements.txt
+2024-05-15 13:23:04,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-15 13:23:04,263 INFO wandb-upload_0:21069 [upload_job.py:push():130] Uploaded file /tmp/tmpaf362022wandb/z0thaxnb-wandb-metadata.json
+2024-05-15 13:23:04,938 INFO Thread-12 :21069 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/tmp/axolotl_config_dayfyzi6.yml
+2024-05-15 13:23:04,939 INFO Thread-12 :21069 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/tmp
+2024-05-15 13:23:09,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:23:14,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:23:14,145 DEBUG SenderThread:21069 [sender.py:send():378] send: telemetry
+2024-05-15 13:23:14,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:23:14,145 DEBUG SenderThread:21069 [sender.py:send():378] send: config
+2024-05-15 13:23:14,149 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:23:14,149 DEBUG SenderThread:21069 [sender.py:send():378] send: telemetry
+2024-05-15 13:23:14,149 DEBUG SenderThread:21069 [sender.py:send():378] send: telemetry
+2024-05-15 13:23:14,149 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:23:14,149 WARNING SenderThread:21069 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-15 13:23:14,149 DEBUG SenderThread:21069 [sender.py:send():378] send: telemetry
+2024-05-15 13:23:14,150 DEBUG SenderThread:21069 [sender.py:send():378] send: telemetry
+2024-05-15 13:23:14,150 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: log_artifact
+2024-05-15 13:23:15,389 INFO wandb-upload_0:21069 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmp9xqpg5n2
+2024-05-15 13:23:16,110 INFO SenderThread:21069 [sender.py:send_request_log_artifact():1455] logged artifact config-ky6h7dv9 - {'id': 'QXJ0aWZhY3Q6ODM3NDI1NDAx', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTE3NDA1MQ==', 'latestArtifact': None}}
+2024-05-15 13:23:16,111 DEBUG SenderThread:21069 [sender.py:send():378] send: telemetry
+2024-05-15 13:23:16,112 DEBUG SenderThread:21069 [sender.py:send():378] send: files
+2024-05-15 13:23:16,112 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file tmp/axolotl_config_dayfyzi6.yml with policy live
+2024-05-15 13:23:16,481 INFO wandb-upload_0:21069 [upload_job.py:push():130] Uploaded file /tmp/tmpaf362022wandb/1a80sfmb-tmp/axolotl_config_dayfyzi6.yml
+2024-05-15 13:23:16,945 INFO Thread-12 :21069 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:23:18,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:23:18,892 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:23:18,946 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:23:20,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:23:24,950 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:23:25,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:23:27,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:23:27,094 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:23:27,096 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:23:27,096 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:23:27,096 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:23:27,096 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:23:27,097 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:23:27,100 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:23:27,952 INFO Thread-12 :21069 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:23:28,953 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:23:31,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:23:33,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:23:33,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:23:33,893 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:23:37,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:23:37,958 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/config.yaml
+2024-05-15 13:23:42,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:23:44,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:23:44,022 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:23:44,023 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:23:44,027 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:23:44,963 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:23:46,964 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:23:48,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:23:48,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:23:48,893 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:23:53,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:23:58,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:03,067 DEBUG SystemMonitor:21069 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-15 13:24:03,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:03,072 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:24:03,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:24:03,868 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:24:03,870 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:24:03,871 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:24:03,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:24:03,893 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:24:03,975 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:24:06,977 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:24:08,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:13,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:18,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:18,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:24:18,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:24:21,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:24:21,582 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:24:21,582 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:24:21,583 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:24:21,986 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:24:22,987 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:24:23,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:28,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:33,075 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:24:33,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:24:33,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:24:33,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:24:34,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:39,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:39,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:24:39,573 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:24:39,575 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:24:39,576 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:24:39,998 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:24:40,999 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:24:44,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:48,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:24:48,893 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:24:50,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:55,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:24:58,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:24:58,263 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:24:58,264 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:24:58,265 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:24:59,010 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:25:00,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:01,011 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:25:03,079 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:25:03,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:25:03,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:25:06,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:11,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:15,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:25:15,176 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:25:15,177 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:25:15,179 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:25:16,021 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:25:16,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:17,022 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:25:18,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:25:18,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:25:18,895 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:25:22,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:27,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:32,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:33,084 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:25:34,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:25:34,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:25:34,588 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:25:34,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:25:34,822 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:25:34,823 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:25:34,824 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:25:35,033 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:25:37,034 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:25:37,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:42,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:47,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:51,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:25:51,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:25:51,387 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:25:51,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:25:51,551 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:25:51,552 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:25:51,553 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:25:52,043 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:25:53,044 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:25:53,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:25:58,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:03,089 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:26:04,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:09,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:09,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:26:09,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:26:09,208 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:26:09,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:26:09,367 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:26:09,368 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:26:09,368 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:26:10,054 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:26:11,055 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:26:14,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:19,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:24,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:26,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:26:26,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:26:26,659 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:26:26,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:26:26,903 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:26:26,904 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:26:26,905 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:26:27,065 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:26:29,066 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:26:29,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:33,095 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:26:35,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:40,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:43,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:26:43,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:26:43,752 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:26:43,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:26:43,894 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:26:43,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:26:43,898 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:26:44,076 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:26:45,076 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:26:45,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:50,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:55,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:26:59,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:26:59,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:26:59,072 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:27:01,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:02,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:27:02,123 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:27:02,124 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:27:02,125 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:27:03,088 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:27:03,101 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:27:05,089 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:27:07,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:12,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:14,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:27:14,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:27:14,940 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:27:18,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:19,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:27:19,170 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:27:19,172 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:27:19,173 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:27:20,099 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:27:21,100 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:27:23,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:28,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:29,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:27:29,938 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:27:33,109 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:27:34,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:37,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:27:37,274 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:27:37,275 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:27:37,276 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:27:38,111 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:27:39,112 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:27:39,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:44,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:44,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:27:44,939 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:27:50,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:54,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:27:54,063 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:27:54,066 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:27:54,067 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:27:54,122 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:27:55,122 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:27:56,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:27:57,124 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:27:59,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:27:59,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:27:59,939 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:28:01,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:03,113 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:28:07,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:11,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:28:11,612 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:28:11,613 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:28:11,614 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:28:12,133 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:28:12,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:13,134 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:28:14,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:28:14,939 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:28:18,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:23,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:28,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:29,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:28:29,536 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:28:29,537 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:28:29,541 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:28:29,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:28:29,939 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:28:30,145 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:28:31,145 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:28:33,120 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:28:33,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:38,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:43,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:46,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:28:46,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:28:46,215 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:28:46,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:28:46,423 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:28:46,424 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:28:46,425 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:28:47,155 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:28:48,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:49,156 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:28:53,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:28:58,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:01,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:29:01,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:29:01,497 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:29:03,125 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:29:04,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:04,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:29:04,719 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:29:04,720 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:29:04,721 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:29:05,166 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:29:07,168 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:29:09,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:14,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:17,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:29:17,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:29:17,248 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:29:20,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:21,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:29:21,412 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:29:21,412 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:29:21,413 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:29:22,178 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:29:23,179 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:29:26,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:31,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:32,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:29:32,247 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:29:33,132 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:29:37,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:38,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:29:38,807 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:29:38,807 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:29:38,808 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:29:39,189 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:29:41,190 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:29:42,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:47,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:29:47,248 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:29:48,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:53,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:29:56,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:29:56,198 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:29:56,200 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:29:56,201 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:29:56,202 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:29:57,203 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:29:59,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:02,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:30:02,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:30:02,248 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:30:03,137 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:30:05,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:10,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:13,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:30:13,140 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:30:13,141 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:30:13,143 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:30:13,214 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:30:15,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:15,215 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:30:17,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:30:17,248 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:30:20,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:25,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:30,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:31,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:30:31,928 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:30:31,929 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:30:31,930 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:30:32,226 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:30:32,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:30:32,248 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:30:33,143 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:30:33,227 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:30:36,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:41,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:46,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:48,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:30:48,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:30:48,700 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:30:48,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:30:48,849 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:30:48,849 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:30:48,850 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:30:49,237 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:30:51,238 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:30:51,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:30:56,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:01,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:03,149 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:31:03,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:31:03,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:31:03,700 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:31:06,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:31:06,130 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:31:06,132 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:31:06,133 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:31:06,247 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:31:07,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:07,248 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:31:12,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:17,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:19,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:31:19,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:31:19,137 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:31:22,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:23,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:31:23,157 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:31:23,158 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:31:23,159 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:31:23,275 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:31:25,279 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:31:28,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:33,154 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:31:34,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:31:34,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:31:34,136 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:31:34,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:39,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:40,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:31:40,517 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:31:40,518 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:31:40,518 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:31:41,305 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:31:41,305 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:31:44,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:49,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:31:49,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:31:49,136 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:31:50,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:55,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:31:59,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:31:59,414 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:31:59,414 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:31:59,415 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:32:00,316 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:32:00,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:01,317 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:32:03,162 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:32:04,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:32:04,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:32:04,137 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:32:06,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:11,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:16,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:32:16,067 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:32:16,069 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:32:16,070 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:32:16,327 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:32:17,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:17,328 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:32:19,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:32:19,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:32:19,136 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:32:22,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:27,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:32,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:33,167 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:32:34,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:32:34,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:32:34,224 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:32:34,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:32:34,432 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:32:34,433 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:32:34,434 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:32:35,340 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:32:35,340 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:32:37,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:42,436 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:47,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:50,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:32:50,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:32:50,978 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:32:50,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:32:51,168 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:32:51,169 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:32:51,169 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:32:51,350 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:32:53,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:32:53,351 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:32:58,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:03,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:03,174 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:33:08,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:08,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:33:08,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:33:08,483 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:33:08,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:33:08,668 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:33:08,669 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:33:08,669 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:33:09,361 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:33:09,361 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:33:13,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:18,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:23,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:33:23,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:33:23,484 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:33:23,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:26,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:33:26,548 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:33:26,549 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:33:26,550 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:33:27,372 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:33:27,373 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:33:29,552 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:33,179 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:33:35,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:39,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:33:39,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:33:39,532 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:33:40,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:43,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:33:43,373 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:33:43,374 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:33:43,375 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:33:43,382 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:33:45,383 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:33:46,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:51,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:33:54,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:33:54,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:33:54,531 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:33:56,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:34:01,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:34:02,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:34:02,250 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:34:02,251 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:34:02,252 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:34:02,394 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:34:03,185 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:34:03,395 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:34:07,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:34:09,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:34:09,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:34:09,533 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:34:12,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:34:17,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:34:19,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:34:19,007 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:34:19,008 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:34:19,009 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:34:19,405 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:34:19,405 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:34:21,406 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:34:23,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:34:24,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:34:24,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:34:24,532 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:34:28,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:34:33,191 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:34:34,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:34:36,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:34:36,566 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:34:36,567 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:34:36,569 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:34:37,416 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:34:37,416 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:34:39,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:34:39,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:34:39,532 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:34:44,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:34:49,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:34:50,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:34:53,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:34:53,577 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:34:53,578 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:34:53,580 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:34:54,426 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:34:54,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:34:54,533 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:34:55,427 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:34:55,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:00,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:03,197 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:35:06,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:10,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:35:10,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:35:10,751 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:35:10,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:35:10,919 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:35:10,920 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:35:10,921 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:35:11,437 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:35:11,438 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:35:11,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:13,439 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:35:16,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:21,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:25,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:35:25,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:35:25,967 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:35:27,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:29,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:35:29,577 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:35:29,578 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:35:29,578 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:35:30,449 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:35:31,450 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:35:32,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:33,203 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:35:38,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:40,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:35:40,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:35:40,968 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:35:43,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:46,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:35:46,253 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:35:46,254 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:35:46,255 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:35:46,459 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:35:47,460 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:35:48,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:53,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:35:55,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:35:55,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:35:55,967 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:35:59,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:03,208 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:36:04,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:04,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:36:04,146 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:36:04,146 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:36:04,147 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:36:04,471 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:36:05,471 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:36:09,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:10,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:36:10,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:36:10,967 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:36:15,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:20,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:21,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:36:21,126 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:36:21,127 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:36:21,128 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:36:21,481 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:36:21,482 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:36:23,483 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:36:25,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:36:25,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:36:25,968 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:36:26,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:31,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:33,214 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:36:36,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:38,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:36:38,591 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:36:38,591 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:36:38,592 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:36:39,493 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:36:39,493 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:36:40,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:36:40,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:36:40,968 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:36:42,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:47,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:52,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:36:56,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:36:56,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:36:56,464 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:36:56,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:36:56,633 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:36:56,633 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:36:56,634 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:36:57,504 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:36:57,505 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:36:57,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:02,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:03,220 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:37:08,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:13,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:37:13,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:37:13,108 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:37:13,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:37:13,339 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:37:13,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:13,340 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:37:13,341 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:37:13,514 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:37:15,515 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:37:18,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:23,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:28,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:28,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:37:28,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:37:28,590 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:37:31,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:37:32,120 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:37:32,121 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:37:32,122 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:37:32,528 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:37:33,226 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:37:33,529 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:37:34,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:39,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:44,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:44,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:37:44,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:37:44,341 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:37:48,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:37:48,585 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:37:48,587 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:37:48,588 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:37:49,556 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:37:49,556 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:37:49,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:54,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:37:59,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:37:59,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:37:59,341 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:38:00,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:03,232 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:38:05,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:05,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:38:05,802 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:38:05,803 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:38:05,805 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:38:06,566 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:38:08,568 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:38:10,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:14,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:38:14,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:38:14,341 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:38:16,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:21,552 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:23,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:38:23,043 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:38:23,044 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:38:23,045 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:38:23,577 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:38:24,578 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:38:27,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:29,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:38:29,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:38:29,341 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:38:32,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:33,238 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:38:38,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:39,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:38:39,948 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:38:39,949 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:38:39,951 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:38:40,588 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:38:42,590 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:38:43,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:44,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:38:44,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:38:44,341 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:38:49,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:54,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:38:58,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:38:58,821 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:38:58,822 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:38:58,823 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:38:59,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:38:59,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:38:59,341 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:38:59,601 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:39:00,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:00,601 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:39:03,244 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:39:06,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:11,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:15,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:39:15,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:39:15,743 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:39:15,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:39:15,898 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:39:15,899 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:39:15,899 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:39:16,612 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:39:16,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:18,613 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:39:21,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:26,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:30,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:39:30,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:39:30,741 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:39:31,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:33,250 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:39:33,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:39:33,335 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:39:33,337 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:39:33,338 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:39:33,622 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:39:34,623 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:39:36,624 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:39:37,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:42,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:46,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:39:46,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:39:46,386 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:39:47,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:50,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:39:50,415 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:39:50,416 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:39:50,417 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:39:50,632 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:39:52,633 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:39:53,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:39:58,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:01,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:40:01,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:40:01,386 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:40:03,256 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:40:04,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:07,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:40:07,799 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:40:07,800 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:40:07,801 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:40:08,644 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:40:09,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:10,645 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:40:14,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:16,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:40:16,385 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:40:20,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:25,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:26,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:40:26,024 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:40:26,025 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:40:26,026 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:40:26,655 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:40:28,656 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:40:31,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:31,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:40:31,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:40:31,386 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:40:33,261 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:40:36,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:41,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:42,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:40:42,812 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:40:42,813 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:40:42,814 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:40:43,666 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:40:44,666 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:40:46,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:40:46,386 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:40:46,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:51,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:40:56,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:00,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:41:00,970 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:41:00,971 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:41:00,972 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:41:01,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:41:01,386 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:41:01,686 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:41:02,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:02,687 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:41:03,266 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:41:08,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:13,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:17,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:41:17,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:41:17,795 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:41:17,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:41:18,036 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:41:18,036 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:41:18,037 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:41:18,701 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:41:19,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:20,703 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:41:24,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:29,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:32,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:41:32,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:41:32,794 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:41:33,272 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:41:34,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:35,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:41:35,300 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:41:35,301 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:41:35,303 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:41:35,712 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:41:36,713 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:41:39,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:44,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:47,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:41:47,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:41:47,798 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:41:52,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:41:53,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:41:53,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:41:53,071 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:41:53,073 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:41:53,074 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:41:53,724 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:41:54,724 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:41:58,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:02,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:42:02,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:42:02,795 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:42:03,278 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:42:03,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:08,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:10,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:42:10,147 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:42:10,148 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:42:10,150 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:42:10,735 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:42:12,736 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:42:14,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:17,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:42:17,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:42:17,795 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:42:20,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:25,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:28,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:42:28,437 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:42:28,439 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:42:28,440 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:42:28,746 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:42:30,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:30,748 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:42:32,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:42:32,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:42:32,795 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:42:33,404 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:42:36,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:41,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:45,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:42:45,309 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:42:45,310 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:42:45,311 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:42:45,757 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:42:46,758 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:42:47,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:47,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:42:47,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:42:47,795 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:42:52,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:42:57,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:02,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:43:02,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:43:02,810 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:43:02,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:43:03,046 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:43:03,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:03,047 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:43:03,048 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:43:03,290 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:43:03,769 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:43:04,769 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:43:08,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:13,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:18,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:20,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:43:20,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:43:20,074 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:43:20,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:43:20,320 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:43:20,321 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:43:20,322 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:43:20,780 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:43:22,781 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:43:23,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:28,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:33,296 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:43:34,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:37,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:43:37,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:43:37,421 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:43:37,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:43:37,599 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:43:37,600 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:43:37,604 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:43:37,790 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:43:38,791 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:43:39,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:44,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:49,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:52,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:43:52,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:43:52,421 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:43:54,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:43:55,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:43:55,516 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:43:55,517 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:43:55,518 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:43:55,802 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:43:56,802 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:44:00,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:03,302 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:44:06,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:08,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:44:08,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:44:08,401 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:44:11,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:12,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:44:12,433 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:44:12,434 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:44:12,438 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:44:12,813 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:44:14,814 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:44:16,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:21,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:23,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:44:23,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:44:23,401 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:44:26,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:30,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:44:30,515 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:44:30,516 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:44:30,518 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:44:30,826 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:44:32,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:32,827 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:44:33,308 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:44:38,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:38,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:44:38,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:44:38,401 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:44:43,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:47,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:44:47,353 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:44:47,354 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:44:47,359 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:44:47,837 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:44:48,837 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:44:49,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:53,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:44:53,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:44:53,401 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:44:54,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:44:59,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:03,314 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:45:05,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:05,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:45:05,002 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:45:05,004 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:45:05,005 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:45:05,848 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:45:06,849 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:45:08,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:45:08,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:45:08,401 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:45:10,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:15,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:20,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:22,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:45:22,829 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:45:22,829 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:45:22,833 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:45:22,859 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:45:23,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:45:23,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:45:23,402 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:45:24,861 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:45:25,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:30,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:33,319 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:45:36,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:39,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:45:39,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:45:39,873 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:45:39,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:45:40,123 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:45:40,123 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:45:40,124 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:45:40,871 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:45:42,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:42,872 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:45:47,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:52,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:54,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:45:54,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:45:54,873 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:45:57,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:45:57,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:45:57,762 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:45:57,763 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:45:57,768 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:45:57,882 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:45:58,882 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:46:02,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:03,325 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:46:08,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:10,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:46:10,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:46:10,571 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:46:13,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:14,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:46:14,774 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:46:14,775 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:46:14,776 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:46:14,892 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:46:16,893 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:46:18,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:23,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:25,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:46:25,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:46:25,571 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:46:28,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:32,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:46:32,508 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:46:32,509 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:46:32,513 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:46:32,902 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:46:33,331 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:46:34,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:34,904 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:46:39,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:40,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:46:40,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:46:40,571 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:46:45,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:49,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:46:49,848 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:46:49,849 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:46:49,851 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:46:49,913 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:46:50,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:46:50,914 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:46:55,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:46:55,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:46:55,571 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:46:56,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:01,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:03,337 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:47:07,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:07,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:47:07,299 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:47:07,300 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:47:07,301 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:47:07,924 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:47:08,925 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:47:10,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:47:10,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:47:10,571 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:47:12,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:17,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:22,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:24,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:47:24,986 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:47:24,987 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:47:24,988 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:47:25,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:47:25,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:47:25,571 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:47:25,936 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:47:26,936 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:47:28,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:33,343 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:47:34,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:39,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:42,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:47:42,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:47:42,052 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:47:42,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:47:42,271 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:47:42,272 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:47:42,273 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:47:42,946 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:47:44,948 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:47:45,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:50,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:55,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:47:57,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:47:57,051 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:48:00,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:48:00,183 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:48:00,184 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:48:00,185 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:48:00,958 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:48:01,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:02,959 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:48:03,348 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:48:06,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:11,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:13,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:48:13,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:48:13,303 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:48:16,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:17,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:48:17,069 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:48:17,070 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:48:17,071 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:48:17,969 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:48:18,969 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:48:22,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:27,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:28,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:48:28,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:48:28,303 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:48:32,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:33,354 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:48:34,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:48:34,810 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:48:34,811 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:48:34,813 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:48:34,979 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:48:36,980 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:48:37,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:42,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:43,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:48:43,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:48:43,303 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:48:48,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:52,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:48:52,053 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:48:52,054 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:48:52,055 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:48:52,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:48:52,991 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:48:54,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:48:58,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:48:58,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:48:58,303 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:48:59,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:03,360 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:49:05,363 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:09,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:49:09,397 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:49:09,398 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:49:09,399 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:49:10,001 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:49:10,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:11,001 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:49:13,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:49:13,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:49:13,303 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:49:15,560 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:20,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:25,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:27,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:49:27,110 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:49:27,111 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:49:27,112 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:49:28,012 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:49:28,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:49:28,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:49:28,303 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:49:29,012 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:49:31,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:33,366 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:49:37,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:42,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:43,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:49:43,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:49:43,816 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:49:43,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:49:44,043 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:49:44,043 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:49:44,044 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:49:45,022 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:49:45,023 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:49:48,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:53,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:49:58,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:01,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:50:01,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:50:01,610 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:50:01,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:50:01,796 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:50:01,797 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:50:01,798 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:50:02,052 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:50:03,056 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:50:03,372 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:50:03,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:08,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:13,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:18,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:18,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:50:18,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:50:18,847 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:50:18,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:50:23,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:50:26,824 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:50:26,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:26,825 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:50:26,827 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:50:27,117 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:50:29,122 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:50:31,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:33,378 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:50:36,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:50:36,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:50:36,448 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:50:36,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:50:36,694 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:50:36,695 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:50:36,695 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:50:37,140 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:50:37,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:39,140 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:50:42,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:47,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:51,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:50:51,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:50:51,447 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:50:53,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:50:53,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:50:53,742 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:50:53,743 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:50:53,745 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:50:54,150 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:50:55,151 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:50:58,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:03,384 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:51:04,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:06,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:51:06,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:51:06,889 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:51:10,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:10,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:51:10,666 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:51:10,667 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:51:10,668 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:51:11,161 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:51:13,162 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:51:15,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:20,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:21,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:51:21,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:51:21,889 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:51:26,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:28,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:51:28,508 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:51:28,509 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:51:28,511 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:51:29,172 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:51:31,174 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:51:31,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:33,390 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:51:36,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:51:36,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:51:36,889 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:51:37,081 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:42,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:45,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:51:45,463 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:51:45,463 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:51:45,465 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:51:46,183 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:51:47,184 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:51:47,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:51,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:51:51,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:51:51,889 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:51:53,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:51:58,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:03,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:03,395 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:52:03,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:03,463 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 13:52:03,464 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 13:52:03,465 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 13:52:04,195 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 13:52:05,196 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 13:52:06,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:52:06,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:52:06,890 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:52:08,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:13,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:18,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:20,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:20,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:20,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:21,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:21,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:21,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:52:21,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:52:21,890 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:52:24,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:24,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:26,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:26,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:26,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:27,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:27,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:27,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:27,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:29,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:31,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:33,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:33,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:33,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:33,401 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:52:33,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:33,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:33,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:34,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:34,404 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:36,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:52:36,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:52:36,890 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:52:37,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:39,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:39,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:39,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:39,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:39,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:39,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:40,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:40,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:43,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:45,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:45,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:45,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:45,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:46,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:46,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:46,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:46,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:50,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:50,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:51,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:51,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:51,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:51,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:52:51,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:52:51,890 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:52:52,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:52,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:52,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:52,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:55,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:52:56,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:58,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:58,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:58,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:58,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:58,786 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:58,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:52:59,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:00,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:02,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:03,407 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:53:04,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:04,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:04,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:04,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:05,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:05,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:05,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:05,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:06,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:53:06,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:53:06,891 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:53:09,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:10,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:10,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:10,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:11,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:11,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:11,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:11,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:11,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:15,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:16,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:17,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:17,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:17,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:17,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:17,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:17,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:18,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:21,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:21,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:21,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:53:21,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:53:21,891 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:53:23,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:23,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:23,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:23,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:24,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:24,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:24,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:26,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:28,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:29,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:29,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:29,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:30,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:30,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:30,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:30,874 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:31,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:33,413 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:53:34,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:36,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:36,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:36,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:36,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:36,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:36,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:36,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:53:36,891 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:53:37,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:37,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:40,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:42,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:42,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:42,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:42,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:42,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:43,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:43,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:43,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:47,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:47,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:48,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:48,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:48,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:49,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:49,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:49,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:49,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:51,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:53:51,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:53:51,892 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:53:52,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:53,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:55,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:55,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:55,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:55,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:55,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:55,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:56,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:53:57,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:53:59,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:01,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:01,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:01,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:01,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:02,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:02,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:02,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:02,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:03,419 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:54:05,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:06,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:54:06,891 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:54:07,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:07,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:07,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:07,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:08,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:08,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:08,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:08,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:12,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:12,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:14,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:14,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:14,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:14,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:14,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:14,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:15,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:17,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:18,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:20,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:20,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:20,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:20,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:21,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:21,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:21,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:21,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:54:21,892 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:54:25,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:25,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:26,832 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:26,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:26,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:27,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:27,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:27,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:27,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:30,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:31,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:33,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:33,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:33,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:33,424 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:54:33,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:33,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:33,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:34,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:35,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:36,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:54:36,892 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:54:37,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:39,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:39,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:39,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:39,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:40,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:40,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:40,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:41,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:44,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:45,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:45,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:45,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:46,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:46,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:46,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:46,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:46,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:50,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:51,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:51,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:54:51,892 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:54:52,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:52,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:52,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:52,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:52,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:52,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:53,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:56,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:54:56,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:58,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:58,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:58,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:58,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:59,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:59,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:54:59,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:01,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:03,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:03,431 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:55:04,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:04,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:04,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:05,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:05,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:05,456 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:05,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:06,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:55:06,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:55:06,893 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:55:07,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:09,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:11,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:11,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:11,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:11,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:11,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:11,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:12,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:12,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:15,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:17,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:17,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:17,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:17,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:17,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:18,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:18,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:18,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:21,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:55:21,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:55:21,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:55:22,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:22,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:23,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:23,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:23,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:24,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:24,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:24,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:24,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:27,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:28,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:30,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:30,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:30,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:30,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:30,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:30,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:31,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:32,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:33,436 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:55:34,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:36,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:36,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:36,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:36,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:55:36,893 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:55:36,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:37,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:37,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:37,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:38,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:41,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:42,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:42,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:42,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:42,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:43,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:48,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:51,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:55:51,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:55:51,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:55:54,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:55:58,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:55:59,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:03,442 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:56:04,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:06,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:56:06,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:56:06,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:56:10,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:14,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:14,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:15,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:16,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:19,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:19,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:19,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:20,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:20,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:20,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:20,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:20,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:21,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:56:21,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:56:22,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:25,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:25,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:25,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:26,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:26,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:26,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:26,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:26,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:28,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:30,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:32,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:32,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:32,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:32,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:32,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:33,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:33,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:33,448 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:56:35,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:36,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:36,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:56:36,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:56:38,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:38,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:38,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:38,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:39,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:39,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:39,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:41,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:42,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:44,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:44,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:45,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:45,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:45,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:45,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:45,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:47,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:47,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:51,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:51,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:51,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:51,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:51,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:51,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:56:51,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:56:52,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:52,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:52,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:53,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:57,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:56:57,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:57,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:57,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:58,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:58,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:58,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:56:58,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:00,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:02,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:03,453 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:57:03,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:03,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:04,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:04,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:04,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:04,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:04,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:06,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:06,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:57:06,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:57:08,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:10,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:10,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:10,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:10,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:10,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:11,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:11,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:13,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:13,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:16,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:16,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:16,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:16,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:17,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:17,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:17,552 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:18,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:19,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:21,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:57:21,895 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:57:22,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:22,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:23,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:23,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:23,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:23,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:23,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:23,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:25,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:28,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:29,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:29,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:29,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:29,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:29,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:30,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:30,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:32,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:33,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:33,460 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:57:35,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:35,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:35,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:36,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:36,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:36,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:36,620 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:36,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:57:36,895 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:57:38,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:39,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:41,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:41,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:42,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:42,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:42,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:42,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:42,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:44,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:44,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:48,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:48,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:48,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:48,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:48,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:49,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:49,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:49,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:51,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:51,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:57:51,895 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:57:54,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:57:54,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:54,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:54,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:55,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:55,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:55,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:55,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:57,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:57:59,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:00,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:00,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:01,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:01,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:01,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:02,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:02,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:03,466 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:58:03,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:04,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:06,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:58:06,896 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:58:07,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:07,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:07,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:07,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:07,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:08,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:08,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:10,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:10,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:13,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:13,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:14,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:14,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:14,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:14,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:14,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:15,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:16,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:19,874 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:19,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:20,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:20,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:20,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:20,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:21,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:21,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:21,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:58:21,896 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:58:22,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:26,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:26,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:26,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:26,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:26,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:26,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:27,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:27,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:29,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:31,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:32,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:32,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:33,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:33,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:33,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:33,471 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:58:33,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:33,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:35,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:36,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:36,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:58:36,896 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:58:38,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:38,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:39,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:39,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:39,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:40,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:40,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:41,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:42,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:45,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:45,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:45,626 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:45,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:45,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:46,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:46,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:47,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:48,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:51,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:51,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:51,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:58:51,896 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:58:52,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:52,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:52,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:52,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:52,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:53,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:54,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:57,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:57,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:58,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:58,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:58,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:58,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:58:58,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:58:58,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:00,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:03,478 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:59:04,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:04,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:04,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:04,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:04,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:04,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:05,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:05,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:06,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:59:06,897 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:59:07,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:10,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:10,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:10,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:10,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:11,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:11,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:11,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:11,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:13,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:15,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:16,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:16,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:17,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:17,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:17,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:18,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:18,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:19,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:20,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:21,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:59:21,897 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:59:23,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:23,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:23,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:23,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:23,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:24,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:24,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:26,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:26,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:29,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:29,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:30,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:30,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:30,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:30,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:30,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:31,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:32,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:33,483 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 13:59:35,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:35,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:36,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:36,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:36,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:36,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:36,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:59:36,897 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:59:36,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:36,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:42,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:47,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:51,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 13:59:51,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 13:59:51,898 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 13:59:51,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:53,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:58,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 13:59:59,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 13:59:59,530 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,530 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,531 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,531 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,532 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,532 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,532 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,533 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,533 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,534 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,534 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,535 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,535 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,536 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,536 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,536 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,537 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,537 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,537 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,538 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,538 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,539 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,539 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,540 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,540 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,540 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,541 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,541 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,542 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,542 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,542 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,543 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,543 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,544 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,544 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,544 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,545 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,545 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,546 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,546 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,546 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,547 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,547 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,548 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,548 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,548 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,549 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,549 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,549 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,550 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,550 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,551 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,551 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,552 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,552 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,552 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,553 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,553 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,554 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,554 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,554 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,555 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,555 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,556 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,556 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,556 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,557 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,557 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,558 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,558 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,558 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,559 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,559 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,560 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,560 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,560 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,561 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,561 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,561 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,562 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,562 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,563 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,563 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,564 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,564 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,564 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,565 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,565 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,566 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,566 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,566 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,567 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,567 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,568 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,568 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,568 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,569 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,569 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,570 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,570 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,570 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,571 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,571 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,572 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,572 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,572 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,573 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,573 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,574 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,574 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,574 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,575 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,575 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,576 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,576 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,577 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,577 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,577 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,578 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,578 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,579 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,579 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,579 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,580 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,580 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,581 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,581 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,582 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,582 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,582 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,583 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,583 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,584 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,584 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,584 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,585 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,585 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,586 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,586 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,586 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,587 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,587 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,588 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,588 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,588 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,589 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,589 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,590 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,590 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,590 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,591 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,591 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,592 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,592 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,592 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,593 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,593 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,594 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,594 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,594 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,595 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,595 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,596 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,596 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,596 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,597 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,597 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,598 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,598 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,598 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,599 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,599 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,600 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,600 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,600 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,601 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,601 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,602 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,602 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,602 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,603 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,603 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,604 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,604 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,604 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,605 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,605 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,606 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,606 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,606 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,607 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,607 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,608 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,608 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,608 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,609 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,609 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,609 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,610 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,610 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,611 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,611 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,612 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,612 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,612 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,613 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,613 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,614 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,614 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,614 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,615 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,615 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,615 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,616 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,616 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,617 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,617 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,617 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,618 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,618 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,619 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,619 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,619 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,620 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,620 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,621 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,621 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,621 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,622 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,622 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,623 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,623 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,623 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,624 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,624 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,625 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,625 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,625 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,626 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,626 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,627 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,627 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,627 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,628 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,628 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,629 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,629 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,629 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,630 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,630 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,631 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,631 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,631 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,632 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,632 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,633 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,633 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,633 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,634 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,634 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,635 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,635 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,635 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,636 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,636 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,637 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,637 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,638 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,638 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,639 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,639 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,639 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,640 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,640 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,641 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,641 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,641 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,642 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,642 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,643 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,643 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,643 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,644 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,644 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,645 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,645 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,645 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,646 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,646 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,647 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,647 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,647 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,648 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,648 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,648 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,649 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,649 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,650 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,650 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,651 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,651 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,651 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,652 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,652 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,653 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,653 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,653 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,654 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,654 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,655 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,655 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,655 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,656 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,656 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,657 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,657 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,657 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,658 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,658 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,659 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,659 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,659 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,660 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,660 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,661 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,661 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,661 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,662 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,662 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,663 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,663 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,663 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,664 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,664 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,665 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,665 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,665 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,666 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,667 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,667 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,668 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,668 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,669 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,669 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,669 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,670 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,670 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,671 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,671 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,671 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,672 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,672 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,673 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,673 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,673 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,674 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,674 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,675 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,675 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,675 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,676 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,676 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,677 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,677 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,677 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,678 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,678 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,679 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,679 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,679 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,680 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,680 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,681 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,681 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,683 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,683 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,684 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,684 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,685 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,685 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,686 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,686 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,686 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,687 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,687 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,687 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,688 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,688 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,689 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,689 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,692 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,693 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,693 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,694 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,694 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,694 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,695 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,695 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,696 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,696 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,696 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,697 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,697 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,698 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,698 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,698 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,699 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,699 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,699 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,700 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,700 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,701 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,701 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,701 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,702 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,702 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,703 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,703 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,703 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,704 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,704 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,705 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,705 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,705 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,706 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,706 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,707 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,707 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,707 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,708 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,708 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,709 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,709 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,709 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,710 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,710 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,711 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,711 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,711 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,712 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,712 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,713 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,713 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,713 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,714 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,714 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,715 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,715 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,715 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,716 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,716 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,717 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,717 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,717 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,718 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,718 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,719 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,719 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,719 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,720 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,720 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,721 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,721 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,721 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,722 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,722 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,723 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,725 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,726 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,726 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,727 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,727 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,728 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,728 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,728 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,729 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,729 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,729 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,730 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,730 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,731 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,731 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,731 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,732 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,732 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,733 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,733 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,733 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,734 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,734 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,735 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,735 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,735 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,736 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,736 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,737 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,737 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,737 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,738 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,738 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,739 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,739 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,739 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,740 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,740 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,741 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,741 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,741 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,742 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,742 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,742 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,743 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,743 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,744 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,744 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,744 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,745 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,745 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,746 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,746 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,747 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,747 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,747 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,748 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,748 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,748 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,749 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,749 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,750 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,750 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,750 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,751 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,751 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,752 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,752 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,752 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,753 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,753 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,754 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,754 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,754 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,755 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,755 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,756 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,756 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,756 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,757 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,757 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,758 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,758 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,758 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,759 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,759 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,760 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,760 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,760 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,761 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,761 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,762 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,762 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,762 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,763 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,763 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,764 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,764 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,764 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,765 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,765 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,765 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,766 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,766 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,767 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,767 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,768 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,768 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,768 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,769 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,769 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,769 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,770 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,770 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,771 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,771 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,771 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,772 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,772 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,773 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,773 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,773 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,774 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,774 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,775 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,775 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,775 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,776 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,776 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,777 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,777 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,777 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,778 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,778 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,778 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,779 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,779 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,780 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,780 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,781 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,781 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,781 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,782 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,782 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,782 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,783 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,783 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,784 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,784 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,784 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,785 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,785 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,786 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,786 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,786 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,787 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,787 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,788 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,788 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,788 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,789 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,789 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,790 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,790 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,790 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,791 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,791 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,792 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,792 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,792 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,793 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,793 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,794 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,794 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,794 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,795 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,795 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,796 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,796 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,796 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,797 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,797 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,798 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,798 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,798 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,799 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,799 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,800 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,800 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,800 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,801 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,801 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,801 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,802 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,802 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,803 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,803 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,804 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,804 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,804 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,805 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,805 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,806 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,806 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,806 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,807 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,807 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,808 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,808 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,808 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,809 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,809 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,810 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,810 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,810 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,811 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,811 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,812 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,812 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,812 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,813 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,813 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,814 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,814 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,814 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,815 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,815 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,816 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,816 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,816 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,817 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,817 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,818 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,818 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,818 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,819 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,819 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,820 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,820 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,820 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,821 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,821 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,822 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,822 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,822 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,823 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,823 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,824 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,824 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,824 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,825 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,825 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,825 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,826 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,826 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,827 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,827 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,827 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,828 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,828 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,829 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,829 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,829 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,830 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,830 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,831 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,831 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,831 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,832 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,832 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,833 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,833 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,833 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,834 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,834 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,835 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,835 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,835 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,836 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,836 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,837 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,837 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,837 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,838 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,838 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,839 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,839 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,839 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,840 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,840 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,841 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,841 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,841 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,842 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,842 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,842 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,843 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,843 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,844 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,844 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,845 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,845 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,845 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,846 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,846 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,847 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,848 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,848 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,848 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,849 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,849 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,849 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,850 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,850 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,851 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,851 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,852 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,852 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,852 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,853 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,853 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,853 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,854 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,854 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,855 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,855 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,855 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,856 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,856 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,857 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,857 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,858 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,858 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,858 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,859 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,859 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,859 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,860 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,860 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,861 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,861 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,862 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,862 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,862 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,863 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,863 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,864 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,864 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,864 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,865 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,865 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,866 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,866 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,866 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,867 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,867 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,868 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,868 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,868 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,869 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,869 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,870 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,870 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,870 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,871 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,871 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,872 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,872 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,872 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,873 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,873 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,873 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,874 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,874 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,875 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,875 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,876 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,876 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,876 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,877 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,877 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,877 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,878 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,878 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,879 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,879 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,880 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,880 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,880 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,881 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,881 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,882 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,882 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,882 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,883 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,883 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,883 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,884 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,884 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,885 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,885 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,885 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,886 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,886 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,887 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,887 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,887 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,888 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,888 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,889 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,889 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,889 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,890 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,890 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,891 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,891 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,891 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,892 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,892 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,893 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,893 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,893 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,894 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,894 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,894 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,895 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,895 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,896 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,896 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,897 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,897 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,897 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,898 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,898 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,899 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,899 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,899 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,900 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,900 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,901 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,901 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,901 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,902 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,902 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,903 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,903 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,903 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,904 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,904 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,904 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,905 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,905 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,906 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,906 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,906 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,907 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,907 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,908 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,908 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,908 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,909 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,909 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,910 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,910 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,910 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,911 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,911 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,912 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,912 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,912 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,913 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,913 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,914 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,914 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,914 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,915 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,915 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,916 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,916 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,916 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,917 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,917 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,918 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,918 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,918 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,919 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,919 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,920 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,920 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,920 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,921 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,921 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,922 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,922 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,922 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,923 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,923 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,924 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,924 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,924 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,925 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,925 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,926 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,926 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,926 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,927 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,927 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,928 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,928 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,928 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,929 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,929 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,930 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,930 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,930 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,931 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,931 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,932 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,932 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,933 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,933 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,933 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,934 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,934 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,935 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,935 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,935 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,936 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,936 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,937 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,937 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,937 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,938 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,938 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,939 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,939 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,939 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,940 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,940 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,941 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,941 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,941 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,942 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,942 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,943 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,943 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,943 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,944 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,944 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,945 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,945 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,945 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,946 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,946 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,947 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,947 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,947 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,948 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,948 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,949 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,949 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,949 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,950 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,950 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,951 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,951 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,951 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,952 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,952 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,953 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,953 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,957 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,957 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,958 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,958 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,959 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,959 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,960 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,960 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,960 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,961 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,961 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,962 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,962 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,963 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,963 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,963 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,964 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,964 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,965 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,965 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,965 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,966 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,966 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,967 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,967 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,967 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,968 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,968 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,969 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,969 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,969 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,970 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,970 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,971 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,971 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,971 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,972 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,972 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,973 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,973 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,973 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,974 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,974 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,975 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,975 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,975 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,976 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,976 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,977 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,977 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,977 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,978 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,978 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,979 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,979 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,979 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,980 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,980 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,981 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,981 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,982 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,982 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,982 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,983 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,983 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,983 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,984 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,984 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,985 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,985 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,986 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,986 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,986 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,987 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,987 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,988 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,988 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,988 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,989 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,989 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,990 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,991 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,991 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,992 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,992 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,993 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,993 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,994 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,994 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,995 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,995 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,996 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,996 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,997 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,997 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,997 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,998 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,998 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,999 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,999 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 13:59:59,999 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,000 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,000 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,001 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,001 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,001 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,002 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,002 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,003 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,003 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,003 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,004 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,004 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,005 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,005 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,005 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,006 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,006 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,007 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,007 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,008 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,008 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,008 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,009 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,009 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,010 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,010 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,010 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,011 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,011 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,012 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,012 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,012 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,013 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,013 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,014 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,014 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,014 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,015 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,015 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,016 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,016 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,016 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,017 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,017 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,018 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,018 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,018 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,019 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,019 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,030 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,045 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,051 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,066 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,067 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,067 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,067 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,067 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,067 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,067 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,068 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,068 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,068 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,068 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,068 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,069 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,069 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,069 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,069 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,069 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,069 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,070 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,070 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,070 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,070 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,070 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,071 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,071 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,071 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,071 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,071 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,071 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,072 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,072 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,072 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,072 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,072 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,073 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,073 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,073 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,073 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,073 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,073 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,074 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,074 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,089 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,109 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,117 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,118 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,118 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,118 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,118 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,118 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,119 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,119 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,119 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,119 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,119 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,120 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,120 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,120 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,120 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,120 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,120 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,121 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,121 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,121 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,121 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,121 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,122 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,122 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,122 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,122 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,122 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,122 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,123 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,123 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,123 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,123 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,123 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,123 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,124 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,124 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,124 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,124 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,124 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,125 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,125 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,125 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,125 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,125 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,131 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,141 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,147 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,147 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,147 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,147 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,148 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,148 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,148 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,148 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,148 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,148 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,149 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,149 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,149 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,149 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,155 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,165 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,176 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,176 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,176 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,176 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,176 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,176 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,177 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,177 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,177 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,177 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,177 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,178 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,178 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,178 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,178 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,178 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,178 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,179 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,179 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,179 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,179 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,179 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,179 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,180 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,180 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,180 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,180 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,180 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,180 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,181 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,186 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,191 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,202 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,202 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,202 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,203 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,203 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,203 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,203 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,203 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,204 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,204 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,204 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,204 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,204 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,204 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,205 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,205 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,205 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,205 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,216 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,226 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,226 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,227 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,227 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,227 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,227 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,227 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,227 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,228 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,228 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,228 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,228 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,228 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,229 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,239 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,239 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,240 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,240 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,240 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,240 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,240 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,241 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,241 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,241 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,241 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,241 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,247 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,252 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,265 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,265 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,265 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,265 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,265 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,265 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,266 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,266 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,266 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,266 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,266 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,267 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,267 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,267 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,267 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,267 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,267 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,268 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,268 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,268 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,268 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,268 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,269 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,269 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,269 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,269 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,269 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,269 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,270 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,270 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,270 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,270 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,270 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,271 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,271 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,271 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,271 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,271 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,271 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,272 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,272 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,272 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,272 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,272 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,272 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,273 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,273 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,273 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,273 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,273 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,274 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,274 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,274 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,274 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,274 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,274 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,275 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,275 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,275 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,275 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,275 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,275 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,276 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,276 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,276 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,276 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,276 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,277 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,277 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,277 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,277 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,277 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,278 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,278 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,278 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,278 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,278 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,279 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,279 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,279 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,279 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,279 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,279 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,280 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,280 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,280 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,280 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,280 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,280 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,281 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,281 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,281 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,281 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,281 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,282 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,282 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,282 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,282 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,282 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,282 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,283 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,283 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,283 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,283 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,283 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,284 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,284 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,284 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,284 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,284 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,284 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,285 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,285 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,285 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,285 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,285 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,286 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,286 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,286 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,286 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,286 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,286 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,287 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,287 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,287 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,287 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,287 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,287 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,288 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,288 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,288 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,288 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,288 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,289 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,289 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,289 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,289 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,289 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,289 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,290 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,290 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,290 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,290 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,290 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,291 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,291 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,291 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,291 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,291 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,291 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,292 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,292 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,292 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,292 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,292 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,292 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,293 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,293 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,293 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,293 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,293 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,294 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,294 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,294 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,294 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,294 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,294 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,295 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,295 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,295 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,295 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,295 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,296 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,296 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,296 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,296 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,296 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,296 DEBUG SenderThread:21069 [sender.py:send():378] send: metric
+2024-05-15 14:00:00,302 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:00:00,388 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:00:00,434 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:00:00,535 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:00:00,535 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:00:02,536 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:00:03,436 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:03,489 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:00:06,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:00:06,897 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:00:09,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:14,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:16,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:00:16,225 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:00:16,226 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:00:16,307 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:00:16,547 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:00:18,548 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:00:19,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:21,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:00:21,897 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:00:25,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:26,553 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/config.yaml
+2024-05-15 14:00:31,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:33,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:00:33,479 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:00:33,480 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:00:33,578 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:00:33,579 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:00:33,581 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:00:34,581 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:00:36,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:36,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:00:36,897 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:00:42,094 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:47,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:51,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:00:51,451 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:00:51,452 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:00:51,532 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:00:51,592 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:00:51,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:00:51,897 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:00:52,592 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:00:53,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:00:58,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:03,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:03,581 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:01:08,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:08,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:01:08,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:01:08,627 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:01:08,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:01:08,864 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:01:08,865 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:01:08,944 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:01:09,603 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:01:10,603 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:01:13,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:18,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:23,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:01:23,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:01:23,626 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:01:24,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:26,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:01:26,880 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:01:26,881 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:01:26,961 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:01:27,614 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:01:28,614 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:01:29,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:33,587 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:01:35,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:39,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:01:39,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:01:39,843 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:01:41,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:43,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:01:43,402 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:01:43,403 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:01:43,483 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:01:43,624 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:01:44,624 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:01:46,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:51,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:01:54,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:01:54,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:01:54,843 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:01:57,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:00,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:02:00,565 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:02:00,566 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:02:00,646 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:02:00,647 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:02:01,647 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:02:02,647 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:02:02,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:03,593 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:02:08,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:09,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:02:09,842 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:02:14,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:17,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:02:17,739 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:02:17,741 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:02:17,824 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:02:18,657 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:02:18,658 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:02:19,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:20,658 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:02:24,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:24,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:02:24,842 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:02:29,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:33,599 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:02:34,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:34,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:02:35,000 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:02:35,001 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:02:35,082 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:02:35,668 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:02:36,669 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:02:39,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:02:39,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:02:39,843 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:02:40,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:45,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:50,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:53,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:02:53,836 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:02:53,837 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:02:53,918 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:02:54,680 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:02:54,680 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:02:54,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:02:54,842 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:02:55,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:02:56,681 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:03:00,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:03,603 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:03:05,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:10,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:03:10,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:03:10,283 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:03:10,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:03:10,503 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:03:10,503 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:03:10,583 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:03:10,690 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:03:11,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:12,692 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:03:16,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:21,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:26,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:27,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:03:27,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:03:27,960 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:03:27,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:03:28,193 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:03:28,193 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:03:28,272 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:03:28,702 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:03:30,703 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:03:32,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:33,609 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:03:37,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:42,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:44,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:03:44,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:03:44,756 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:03:44,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:03:44,972 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:03:44,973 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:03:45,053 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:03:45,712 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:03:46,713 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:03:48,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:53,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:03:58,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:02,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:04:02,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:04:02,131 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:04:02,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:04:02,302 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:04:02,302 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:04:02,381 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:04:02,723 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:04:03,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:03,614 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:04:04,725 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:04:08,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:13,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:17,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:04:17,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:04:17,131 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:04:19,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:20,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:04:20,430 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:04:20,431 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:04:20,511 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:04:20,735 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:04:22,736 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:04:24,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:29,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:33,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:04:33,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:04:33,602 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:04:33,845 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:04:34,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:37,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:04:37,131 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:04:37,132 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:04:37,217 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:04:37,745 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:04:38,746 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:04:40,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:45,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:48,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:04:48,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:04:48,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:04:50,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:55,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:04:55,517 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:04:55,518 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:04:55,598 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:04:55,756 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:04:56,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:04:56,757 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:05:01,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:03,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:05:03,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:05:03,602 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:05:03,835 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:05:06,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:11,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:12,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:05:12,203 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:05:12,204 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:05:12,299 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:05:12,767 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:05:14,768 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:05:17,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:18,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:05:18,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:05:18,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:05:22,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:27,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:29,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:05:29,338 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:05:29,339 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:05:29,419 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:05:29,778 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:05:30,778 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:05:33,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:33,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:05:33,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:05:33,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:05:33,823 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:05:38,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:43,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:46,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:05:46,605 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:05:46,606 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:05:46,698 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:05:46,788 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:05:48,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:05:48,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:05:48,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:05:48,789 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:05:49,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:54,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:05:59,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:03,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:06:03,324 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:06:03,325 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:06:03,405 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:06:03,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:06:03,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:06:03,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:06:03,799 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:06:03,812 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:06:04,800 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:06:04,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:09,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:14,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:18,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:06:18,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:06:18,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:06:20,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:22,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:06:22,379 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:06:22,380 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:06:22,473 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:06:22,811 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:06:24,812 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:06:26,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:31,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:33,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:06:33,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:06:33,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:06:33,799 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:06:36,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:39,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:06:39,301 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:06:39,302 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:06:39,384 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:06:39,821 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:06:40,821 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:06:42,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:47,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:48,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:06:48,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:06:48,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:06:52,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:56,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:06:56,778 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:06:56,779 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:06:56,872 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:06:56,872 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:06:57,874 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:06:58,872 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:07:02,874 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:03,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:07:03,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:07:03,604 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:07:03,852 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:07:08,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:13,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:07:13,523 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:07:13,524 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:07:13,607 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:07:13,882 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:07:14,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:14,883 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:07:18,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:07:18,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:07:18,604 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:07:19,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:24,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:29,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:30,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:07:30,146 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:07:30,147 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:07:30,240 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:07:30,893 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:07:32,894 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:07:33,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:07:33,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:07:33,604 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:07:33,832 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:07:35,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:40,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:45,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:49,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:07:49,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:07:49,407 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:07:49,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:07:49,654 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:07:49,654 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:07:49,742 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:07:49,905 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:07:50,906 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:07:51,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:07:56,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:01,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:03,662 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:08:06,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:08:06,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:08:06,227 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:08:06,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:08:07,444 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:08:07,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:07,445 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:08:07,523 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:08:07,916 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:08:08,917 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:08:12,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:17,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:21,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:08:21,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:08:21,227 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:08:23,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:24,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:08:24,616 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:08:24,617 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:08:24,698 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:08:24,927 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:08:26,928 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:08:28,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:33,667 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:08:34,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:36,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:08:36,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:08:36,227 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:08:40,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:41,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:08:41,354 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:08:41,355 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:08:41,435 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:08:41,938 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:08:42,939 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:08:46,436 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:51,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:08:51,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:08:51,227 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:08:51,456 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:56,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:08:57,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:08:57,930 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:08:57,931 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:08:58,012 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:08:59,012 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:08:59,012 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:09:02,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:03,673 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:09:06,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:09:06,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:09:06,228 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:09:07,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:12,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:15,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:09:15,979 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:09:15,979 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:09:16,059 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:09:17,059 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:09:17,059 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:09:18,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:21,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:09:21,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:09:21,228 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:09:23,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:28,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:32,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:09:32,638 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:09:32,639 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:09:32,722 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:09:33,069 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:09:33,679 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:09:33,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:35,070 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:09:36,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:09:36,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:09:36,228 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:09:39,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:44,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:49,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:51,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:09:51,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:09:51,601 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:09:51,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:09:51,836 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:09:51,837 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:09:51,916 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:09:52,081 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:09:53,082 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:09:54,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:09:59,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:03,685 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:10:05,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:08,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:10:08,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:10:08,617 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:10:08,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:10:08,835 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:10:08,835 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:10:08,915 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:10:09,092 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:10:10,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:11,093 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:10:15,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:20,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:25,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:10:25,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:10:25,673 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:10:25,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:10:25,911 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:10:25,911 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:10:25,992 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:10:25,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:26,125 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:10:27,126 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:10:30,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:33,691 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:10:36,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:41,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:43,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:10:43,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:10:43,274 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:10:43,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:10:46,124 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:10:46,124 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:10:46,206 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:10:46,207 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:10:47,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:47,208 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:10:52,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:10:57,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:00,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:11:00,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:11:00,045 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:11:00,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:11:00,255 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:11:00,256 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:11:00,337 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:11:01,242 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:11:01,243 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:11:02,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:03,247 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:11:03,696 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:11:07,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:12,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:15,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:11:15,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:11:15,045 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:11:18,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:11:19,625 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:11:19,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:19,626 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:11:19,706 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:11:20,292 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:11:21,293 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:11:24,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:29,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:30,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:11:30,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:11:30,045 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:11:33,702 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:11:35,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:35,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:11:35,590 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:11:35,591 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:11:35,670 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:11:36,332 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:11:37,333 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:11:40,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:45,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:11:45,045 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:11:46,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:51,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:11:53,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:11:53,649 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:11:53,649 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:11:53,688 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:11:54,348 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:11:55,349 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:11:56,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:00,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:12:00,045 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:12:02,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:03,708 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:12:07,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:11,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:12:11,041 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:12:11,042 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:12:11,079 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:12:11,359 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:12:13,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:13,360 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:12:15,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:12:15,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:12:15,046 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:12:18,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:23,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:27,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:12:27,642 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:12:27,643 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:12:27,680 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:12:28,370 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:12:28,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:29,371 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:12:30,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:12:30,046 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:12:33,714 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:12:33,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:38,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:43,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:45,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:12:45,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:12:45,346 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:12:45,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:12:45,580 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:12:45,581 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:12:45,616 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:12:46,381 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:12:47,382 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:12:49,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:54,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:12:59,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:02,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:13:02,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:13:02,177 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:13:02,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:13:03,673 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:13:03,673 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:13:03,767 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:13:03,768 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:13:04,393 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:13:04,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:05,393 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:13:09,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:14,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:17,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:13:17,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:13:17,177 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:13:20,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:20,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:13:20,601 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:13:20,602 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:13:20,639 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:13:21,403 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:13:23,405 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:13:25,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:30,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:32,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:13:32,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:13:32,178 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:13:36,750 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:13:36,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:38,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:13:38,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:13:38,125 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:13:38,127 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:13:38,207 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:13:38,415 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:13:39,416 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:13:42,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:47,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:13:47,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:13:47,178 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:13:47,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:52,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:13:55,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:13:55,281 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:13:55,283 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:13:55,363 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:13:55,426 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:13:57,426 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:13:58,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:02,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:14:02,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:14:02,178 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:14:03,417 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:03,767 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:14:08,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:12,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:14:12,767 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:14:12,768 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:14:12,866 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:14:13,437 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:14:13,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:15,438 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:14:17,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:14:17,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:14:17,179 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:14:19,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:24,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:29,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:29,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:14:29,482 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:14:29,483 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:14:29,563 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:14:30,447 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:14:31,448 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:14:32,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:14:32,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:14:32,178 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:14:33,773 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:14:34,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:39,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:44,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:47,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:14:47,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:14:47,460 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:14:47,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:14:47,683 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:14:47,683 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:14:47,776 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:14:48,459 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:14:49,460 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:14:50,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:14:55,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:00,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:03,779 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:15:05,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:15:05,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:15:05,061 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:15:05,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:15:05,309 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:15:05,310 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:15:05,388 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:15:05,470 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:15:06,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:07,471 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:15:11,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:16,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:20,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:15:20,060 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:15:22,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:22,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:15:22,921 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:15:22,922 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:15:23,014 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:15:23,481 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:15:25,482 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:15:28,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:33,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:33,785 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:15:35,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:15:35,060 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:15:38,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:40,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:15:40,448 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:15:40,449 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:15:40,530 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:15:40,531 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:15:41,530 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:15:43,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:48,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:50,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:15:50,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:15:50,061 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:15:54,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:15:57,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:15:57,076 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:15:57,078 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:15:57,170 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:15:57,540 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:15:59,541 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:16:00,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:16:03,790 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:16:05,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:16:05,060 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:16:10,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:16:10,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:16:14,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:16:14,092 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:16:14,094 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:16:14,174 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:16:14,551 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:16:15,551 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:16:16,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:16:20,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:16:20,060 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:16:21,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:16:26,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:16:31,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:16:31,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:16:31,725 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:16:31,726 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:16:31,819 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:16:32,562 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:16:33,563 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:16:33,795 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:16:35,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:16:35,060 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:16:40,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:16:43,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:16:48,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:16:50,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:16:50,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:16:50,135 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:16:50,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:16:50,376 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:16:50,377 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:16:50,456 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:16:50,573 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:16:51,573 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:16:53,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:16:58,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:03,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:03,802 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:17:07,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:17:07,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:17:07,386 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:17:07,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:17:07,581 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:17:07,582 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:17:07,661 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:17:07,661 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:17:08,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:09,662 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:17:13,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:18,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:23,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:24,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:17:24,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:17:24,938 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:17:24,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:17:25,105 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:17:25,106 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:17:25,189 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:17:25,672 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:17:27,673 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:17:29,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:33,807 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:17:34,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:39,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:42,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:17:42,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:17:42,073 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:17:42,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:17:42,323 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:17:42,324 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:17:42,404 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:17:42,683 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:17:43,683 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:17:45,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:50,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:55,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:17:59,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:17:59,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:17:59,579 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:17:59,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:17:59,792 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:17:59,793 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:17:59,872 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:18:00,694 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:18:00,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:01,695 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:18:03,813 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:18:06,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:11,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:16,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:17,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:18:17,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:18:17,138 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:18:17,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:18:17,396 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:18:17,397 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:18:17,481 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:18:17,705 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:18:19,706 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:18:22,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:27,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:32,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:33,819 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:18:34,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:18:34,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:18:34,527 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:18:34,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:18:34,709 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:18:34,710 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:18:34,790 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:18:35,790 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:18:35,791 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:18:37,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:42,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:47,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:49,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:18:49,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:18:49,528 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:18:52,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:18:52,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:18:52,864 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:18:52,865 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:18:52,946 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:18:53,802 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:18:53,802 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:18:57,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:02,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:03,825 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:19:04,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:19:04,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:19:04,528 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:19:08,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:09,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:19:09,916 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:19:09,918 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:19:09,998 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:19:10,812 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:19:11,813 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:19:13,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:19,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:19,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:19:19,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:19:19,528 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:19:24,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:27,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:19:27,180 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:19:27,181 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:19:27,262 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:19:27,823 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:19:29,824 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:19:30,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:33,831 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:19:34,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:19:34,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:19:34,528 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:19:35,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:40,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:44,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:19:44,015 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:19:44,016 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:19:44,097 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:19:44,851 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:19:45,852 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:19:46,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:49,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:19:49,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:19:49,528 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:19:51,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:19:56,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:01,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:20:01,271 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:20:01,272 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:20:01,352 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:20:01,862 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:20:02,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:03,837 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:20:03,863 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:20:04,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:20:04,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:20:04,528 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:20:07,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:12,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:17,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:19,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:20:19,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:20:19,731 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:20:19,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:20:19,926 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:20:19,927 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:20:20,007 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:20:20,874 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:20:21,874 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:20:23,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:28,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:33,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:33,842 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:20:36,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:20:36,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:20:36,870 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:20:36,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:20:37,066 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:20:37,066 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:20:37,146 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:20:37,884 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:20:37,884 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:20:38,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:43,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:48,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:53,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:20:54,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:20:54,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:20:54,993 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:20:54,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:20:55,231 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:20:55,232 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:20:55,311 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:20:55,895 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:20:57,896 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:20:58,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:03,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:03,848 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:21:08,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:12,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:21:12,167 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:21:12,167 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:21:12,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:21:12,378 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:21:12,379 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:21:12,458 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:21:12,905 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:21:13,906 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:21:14,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:19,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:24,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:29,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:21:29,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:21:29,058 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:21:29,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:21:29,316 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:21:29,317 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:21:29,397 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:21:29,916 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:21:30,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:31,917 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:21:33,854 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:21:35,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:40,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:45,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:46,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:21:46,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:21:46,724 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:21:46,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:21:46,911 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:21:46,912 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:21:46,991 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:21:46,992 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:21:47,991 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:21:50,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:21:55,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:00,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:03,860 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:22:04,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:22:04,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:22:04,017 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:22:04,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:22:04,238 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:22:04,239 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:22:04,320 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:22:05,002 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:22:06,003 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:22:06,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:11,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:16,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:19,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:22:19,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:22:19,018 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:22:22,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:22,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:22:22,633 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:22:22,635 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:22:22,715 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:22:23,013 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:22:24,014 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:22:27,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:32,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:33,866 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:22:34,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:22:34,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:22:34,018 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:22:38,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:39,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:22:39,763 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:22:39,764 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:22:39,846 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:22:40,027 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:22:42,029 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:22:43,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:48,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:49,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:22:49,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:22:49,018 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:22:54,167 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:22:57,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:22:57,073 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:22:57,074 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:22:57,155 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:22:58,038 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:22:58,039 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:23:00,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:03,872 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:23:04,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:23:04,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:23:04,019 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:23:05,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:10,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:14,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:23:14,502 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:23:14,503 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:23:14,584 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:23:15,057 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:23:15,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:16,061 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:23:19,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:23:19,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:23:19,018 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:23:21,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:26,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:31,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:31,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:23:31,375 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:23:31,376 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:23:31,458 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:23:32,101 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:23:32,102 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:23:33,878 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:23:34,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:23:34,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:23:34,018 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:23:34,107 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:23:36,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:41,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:46,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:49,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:23:49,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:23:49,631 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:23:49,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:23:49,841 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:23:49,841 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:23:49,916 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:23:50,128 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:23:51,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:23:52,129 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:23:56,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:01,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:03,883 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:24:06,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:24:06,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:24:06,855 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:24:06,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:24:07,068 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:24:07,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:07,069 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:24:07,148 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:24:08,148 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:24:08,149 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:24:12,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:17,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:21,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:24:21,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:24:21,855 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:24:23,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:24,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:24:24,641 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:24:24,642 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:24:24,726 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:24:25,160 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:24:26,160 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:24:28,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:33,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:33,890 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:24:36,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:24:36,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:24:36,857 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:24:39,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:42,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:24:42,231 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:24:42,232 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:24:42,319 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:24:43,171 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:24:44,172 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:24:44,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:49,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:51,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:24:51,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:24:51,856 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:24:55,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:24:58,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:24:58,921 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:24:58,922 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:24:59,002 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:24:59,181 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:25:00,182 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:25:01,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:03,895 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:25:06,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:25:06,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:25:06,856 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:25:07,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:12,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:16,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:25:16,409 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:25:16,410 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:25:16,492 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:25:17,192 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:25:17,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:18,192 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:25:21,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:25:21,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:25:21,856 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:25:23,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:28,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:33,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:33,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:25:33,569 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:25:33,570 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:25:33,653 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:25:33,901 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:25:34,202 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:25:34,203 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:25:36,203 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:25:36,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:25:36,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:25:36,856 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:25:38,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:43,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:48,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:51,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:25:51,816 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:25:51,816 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:25:51,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:25:51,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:25:51,898 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:25:51,899 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:25:52,213 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:25:53,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:25:54,214 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:25:58,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:03,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:03,907 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:26:08,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:09,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:26:09,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:26:09,333 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:26:09,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:26:09,534 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:26:09,534 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:26:09,612 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:26:10,225 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:26:10,225 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:26:14,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:19,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:24,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:26:24,332 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:26:25,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:26,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:26:26,876 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:26:26,877 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:26:26,958 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:26:27,236 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:26:28,236 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:26:30,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:33,913 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:26:36,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:39,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:26:39,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:26:39,333 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:26:42,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:44,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:26:44,560 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:26:44,561 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:26:44,646 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:26:45,258 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:26:46,264 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:26:48,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:53,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:26:54,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:26:54,332 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:26:59,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:01,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:27:01,355 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:27:01,356 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:27:01,436 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:27:02,274 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:27:03,275 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:27:03,919 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:27:04,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:09,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:27:09,333 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:27:10,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:15,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:18,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:27:18,938 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:27:18,940 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:27:19,019 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:27:19,285 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:27:21,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:21,286 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:27:24,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:27:24,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:27:24,334 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:27:26,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:31,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:33,924 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:27:36,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:27:36,362 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:27:36,363 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:27:36,448 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:27:37,296 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:27:37,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:39,298 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:27:39,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:27:39,333 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:27:42,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:47,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:52,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:27:54,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:27:54,177 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:27:54,178 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:27:54,259 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:27:54,307 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:27:54,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:27:54,333 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:27:55,308 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:27:57,309 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:27:57,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:02,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:03,931 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:28:07,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:12,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:28:12,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:28:12,312 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:28:12,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:28:12,528 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:28:12,529 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:28:12,609 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:28:13,322 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:28:13,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:15,323 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:28:18,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:23,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:28,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:29,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:28:29,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:28:29,504 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:28:29,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:28:29,747 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:28:29,748 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:28:29,827 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:28:30,333 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:28:31,334 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:28:33,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:33,937 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:28:38,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:43,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:46,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:28:46,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:28:46,706 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:28:46,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:28:46,971 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:28:46,972 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:28:47,051 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:28:47,344 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:28:49,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:49,345 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:28:54,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:28:59,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:01,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:29:01,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:29:01,707 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:29:03,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:03,942 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:29:04,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:04,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:04,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:04,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:04,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:08,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:09,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:10,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:10,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:10,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:10,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:10,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:10,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:11,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:14,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:14,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:16,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:16,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:16,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:16,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:29:16,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:29:16,706 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:29:16,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:16,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:17,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:17,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:20,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:20,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:22,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:22,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:22,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:23,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:23,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:23,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:23,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:25,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:27,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:28,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:28,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:28,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:29,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:29,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:29,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:30,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:30,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:31,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:29:31,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:29:31,707 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:29:33,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:33,949 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:29:35,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:35,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:35,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:35,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:35,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:35,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:35,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:36,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:39,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:40,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:41,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:41,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:41,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:41,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:42,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:42,167 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:42,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:45,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:46,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:46,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:29:46,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:29:46,707 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:29:47,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:47,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:47,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:48,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:48,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:48,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:48,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:51,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:52,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:54,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:54,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:54,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:54,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:54,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:54,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:55,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:29:56,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:29:58,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:00,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:00,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:00,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:00,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:00,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:00,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:01,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:01,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:30:01,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:30:01,707 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:30:01,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:03,954 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:30:04,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:06,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:06,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:06,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:06,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:07,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:07,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:07,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:07,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:11,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:11,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:12,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:12,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:12,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:13,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:13,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:13,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:14,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:16,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:30:16,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:30:16,707 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:30:17,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:17,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:19,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:19,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:19,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:19,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:19,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:19,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:20,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:22,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:23,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:25,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:25,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:25,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:25,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:26,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:26,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:26,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:27,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:30,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:31,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:30:31,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:30:31,708 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:30:31,844 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:31,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:31,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:32,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:32,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:32,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:32,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:33,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:33,960 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:30:36,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:38,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:38,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:38,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:38,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:38,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:38,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:38,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:39,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:42,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:43,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:44,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:44,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:44,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:44,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:45,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:45,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:45,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:46,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:30:46,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:30:46,708 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:30:49,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:49,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:50,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:50,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:50,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:51,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:51,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:51,467 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:51,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:54,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:30:55,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:57,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:57,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:57,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:57,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:57,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:57,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:58,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:30:59,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:01,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:01,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:31:01,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:31:01,708 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:31:03,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:03,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:03,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:03,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:03,966 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:31:03,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:04,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:04,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:04,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:07,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:09,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:09,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:09,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:09,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:10,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:10,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:10,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:10,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:14,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:14,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:16,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:16,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:16,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:16,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:16,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:16,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:16,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:31:16,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:31:16,709 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:31:17,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:19,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:20,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:22,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:22,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:22,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:22,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:22,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:22,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:23,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:24,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:26,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:28,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:28,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:28,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:29,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:29,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:29,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:29,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:29,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:31,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:31:31,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:31:31,709 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:31:33,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:33,972 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:31:34,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:34,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:34,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:35,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:35,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:35,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:35,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:36,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:39,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:40,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:41,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:41,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:41,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:41,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:41,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:41,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:42,363 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:45,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:45,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:46,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:31:46,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:31:46,709 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:31:47,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:47,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:47,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:48,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:48,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:48,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:48,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:51,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:52,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:53,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:53,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:53,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:54,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:54,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:54,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:54,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:31:56,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:31:58,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:00,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:00,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:00,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:00,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:00,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:00,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:01,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:01,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:32:01,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:32:01,709 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:32:01,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:03,977 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:32:04,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:06,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:06,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:06,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:06,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:06,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:07,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:07,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:07,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:11,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:11,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:12,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:12,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:12,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:13,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:13,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:13,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:13,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:16,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:32:16,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:32:16,710 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:32:17,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:17,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:19,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:19,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:19,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:19,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:19,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:19,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:20,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:22,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:23,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:25,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:25,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:25,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:25,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:27,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:31,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:32:31,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:32:31,711 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:32:32,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:33,983 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:32:37,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:41,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:42,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:46,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:32:46,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:32:46,710 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:32:48,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:53,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:32:56,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:56,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:58,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:32:58,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:01,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:01,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:01,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:33:01,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:33:01,711 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:33:02,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:02,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:02,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:02,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:02,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:03,989 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:33:03,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:04,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:07,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:08,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:08,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:08,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:08,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:08,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:09,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:09,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:10,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:13,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:14,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:14,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:14,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:14,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:14,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:15,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:15,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:16,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:33:16,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:33:16,711 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:33:17,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:19,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:20,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:20,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:21,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:21,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:21,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:21,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:21,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:23,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:24,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:26,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:26,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:27,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:27,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:27,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:28,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:28,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:29,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:29,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:31,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:33:31,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:33:31,711 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:33:33,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:33,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:33,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:33,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:33,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:33,995 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:33:34,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:34,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:34,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:36,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:39,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:39,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:39,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:39,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:40,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:40,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:40,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:40,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:42,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:44,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:45,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:45,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:46,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:46,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:46,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:46,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:33:46,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:33:46,711 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:33:46,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:46,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:48,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:50,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:52,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:52,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:52,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:52,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:52,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:53,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:53,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:55,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:55,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:33:58,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:58,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:58,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:58,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:59,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:59,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:33:59,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:00,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:01,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:01,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:34:01,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:34:01,712 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:34:04,001 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:34:04,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:04,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:05,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:05,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:05,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:05,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:05,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:06,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:07,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:10,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:10,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:11,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:11,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:11,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:11,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:12,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:12,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:13,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:16,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:16,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:34:16,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:34:16,712 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:34:17,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:17,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:17,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:17,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:17,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:18,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:18,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:20,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:21,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:23,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:23,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:24,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:24,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:24,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:24,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:24,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:26,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:26,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:29,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:29,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:30,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:30,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:30,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:31,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:31,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:31,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:34:31,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:34:31,713 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:34:31,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:32,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:34,006 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:34:36,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:36,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:36,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:36,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:36,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:37,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:37,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:37,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:39,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:42,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:42,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:42,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:42,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:43,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:43,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:43,626 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:43,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:45,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:46,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:34:46,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:34:46,713 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:34:47,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:48,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:48,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:49,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:49,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:49,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:49,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:49,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:51,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:52,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:55,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:55,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:55,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:55,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:55,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:56,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:56,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:34:57,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:34:58,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:01,375 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:01,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:01,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:35:01,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:35:01,713 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:35:01,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:01,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:02,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:02,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:02,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:02,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:04,012 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:35:04,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:07,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:07,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:08,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:08,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:08,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:08,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:08,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:08,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:10,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:13,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:13,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:13,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:14,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:14,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:14,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:15,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:15,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:16,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:35:16,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:35:16,713 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:35:16,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:18,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:20,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:20,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:20,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:20,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:20,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:21,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:21,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:23,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:23,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:26,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:26,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:27,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:27,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:27,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:27,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:27,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:28,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:29,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:31,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:35:31,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:35:31,713 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:35:32,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:32,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:33,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:33,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:33,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:33,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:34,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:34,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:34,019 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:35:35,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:39,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:39,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:39,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:39,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:39,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:39,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:40,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:40,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:41,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:44,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:45,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:45,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:45,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:45,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:46,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:46,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:46,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:46,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:35:46,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:35:46,713 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:35:48,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:49,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:51,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:51,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:52,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:52,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:52,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:52,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:52,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:54,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:54,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:35:57,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:57,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:58,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:58,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:58,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:58,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:59,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:35:59,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:00,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:01,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:36:01,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:36:01,714 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:36:04,024 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:36:04,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:04,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:04,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:04,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:04,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:05,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:05,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:05,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:06,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:10,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:10,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:10,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:10,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:10,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:11,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:11,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:11,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:13,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:15,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:16,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:16,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:16,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:36:16,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:36:16,714 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:36:17,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:17,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:17,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:17,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:17,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:20,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:25,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:30,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:31,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:36:31,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:36:31,715 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:36:32,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:34,030 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:36:36,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:40,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:40,481 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:36:40,581 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:36:40,634 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:36:40,657 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:36:41,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:41,658 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:36:46,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:46,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:36:46,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:36:46,714 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:36:51,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:56,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:36:57,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:36:57,236 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:36:57,237 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:36:57,318 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:36:57,668 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:36:59,669 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:37:01,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:37:01,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:37:01,714 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:37:02,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:04,035 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:37:08,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:13,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:14,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:37:14,609 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:37:14,610 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:37:14,694 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:37:14,695 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:37:15,694 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:37:15,695 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:37:16,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:37:16,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:37:16,714 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:37:18,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:23,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:28,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:32,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:37:32,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:37:32,564 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:37:32,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:37:32,787 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:37:32,788 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:37:32,868 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:37:33,705 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:37:33,706 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:37:34,041 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:37:34,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:39,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:44,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:49,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:49,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:37:49,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:37:49,854 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:37:49,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:37:50,068 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:37:50,069 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:37:50,149 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:37:50,715 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:37:51,716 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:37:54,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:37:59,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:04,048 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:38:05,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:05,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:38:05,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:38:05,449 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:38:08,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:38:08,943 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:38:08,944 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:38:09,025 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:38:09,727 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:38:09,728 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:38:11,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:11,729 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:38:16,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:20,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:38:20,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:38:20,450 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:38:21,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:26,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:26,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:38:26,718 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:38:26,719 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:38:26,802 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:38:27,801 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:38:27,802 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:38:31,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:34,053 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:38:35,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:38:35,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:38:35,449 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:38:37,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:42,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:45,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:38:45,996 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:38:45,996 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:38:46,076 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:38:46,812 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:38:47,813 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:38:48,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:50,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:38:50,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:38:50,449 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:38:53,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:38:58,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:03,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:39:03,141 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:39:03,142 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:39:03,222 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:39:03,823 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:39:03,823 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:39:04,059 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:39:04,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:05,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:39:05,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:39:05,450 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:39:05,824 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:39:09,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:14,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:19,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:19,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:39:19,740 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:39:19,741 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:39:19,820 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:39:19,832 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:39:20,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:39:20,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:39:20,450 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:39:21,833 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:39:24,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:29,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:34,065 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:39:35,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:36,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:39:36,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:39:36,568 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:39:36,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:39:36,718 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:39:36,719 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:39:36,799 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:39:36,841 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:39:37,842 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:39:40,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:45,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:50,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:39:54,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:39:54,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:39:54,340 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:39:54,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:39:54,562 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:39:54,563 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:39:54,647 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:39:54,853 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:39:55,853 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:39:56,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:01,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:04,071 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:40:07,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:09,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:40:09,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:40:09,339 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:40:12,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:12,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:40:12,920 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:40:12,920 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:40:13,002 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:40:13,865 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:40:13,865 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:40:18,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:23,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:24,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:40:24,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:40:24,339 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:40:28,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:30,417 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:40:30,420 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:40:30,421 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:40:30,502 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:40:30,875 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:40:31,876 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:40:34,076 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:40:34,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:39,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:39,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:40:39,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:40:39,340 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:40:44,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:47,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:40:47,835 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:40:47,837 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:40:47,918 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:40:48,917 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:40:49,918 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:40:49,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:40:54,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:40:54,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:40:54,339 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:40:55,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:00,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:04,082 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:41:05,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:41:05,123 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:41:05,123 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:41:05,212 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:41:05,928 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:41:05,928 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:41:06,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:07,929 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:41:09,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:41:09,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:41:09,340 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:41:11,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:16,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:21,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:22,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:41:22,393 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:41:22,393 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:41:22,475 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:41:22,939 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:41:23,939 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:41:24,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:41:24,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:41:24,340 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:41:26,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:31,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:34,089 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:41:37,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:39,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:41:39,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:41:39,633 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:41:39,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:41:39,868 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:41:39,868 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:41:39,948 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:41:39,950 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:41:41,951 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:41:42,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:47,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:52,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:57,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:41:57,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:41:57,321 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:41:57,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:41:57,551 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:41:57,552 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:41:57,632 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:41:57,961 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:41:58,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:41:59,962 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:42:03,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:04,092 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:42:09,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:12,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:42:12,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:42:12,321 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:42:14,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:16,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:42:16,186 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:42:16,187 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:42:16,268 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:42:16,983 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:42:17,984 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:42:20,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:25,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:27,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:42:27,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:42:27,321 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:42:30,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:33,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:42:33,132 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:42:33,133 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:42:33,217 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:42:33,994 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:42:33,994 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:42:34,098 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:42:36,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:41,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:42,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:42:42,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:42:42,322 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:42:46,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:50,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:42:50,567 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:42:50,568 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:42:50,648 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:42:51,005 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:42:51,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:52,006 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:42:56,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:42:57,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:42:57,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:42:57,321 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:43:02,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:04,104 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:43:07,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:43:07,102 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:43:07,103 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:43:07,187 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:43:08,016 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:43:08,016 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:43:08,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:12,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:43:12,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:43:12,321 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:43:13,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:18,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:23,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:24,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:43:24,748 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:43:24,748 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:43:24,830 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:43:25,026 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:43:26,027 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:43:27,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:43:27,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:43:27,322 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:43:28,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:33,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:34,110 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:43:39,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:42,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:43:42,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:43:42,650 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:43:42,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:43:42,837 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:43:42,838 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:43:42,919 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:43:43,037 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:43:44,038 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:43:44,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:49,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:54,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:43:57,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:43:57,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:43:57,650 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:44:00,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:00,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:44:00,934 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:44:00,935 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:44:01,018 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:44:01,049 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:44:02,050 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:44:04,116 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:44:06,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:11,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:12,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:44:12,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:44:12,651 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:44:16,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:19,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:44:19,493 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:44:19,494 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:44:19,576 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:44:20,060 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:44:20,060 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:44:22,061 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:44:22,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:27,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:27,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:44:27,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:44:27,651 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:44:32,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:34,122 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:44:36,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:44:36,578 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:44:36,578 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:44:36,661 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:44:37,070 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:44:38,071 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:44:38,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:42,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:44:42,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:44:42,650 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:44:43,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:48,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:53,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:44:54,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:44:54,124 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:44:54,125 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:44:54,208 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:44:55,082 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:44:56,082 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:44:57,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:44:57,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:44:57,651 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:44:59,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:04,127 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:45:05,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:10,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:10,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:45:10,750 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:45:10,752 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:45:10,832 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:45:11,092 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:45:12,093 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:45:12,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:45:12,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:45:12,651 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:45:15,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:20,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:25,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:28,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:45:28,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:45:28,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:45:28,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:45:28,854 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:45:28,855 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:45:28,935 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:45:29,104 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:45:30,104 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:45:30,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:34,133 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:45:36,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:41,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:43,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:45:43,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:45:43,617 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:45:46,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:47,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:45:47,114 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:45:47,115 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:45:47,195 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:45:47,196 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:45:48,195 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:45:52,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:57,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:45:58,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:45:58,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:46:02,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:04,139 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:46:05,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:46:05,100 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:46:05,101 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:46:05,180 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:46:05,206 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:46:06,206 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:46:08,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:13,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:13,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:46:13,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:46:18,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:23,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:46:23,333 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:46:23,334 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:46:23,415 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:46:24,217 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:46:24,218 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:46:24,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:28,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:46:28,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:46:28,617 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:46:29,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:34,145 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:46:35,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:39,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:46:39,847 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:46:39,847 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:46:39,929 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:46:40,227 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:46:40,228 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:46:40,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:42,229 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:46:43,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:46:43,617 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:46:46,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:51,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:56,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:46:57,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:46:57,141 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:46:57,142 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:46:57,222 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:46:57,238 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:46:58,239 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:46:58,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:46:58,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:47:02,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:04,151 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:47:08,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:13,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:14,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:47:14,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:47:14,225 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:47:14,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:47:14,434 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:47:14,435 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:47:14,518 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:47:15,249 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:47:16,249 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:47:18,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:23,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:28,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:31,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:47:31,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:47:31,470 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:47:31,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:47:31,662 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:47:31,663 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:47:31,743 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:47:32,260 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:47:32,260 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:47:33,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:34,156 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:47:39,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:44,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:46,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:47:46,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:47:46,471 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:47:49,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:47:50,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:47:50,508 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:47:50,509 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:47:50,592 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:47:51,272 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:47:52,272 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:47:55,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:00,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:01,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:48:01,470 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:48:04,163 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:48:06,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:07,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:48:07,418 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:48:07,419 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:48:07,503 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:48:08,283 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:48:08,283 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:48:11,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:16,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:48:16,470 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:48:16,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:21,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:25,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:48:25,719 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:48:25,720 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:48:25,801 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:48:26,294 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:48:26,295 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:48:26,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:31,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:48:31,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:48:31,471 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:48:32,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:34,169 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:48:38,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:42,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:48:42,540 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:48:42,540 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:48:42,620 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:48:43,305 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:48:43,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:44,305 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:48:46,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:48:46,470 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:48:48,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:53,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:58,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:48:59,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:48:59,772 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:48:59,772 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:48:59,852 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:49:00,315 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:49:00,316 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:49:01,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:49:01,470 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:49:04,174 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:49:04,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:09,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:14,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:17,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:49:17,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:49:17,973 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:49:17,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:49:18,178 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:49:18,179 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:49:18,262 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:49:18,327 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:49:18,327 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:49:19,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:21,329 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:49:24,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:29,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:34,180 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:49:35,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:35,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:49:35,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:49:35,323 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:49:35,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:49:35,491 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:49:35,491 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:49:35,572 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:49:36,340 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:49:36,340 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:49:37,340 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:49:40,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:45,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:50,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:49:50,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:49:50,323 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:49:51,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:49:53,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:49:53,784 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:49:53,784 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:49:53,868 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:49:54,351 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:49:55,351 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:49:56,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:01,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:04,186 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:50:05,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:50:05,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:50:05,324 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:50:07,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:10,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:50:10,777 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:50:10,777 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:50:10,858 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:50:11,361 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:50:12,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:13,363 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:50:17,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:20,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:50:20,323 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:50:23,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:28,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:50:28,466 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:50:28,467 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:50:28,547 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:50:28,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:29,373 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:50:31,374 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:50:33,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:34,192 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:50:35,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:50:35,323 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:50:39,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:44,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:45,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:50:45,727 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:50:45,728 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:50:45,808 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:50:46,384 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:50:47,385 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:50:49,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:50:50,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:50:50,324 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:50:55,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:00,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:02,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:51:02,685 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:51:02,685 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:51:02,767 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:51:03,395 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:51:04,198 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:51:05,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:51:05,323 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:51:05,396 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:51:06,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:11,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:16,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:20,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:51:20,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:51:20,936 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:51:20,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:51:21,154 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:51:21,155 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:51:21,236 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:51:21,406 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:51:22,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:23,408 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:51:27,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:32,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:34,203 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:51:38,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:38,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:51:38,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:51:38,236 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:51:38,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:51:38,488 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:51:38,489 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:51:38,575 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:51:39,438 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:51:39,438 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:51:41,442 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:51:43,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:48,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:53,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:51:53,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:51:53,236 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:51:54,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:51:56,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:51:56,466 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:51:56,467 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:51:56,551 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:51:57,460 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:51:57,460 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:51:59,461 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:51:59,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:04,209 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:52:05,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:08,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:52:08,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:52:08,236 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:52:10,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:13,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:52:13,864 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:52:13,865 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:52:13,946 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:52:14,470 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:52:15,471 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:52:15,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:20,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:23,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:52:23,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:52:23,236 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:52:26,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:30,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:52:30,980 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:52:30,980 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:52:31,060 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:52:31,481 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:52:32,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:33,483 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:52:34,215 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:52:37,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:38,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:52:38,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:52:38,236 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:52:42,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:47,467 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:48,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:52:48,133 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:52:48,134 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:52:48,216 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:52:48,492 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:52:49,493 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:52:53,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:52:53,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:52:53,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:52:53,236 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:52:58,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:03,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:04,221 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:53:05,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:53:05,606 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:53:05,607 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:53:05,687 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:53:06,503 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:53:07,504 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:53:08,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:53:08,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:53:08,236 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:53:08,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:13,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:18,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:23,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:53:23,032 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:53:23,033 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:53:23,114 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:53:23,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:53:23,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:53:23,237 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:53:23,514 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:53:24,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:25,515 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:53:29,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:34,227 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:53:35,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:40,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:41,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:53:41,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:53:41,426 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:53:41,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:53:41,657 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:53:41,657 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:53:41,737 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:53:42,526 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:53:43,527 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:53:45,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:50,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:55,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:53:58,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:53:58,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:53:58,864 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:53:58,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:53:59,036 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:53:59,037 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:53:59,117 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:53:59,537 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:54:01,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:01,538 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:54:04,232 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:54:06,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:11,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:15,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:54:15,989 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:54:15,989 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:54:15,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:54:16,209 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:54:16,210 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:54:16,288 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:54:16,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:16,547 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:54:17,548 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:54:21,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:26,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:31,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:33,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:54:33,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:54:33,517 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:54:33,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:54:33,711 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:54:33,712 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:54:33,792 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:54:34,238 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:54:34,559 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:54:35,559 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:54:37,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:42,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:47,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:50,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:54:50,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:54:50,610 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:54:50,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:54:50,802 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:54:50,803 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:54:50,884 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:54:51,569 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:54:52,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:54:53,570 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:54:57,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:02,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:04,244 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:55:05,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:55:05,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:55:05,611 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:55:08,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:09,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:55:09,094 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:55:09,095 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:55:09,176 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:55:09,580 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:55:11,581 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:55:14,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:19,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:20,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:55:20,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:55:20,611 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:55:24,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:26,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:55:26,647 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:55:26,647 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:55:26,728 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:55:27,591 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:55:29,592 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:55:30,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:34,250 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:55:35,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:55:35,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:55:35,611 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:55:35,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:40,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:43,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:55:43,743 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:55:43,744 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:55:43,824 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:55:44,602 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:55:45,603 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:55:45,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:50,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:55:50,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:55:50,611 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:55:51,786 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:55:56,786 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:01,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:56:01,610 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:56:01,610 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:56:01,695 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:56:01,696 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:56:02,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:03,695 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:56:04,256 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:56:05,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:56:05,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:56:05,611 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:56:07,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:12,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:17,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:18,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:56:18,236 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:56:18,237 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:56:18,321 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:56:18,704 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:56:19,705 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:56:20,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:56:20,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:56:20,612 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:56:22,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:27,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:32,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:34,262 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:56:35,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:56:35,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:56:35,949 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:56:35,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:56:36,154 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:56:36,154 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:56:36,235 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:56:36,715 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:56:37,716 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:56:38,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:43,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:48,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:53,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:56:53,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:56:53,187 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:56:53,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:56:53,428 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:56:53,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:56:53,428 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:56:53,509 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:56:53,725 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:56:55,726 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:56:58,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:03,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:04,267 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:57:08,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:57:08,186 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:57:09,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:11,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:57:11,598 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:57:11,598 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:57:11,681 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:57:11,736 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:57:13,737 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:57:14,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:19,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:23,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:57:23,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:57:23,188 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:57:25,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:29,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:57:29,091 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:57:29,092 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:57:29,174 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:57:29,747 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:57:31,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:31,748 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:57:34,273 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:57:36,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:38,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:57:38,187 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:57:41,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:45,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:57:45,770 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:57:45,771 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:57:45,851 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:57:46,757 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:57:46,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:47,758 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:57:51,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:57:53,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:57:53,187 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:57:57,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:02,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:04,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:58:04,128 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:58:04,129 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:58:04,210 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:58:04,279 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:58:04,769 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:58:05,769 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:58:08,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:58:08,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:58:08,187 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:58:08,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:13,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:18,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:20,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:58:20,914 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:58:20,916 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:58:20,996 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:58:21,779 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:58:21,780 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:58:23,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:58:23,187 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:58:23,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:23,781 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:58:28,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:33,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:34,285 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:58:38,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:58:38,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:58:38,458 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:58:38,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:58:38,690 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:58:38,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:38,690 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:58:38,771 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:58:38,790 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:58:39,791 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:58:43,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:48,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:53,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:58:56,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:58:56,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:58:56,479 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:58:56,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:58:56,707 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:58:56,708 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:58:56,789 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:58:56,801 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:58:57,802 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:58:58,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:03,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:04,291 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:59:09,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:13,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:59:13,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:59:13,365 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:59:13,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:59:14,023 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:59:14,024 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:59:14,104 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:59:14,812 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:59:15,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:15,813 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:59:20,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:25,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:30,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:31,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:59:31,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:59:31,886 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:59:31,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:59:32,133 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:59:32,133 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:59:32,212 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:59:32,824 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:59:33,824 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:59:34,297 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 14:59:35,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:40,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:45,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:48,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 14:59:48,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 14:59:48,727 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 14:59:48,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 14:59:48,894 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 14:59:48,895 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 14:59:48,977 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 14:59:49,834 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 14:59:49,834 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 14:59:50,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 14:59:55,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:00,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:03,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:00:03,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:00:03,727 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:00:04,302 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:00:06,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:06,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:00:06,581 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:00:06,582 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:00:06,662 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:00:06,843 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:00:07,844 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:00:11,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:16,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:18,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:00:18,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:00:18,728 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:00:21,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:24,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:00:24,951 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:00:24,953 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:00:25,037 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:00:25,855 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:00:25,855 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:00:27,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:32,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:33,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:00:33,727 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:00:34,308 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:00:37,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:42,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:44,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:00:44,904 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:00:44,904 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:00:44,984 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:00:45,867 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:00:45,868 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:00:47,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:48,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:00:48,727 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:00:53,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:00:58,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:03,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:04,314 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:01:04,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:01:04,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:01:04,722 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:01:04,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:01:04,937 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:01:04,937 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:01:05,017 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:01:05,880 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:01:05,880 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:01:09,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:14,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:19,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:19,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:01:19,722 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:01:23,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:01:23,169 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:01:23,170 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:01:23,250 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:01:23,898 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:01:23,898 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:01:24,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:25,899 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:01:29,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:34,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:34,320 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:01:34,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:01:34,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:01:34,722 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:01:39,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:43,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:01:43,136 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:01:43,137 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:01:43,219 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:01:43,910 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:01:43,911 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:01:45,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:49,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:01:49,722 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:01:50,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:01:55,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:00,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:02:00,654 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:02:00,654 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:02:00,742 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:02:00,920 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:02:01,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:01,921 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:02:04,326 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:02:04,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:02:04,722 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:02:06,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:11,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:16,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:19,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:02:19,072 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:02:19,073 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:02:19,154 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:02:19,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:02:19,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:02:19,722 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:02:19,932 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:02:19,932 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:02:21,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:26,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:31,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:34,332 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:02:36,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:02:36,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:02:36,129 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:02:36,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:02:36,288 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:02:36,288 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:02:36,369 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:02:36,942 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:02:37,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:37,943 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:02:42,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:47,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:51,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:02:51,128 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:02:53,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:02:55,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:02:55,612 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:02:55,613 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:02:55,696 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:02:55,954 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:02:57,956 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:02:58,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:03,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:04,337 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:03:06,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:03:06,130 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:03:06,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:03:09,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:13,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:03:13,270 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:03:13,270 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:03:13,351 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:03:13,965 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:03:13,966 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:03:14,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:19,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:21,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:03:21,128 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:03:24,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:29,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:31,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:03:31,420 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:03:31,420 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:03:31,502 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:03:31,977 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:03:31,977 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:03:33,978 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:03:34,343 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:03:35,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:36,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:03:36,129 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:03:41,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:46,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:48,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:03:48,948 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:03:48,949 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:03:49,030 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:03:49,030 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:03:50,030 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:03:51,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:03:51,129 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:03:51,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:03:56,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:01,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:04,350 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:04:06,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:07,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:04:07,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:04:07,800 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:04:07,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:04:07,988 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:04:07,989 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:04:08,071 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:04:08,071 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:04:10,072 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:04:12,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:17,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:22,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:25,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:04:25,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:04:25,215 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:04:25,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:04:25,407 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:04:25,407 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:04:25,487 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:04:26,083 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:04:26,083 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:04:27,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:32,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:34,355 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:04:38,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:41,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:04:41,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:04:41,906 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:04:41,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:04:42,132 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:04:42,133 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:04:42,213 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:04:43,094 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:04:44,094 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:04:44,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:49,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:54,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:04:57,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:04:57,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:04:57,030 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:04:59,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:00,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:05:00,314 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:05:00,315 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:05:00,397 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:05:01,104 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:05:02,105 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:05:04,361 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:05:05,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:10,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:12,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:05:12,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:05:12,030 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:05:16,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:17,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:05:17,770 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:05:17,771 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:05:17,851 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:05:18,115 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:05:20,116 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:05:21,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:26,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:27,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:05:27,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:05:27,030 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:05:32,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:34,367 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:05:35,560 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:05:35,563 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:05:35,564 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:05:35,646 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:05:36,126 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:05:36,126 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:05:37,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:42,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:05:42,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:05:42,030 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:05:43,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:48,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:52,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:05:52,762 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:05:52,763 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:05:52,845 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:05:53,137 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:05:53,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:05:54,137 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:05:57,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:05:57,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:05:57,031 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:05:59,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:04,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:04,372 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:06:09,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:09,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:10,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:10,456 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:10,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:10,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:12,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:06:12,030 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:06:14,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:15,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:16,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:16,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:16,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:16,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:16,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:16,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:17,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:20,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:20,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:22,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:22,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:22,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:22,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:22,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:23,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:23,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:25,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:26,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:27,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:06:27,031 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:06:28,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:28,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:28,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:29,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:29,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:29,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:29,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:31,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:33,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:34,379 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:06:34,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:34,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:34,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:35,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:35,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:35,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:35,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:36,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:39,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:41,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:41,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:41,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:41,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:41,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:41,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:41,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:42,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:06:42,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:06:42,032 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:06:42,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:45,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:47,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:47,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:47,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:47,363 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:47,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:47,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:48,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:48,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:51,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:52,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:53,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:53,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:53,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:54,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:54,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:54,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:54,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:57,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:06:57,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:06:57,032 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:06:57,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:06:58,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:59,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:59,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:06:59,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:00,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:00,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:00,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:00,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:02,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:04,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:04,384 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:07:06,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:06,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:06,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:06,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:06,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:06,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:07,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:07,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:10,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:12,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:07:12,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:07:12,032 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:07:12,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:12,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:12,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:12,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:12,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:13,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:13,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:13,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:16,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:18,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:18,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:18,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:18,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:19,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:19,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:19,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:19,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:23,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:23,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:24,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:24,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:24,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:25,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:25,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:25,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:26,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:27,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:07:27,032 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:07:28,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:29,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:31,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:31,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:31,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:31,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:31,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:31,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:32,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:33,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:34,390 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:07:35,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:37,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:37,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:37,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:38,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:38,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:38,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:38,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:38,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:42,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:07:42,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:07:42,033 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:07:42,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:43,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:43,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:43,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:44,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:44,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:44,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:44,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:44,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:48,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:49,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:50,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:50,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:50,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:50,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:50,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:50,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:51,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:54,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:07:54,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:56,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:56,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:56,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:56,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:56,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:56,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:57,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:07:57,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:07:57,033 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:07:57,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:07:59,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:00,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:02,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:02,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:02,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:03,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:03,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:03,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:03,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:04,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:04,395 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:08:07,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:08,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:08,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:08,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:09,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:09,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:09,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:09,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:09,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:12,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:08:12,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:08:12,033 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:08:13,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:15,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:15,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:15,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:15,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:15,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:15,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:15,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:16,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:19,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:20,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:21,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:21,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:21,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:21,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:21,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:22,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:22,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:25,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:25,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:27,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:08:27,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:08:27,033 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:08:27,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:27,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:27,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:28,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:28,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:28,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:28,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:30,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:32,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:33,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:34,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:34,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:34,401 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:08:34,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:34,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:34,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:35,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:35,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:38,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:40,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:40,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:40,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:40,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:40,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:40,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:40,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:41,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:42,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:08:42,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:08:42,034 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:08:44,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:46,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:46,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:46,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:46,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:46,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:47,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:47,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:47,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:51,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:51,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:52,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:52,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:52,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:53,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:53,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:53,436 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:53,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:56,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:08:57,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:08:57,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:08:57,034 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:08:57,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:58,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:58,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:58,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:59,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:59,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:08:59,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:00,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:02,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:03,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:04,407 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:09:05,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:05,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:05,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:05,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:05,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:05,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:06,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:07,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:09,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:11,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:11,620 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:11,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:12,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:09:12,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:09:12,034 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:09:12,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:12,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:12,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:12,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:13,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:16,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:17,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:17,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:17,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:18,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:18,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:18,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:18,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:18,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:22,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:23,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:24,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:24,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:24,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:24,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:24,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:24,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:25,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:27,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:09:27,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:09:27,034 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:09:28,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:28,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:30,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:30,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:30,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:30,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:33,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:34,413 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:09:38,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:42,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:09:42,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:09:42,035 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:09:44,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:45,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:09:49,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:54,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:09:57,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:09:57,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:09:57,036 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:10:00,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:01,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:01,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:03,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:04,419 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:10:05,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:06,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:06,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:06,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:07,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:07,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:07,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:07,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:09,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:10,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:12,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:10:12,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:10:12,036 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:10:12,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:12,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:13,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:13,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:13,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:13,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:13,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:15,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:16,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:19,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:19,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:19,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:19,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:19,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:20,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:20,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:21,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:21,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:25,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:25,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:25,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:25,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:25,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:26,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:26,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:26,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:27,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:10:27,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:10:27,036 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:10:28,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:31,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:31,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:32,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:32,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:32,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:32,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:32,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:32,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:34,424 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:10:34,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:37,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:37,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:37,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:38,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:38,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:38,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:38,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:38,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:40,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:42,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:10:42,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:10:42,036 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:10:43,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:44,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:44,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:44,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:44,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:44,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:45,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:45,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:47,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:48,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:50,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:50,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:50,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:50,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:51,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:51,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:51,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:53,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:53,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:56,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:56,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:57,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:10:57,036 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:10:57,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:57,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:57,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:57,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:57,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:10:58,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:10:59,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:02,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:02,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:03,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:03,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:03,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:03,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:03,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:03,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:04,430 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:11:05,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:08,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:09,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:09,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:09,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:09,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:09,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:10,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:10,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:12,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:11:12,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:11:12,037 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:11:12,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:14,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:15,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:15,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:15,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:15,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:16,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:16,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:16,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:18,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:19,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:21,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:21,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:22,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:22,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:22,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:22,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:22,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:24,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:24,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:27,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:11:27,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:11:27,037 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:11:27,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:27,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:28,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:28,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:28,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:29,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:29,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:30,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:30,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:34,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:34,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:34,436 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:11:34,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:34,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:34,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:35,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:35,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:35,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:37,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:40,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:40,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:40,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:40,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:41,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:41,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:41,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:41,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:42,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:11:42,037 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:11:43,363 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:46,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:46,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:46,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:47,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:47,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:47,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:48,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:48,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:49,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:51,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:53,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:53,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:53,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:53,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:53,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:54,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:54,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:55,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:56,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:11:57,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:11:57,037 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:11:59,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:59,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:59,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:11:59,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:00,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:00,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:00,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:02,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:02,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:04,442 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:12:05,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:05,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:06,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:06,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:06,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:06,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:06,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:07,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:08,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:11,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:11,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:12,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:12:12,038 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:12:12,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:12,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:12,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:13,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:13,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:13,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:14,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:18,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:18,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:18,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:18,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:18,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:18,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:19,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:19,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:21,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:23,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:24,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:24,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:24,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:25,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:25,167 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:25,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:25,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:27,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:12:27,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:12:27,038 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:12:27,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:28,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:30,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:30,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:31,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:31,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:31,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:31,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:31,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:33,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:33,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:34,448 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:12:36,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:37,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:37,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:37,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:37,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:38,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:38,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:38,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:39,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:42,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:12:42,038 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:12:43,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:43,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:43,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:43,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:43,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:44,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:44,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:44,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:46,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:49,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:49,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:49,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:50,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:50,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:50,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:50,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:50,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:52,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:54,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:12:55,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:55,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:56,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:56,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:56,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:56,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:56,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:57,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:12:57,038 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:12:58,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:12:59,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:02,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:02,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:02,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:02,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:02,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:03,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:03,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:04,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:04,454 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:13:04,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:08,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:08,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:08,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:08,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:08,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:09,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:09,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:09,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:11,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:12,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:13:12,038 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:13:14,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:14,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:15,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:15,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:15,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:15,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:15,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:15,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:17,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:20,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:20,844 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:20,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:21,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:21,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:21,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:21,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:21,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:25,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:27,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:13:27,039 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:13:31,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:34,459 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:13:36,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:36,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:41,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:46,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:49,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:13:49,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:13:49,429 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:13:49,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:13:49,650 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:13:49,751 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:13:49,816 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:13:50,522 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:13:51,523 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:13:51,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:13:56,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:01,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:04,465 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:14:07,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:14:07,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:14:07,459 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:14:07,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:14:07,665 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:14:07,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:07,666 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:14:07,746 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:14:08,532 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:14:09,533 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:14:12,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:17,749 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:22,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:14:22,458 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:14:23,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:24,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:14:24,921 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:14:24,922 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:14:25,003 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:14:25,542 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:14:27,544 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:14:29,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:34,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:34,471 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:14:37,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:14:37,458 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:14:39,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:42,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:14:42,403 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:14:42,403 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:14:42,484 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:14:42,553 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:14:45,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:45,555 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:14:50,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:52,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:14:52,458 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:14:55,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:14:59,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:14:59,164 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:14:59,165 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:14:59,252 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:14:59,563 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:15:01,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:01,565 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:15:04,477 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:15:06,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:07,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:15:07,458 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:15:11,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:16,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:16,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:15:16,978 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:15:16,978 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:15:17,016 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:15:17,575 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:15:19,576 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:15:22,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:22,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:15:22,458 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:15:27,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:32,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:34,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:15:34,212 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:15:34,212 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:15:34,292 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:15:34,483 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:15:34,586 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:15:35,587 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:15:37,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:15:37,459 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:15:37,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:42,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:47,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:52,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:15:52,216 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:15:52,217 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:15:52,297 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:15:52,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:15:52,459 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:15:52,598 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:15:53,598 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:15:53,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:15:58,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:03,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:04,489 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:16:08,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:16:08,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:16:08,973 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:16:08,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:16:09,134 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:16:09,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:09,135 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:16:09,216 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:16:09,608 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:16:11,610 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:16:14,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:19,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:24,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:26,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:16:26,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:16:26,102 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:16:26,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:16:26,330 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:16:26,331 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:16:26,411 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:16:26,619 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:16:27,620 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:16:29,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:34,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:34,495 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:16:39,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:43,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:16:43,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:16:43,377 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:16:43,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:16:43,558 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:16:43,559 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:16:43,640 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:16:44,639 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:16:44,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:45,640 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:16:49,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:54,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:16:59,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:00,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:17:00,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:17:00,761 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:17:00,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:17:00,975 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:17:00,976 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:17:01,060 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:17:01,650 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:17:03,651 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:17:04,500 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:17:05,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:10,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:15,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:15,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:17:15,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:17:15,800 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:17:19,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:17:19,039 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:17:19,040 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:17:19,123 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:17:19,662 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:17:21,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:21,663 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:17:26,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:30,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:17:30,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:17:30,800 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:17:31,989 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:34,506 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:17:35,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:17:35,766 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:17:35,767 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:17:35,848 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:17:36,672 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:17:37,673 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:17:37,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:42,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:45,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:17:45,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:17:45,801 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:17:47,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:52,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:17:53,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:17:53,404 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:17:53,404 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:17:53,484 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:17:53,683 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:17:55,684 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:17:58,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:00,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:18:00,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:18:00,800 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:18:03,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:04,513 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:18:09,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:09,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:18:09,993 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:18:09,994 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:18:10,075 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:18:10,694 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:18:11,695 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:18:15,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:15,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:18:15,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:18:15,800 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:18:21,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:26,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:27,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:18:27,720 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:18:27,721 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:18:27,802 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:18:28,705 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:18:29,706 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:18:30,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:18:30,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:18:30,801 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:18:31,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:34,518 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:18:37,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:42,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:45,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:18:45,242 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:18:45,242 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:18:45,326 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:18:45,716 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:18:45,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:18:45,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:18:45,801 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:18:47,717 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:18:48,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:53,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:18:58,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:02,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:19:02,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:19:02,252 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:19:02,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:19:02,415 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:19:02,415 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:19:02,495 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:19:02,727 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:19:03,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:03,728 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:19:04,524 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:19:08,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:13,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:17,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:19:17,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:19:17,251 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:19:19,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:20,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:19:20,541 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:19:20,542 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:19:20,627 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:19:20,739 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:19:21,740 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:19:24,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:29,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:33,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:19:33,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:19:33,371 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:19:34,530 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:19:35,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:36,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:19:36,767 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:19:36,767 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:19:36,847 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:19:37,750 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:19:37,750 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:19:40,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:45,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:48,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:19:48,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:19:48,372 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:19:51,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:19:54,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:19:54,504 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:19:54,504 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:19:54,584 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:19:54,760 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:19:55,761 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:19:57,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:02,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:03,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:20:03,371 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:20:04,535 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:20:08,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:11,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:20:11,556 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:20:11,557 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:20:11,638 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:20:11,771 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:20:13,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:13,773 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:20:18,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:20:18,371 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:20:19,626 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:24,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:29,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:20:29,113 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:20:29,114 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:20:29,192 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:20:29,793 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:20:29,794 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:20:30,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:31,798 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:20:33,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:20:33,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:20:33,372 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:20:34,541 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:20:35,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:40,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:45,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:47,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:20:47,165 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:20:47,166 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:20:47,246 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:20:47,838 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:20:47,838 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:20:48,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:20:48,372 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:20:49,843 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:20:51,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:20:56,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:01,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:03,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:21:03,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:21:03,615 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:21:03,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:21:03,772 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:21:03,772 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:21:03,852 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:21:03,852 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:21:04,545 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:21:05,853 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:21:06,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:11,550 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:16,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:18,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:21:18,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:21:18,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:21:21,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:22,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:21:22,268 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:21:22,269 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:21:22,350 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:21:22,864 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:21:23,865 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:21:27,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:32,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:33,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:21:33,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:21:33,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:21:34,551 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:21:37,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:38,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:21:38,631 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:21:38,633 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:21:38,713 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:21:38,875 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:21:39,875 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:21:42,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:47,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:48,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:21:48,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:21:48,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:21:52,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:21:55,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:21:55,977 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:21:55,977 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:21:56,058 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:21:56,886 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:21:57,886 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:21:58,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:03,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:03,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:22:03,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:22:03,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:22:04,557 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:22:08,560 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:13,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:22:13,446 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:22:13,447 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:22:13,532 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:22:13,897 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:22:14,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:15,898 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:22:18,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:22:18,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:22:18,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:22:19,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:24,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:29,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:30,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:22:30,144 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:22:30,144 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:22:30,230 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:22:30,907 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:22:31,908 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:22:33,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:22:33,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:22:33,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:22:34,563 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:22:35,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:40,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:45,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:49,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:22:49,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:22:49,171 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:22:49,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:22:49,405 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:22:49,406 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:22:49,484 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:22:49,920 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:22:51,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:22:51,921 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:22:56,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:01,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:04,568 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:23:05,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:23:05,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:23:05,669 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:23:05,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:23:05,887 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:23:05,888 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:23:05,968 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:23:06,968 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:23:06,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:07,969 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:23:11,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:16,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:21,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:23,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:23:23,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:23:23,501 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:23:23,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:23:23,690 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:23:23,690 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:23:23,773 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:23:23,979 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:23:25,980 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:23:27,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:32,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:34,574 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:23:38,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:40,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:23:40,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:23:40,085 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:23:40,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:23:40,306 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:23:40,307 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:23:40,387 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:23:40,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:23:41,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:23:44,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:49,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:54,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:23:56,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:23:56,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-15 15:23:56,746 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: stop_status
+2024-05-15 15:23:56,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:24:01,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:06,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:15,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:15,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:24:20,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:25,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:31,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:31,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:24:36,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:37,493 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:24:37,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:24:37,494 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:24:37,574 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:24:37,574 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:24:37,576 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:24:37,576 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:24:37,628 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:24:37,629 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:24:37,629 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:24:37,667 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:24:37,667 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:24:38,025 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:24:38,025 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:24:40,026 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:24:41,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:42,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:24:47,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:24:50,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:50,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:24:50,518 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:24:50,518 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:24:50,599 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:24:51,033 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:24:52,034 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:24:53,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:24:55,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:24:58,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:00,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:03,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:04,592 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:25:07,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:07,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:25:07,237 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:25:07,239 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:25:07,320 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:25:08,045 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:25:08,045 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:25:09,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:12,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:14,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:17,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:19,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:24,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:24,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:24,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:25:24,431 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:25:24,432 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:25:24,512 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:25:25,056 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:25:26,056 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:25:29,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:29,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:34,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:34,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:34,598 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:25:39,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:42,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:42,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:25:42,085 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:25:42,086 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:25:42,166 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:25:43,067 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:25:44,068 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:25:45,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:47,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:50,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:52,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:55,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:25:58,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:58,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:25:58,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:25:58,703 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:25:58,703 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:25:58,784 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:25:59,081 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:26:00,081 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:26:00,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:03,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:04,603 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:26:06,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:08,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:11,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:16,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:17,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:17,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:17,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:26:17,300 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:26:17,301 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:26:17,383 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:26:18,092 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:26:18,093 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:26:22,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:22,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:27,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:27,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:32,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:33,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:33,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:26:33,993 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:26:33,994 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:26:34,077 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:26:34,102 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:26:34,609 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:26:36,103 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:26:37,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:38,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:42,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:43,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:47,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:51,001 WARNING FileStreamThread:21069 [file_stream.py:request_with_retry():674] requests_with_retry encountered retryable exception: HTTPSConnectionPool(host='api.wandb.ai', port=443): Read timed out. (read timeout=180.0). func: functools.partial(>, timeout=180.0), args: ('https://api.wandb.ai/files/autometa/pyg-1-m-2.3/ky6h7dv9/file_stream',), kwargs: {'json': {'files': {'output.log': {'offset': 10, 'content': ['ERROR 2024-05-15T15:23:25.777049 2%|█ | 333/18756 [2:00:19<89:32:10, 17.50s/it]\n']}}, 'dropped': 0}}
+2024-05-15 15:26:51,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:51,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:26:51,911 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:26:51,913 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:26:51,993 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:26:52,113 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:26:52,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:26:54,113 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:26:56,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:26:57,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:01,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:02,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:04,617 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:27:08,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:09,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:09,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:27:09,234 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:27:09,235 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:27:09,319 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:27:10,134 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:27:10,134 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:27:14,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:14,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:19,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:19,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:24,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:26,026 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:26,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:27:26,034 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:27:26,034 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:27:26,130 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:27:26,143 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:27:29,145 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:27:30,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:31,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:34,622 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:27:35,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:36,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:40,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:43,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:43,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:43,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:27:43,988 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:27:43,989 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:27:44,070 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:27:44,154 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:27:46,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:47,156 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:27:48,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:51,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:27:53,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:27:56,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:00,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:00,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:00,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:28:00,688 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:28:00,690 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:28:00,783 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:28:01,164 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:28:01,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:03,166 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:28:04,627 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:28:05,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:07,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:10,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:12,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:17,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:19,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:19,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:28:19,029 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:28:19,029 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:28:19,109 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:28:19,176 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:28:21,177 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:28:23,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:24,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:28,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:29,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:33,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:34,633 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:28:36,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:36,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:28:36,131 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:28:36,131 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:28:36,212 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:28:37,211 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:28:38,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:39,212 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:28:41,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:43,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:46,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:48,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:51,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:53,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:28:53,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:28:53,605 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:28:53,605 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:28:53,686 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:28:54,221 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:28:55,222 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:28:56,598 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:28:58,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:01,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:03,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:04,639 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:29:07,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:09,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:11,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:29:11,240 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:29:11,241 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:29:11,323 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:29:12,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:12,232 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:29:13,233 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:29:15,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:17,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:20,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:22,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:25,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:27,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:27,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:29:27,842 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:29:27,843 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:29:27,926 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:29:28,242 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:29:29,243 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:29:30,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:32,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:34,645 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:29:36,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:37,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:41,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:45,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:45,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:29:45,365 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:29:45,366 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:29:45,447 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:29:46,252 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:29:47,252 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:29:47,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:50,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:52,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:29:55,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:29:57,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:02,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:02,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:02,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:30:02,394 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:30:02,396 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:30:02,479 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:30:02,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:03,263 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:30:04,650 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:30:05,264 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:30:07,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:07,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:12,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:12,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:17,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:17,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:17,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:20,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:30:20,711 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:30:20,712 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:30:20,791 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:30:21,275 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:30:22,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:22,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:23,276 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:30:27,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:27,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:32,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:32,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:34,656 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:30:38,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:38,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:38,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:30:38,194 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:30:38,195 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:30:38,280 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:30:38,286 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:30:39,286 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:30:41,287 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:30:43,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:43,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:48,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:48,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:53,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:30:55,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:30:55,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:30:55,712 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:30:55,712 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:30:55,792 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:30:56,311 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:30:57,311 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:30:58,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:00,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:03,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:04,661 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:31:05,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:09,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:13,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:13,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:31:13,033 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:31:13,034 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:31:13,116 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:31:13,321 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:31:15,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:15,322 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:31:18,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:20,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:23,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:25,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:29,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:29,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:31:29,636 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:31:29,639 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:31:29,723 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:31:30,332 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:31:30,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:31,332 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:31:34,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:34,668 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:31:36,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:39,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:41,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:45,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:46,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:47,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:31:47,540 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:31:47,541 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:31:47,621 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:31:48,344 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:31:49,345 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:31:50,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:50,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:52,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:31:55,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:31:57,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:00,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:02,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:04,674 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:32:04,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:32:04,760 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:32:04,760 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:32:04,840 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:32:05,355 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:32:05,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:05,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:07,357 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:32:07,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:10,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:12,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:15,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:17,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:22,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:23,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:23,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:23,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:32:23,135 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:32:23,136 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:32:23,217 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:32:23,367 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:32:25,368 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:32:28,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:28,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:33,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:33,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:34,679 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:32:38,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:40,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:40,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:32:40,388 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:32:40,389 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:32:40,468 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:32:41,376 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:32:41,376 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:32:43,377 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:32:44,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:45,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:49,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:50,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:54,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:32:57,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:32:57,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:32:57,566 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:32:57,566 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:32:57,648 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:32:58,387 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:32:59,387 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:32:59,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:02,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:04,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:04,684 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:33:07,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:09,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:14,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:14,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:33:14,492 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:33:14,492 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:33:14,572 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:33:15,397 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:33:15,398 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:33:15,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:17,398 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:33:19,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:20,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:24,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:25,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:30,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:31,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:31,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:33:31,554 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:33:31,554 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:33:31,635 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:33:32,408 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:33:33,409 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:33:34,690 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:33:35,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:36,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:40,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:41,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:45,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:49,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:49,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:49,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:33:49,841 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:33:49,842 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:33:49,922 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:33:50,420 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:33:50,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:51,420 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:33:54,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:33:55,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:33:59,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:00,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:04,696 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:34:06,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:07,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:07,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:07,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:34:07,213 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:34:07,213 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:34:07,294 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:34:07,430 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:34:09,432 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:34:12,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:12,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:17,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:17,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:22,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:25,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:25,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:25,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:34:25,063 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:34:25,063 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:34:25,143 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:34:25,440 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:34:27,441 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:34:28,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:30,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:33,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:34,702 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:34:35,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:38,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:41,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:41,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:41,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:34:41,999 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:34:41,999 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:34:42,080 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:34:42,451 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:34:43,452 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:34:44,081 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:46,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:49,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:51,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:54,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:59,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:34:59,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:34:59,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:34:59,128 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:34:59,129 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:34:59,209 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:34:59,462 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:35:01,463 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:35:04,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:35:04,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:04,708 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:35:09,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:35:09,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:14,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:16,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:35:16,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:35:16,255 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:35:16,256 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:35:16,336 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:35:16,475 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:35:17,476 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:35:20,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:21,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:35:25,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:26,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:35:30,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:33,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:35:33,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:35:33,544 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:35:33,545 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:35:33,625 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:35:34,486 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:35:34,714 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:35:35,487 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:35:35,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:38,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:35:40,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:43,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:35:45,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:50,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:51,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:35:51,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:35:51,798 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:35:51,799 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:35:51,880 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:35:52,497 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:35:53,498 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:35:55,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:35:56,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:00,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:01,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:04,720 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:36:06,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:08,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:08,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:08,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:36:08,726 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:36:08,727 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:36:08,807 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:36:09,508 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:36:09,508 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:36:11,509 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:36:11,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:13,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:16,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:18,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:21,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:26,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:26,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:26,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:36:26,581 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:36:26,581 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:36:26,665 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:36:27,519 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:36:27,519 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:36:27,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:31,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:32,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:34,725 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:36:36,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:37,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:42,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:43,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:43,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:43,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:36:43,412 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:36:43,413 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:36:43,494 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:36:43,529 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:36:45,529 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:36:48,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:48,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:53,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:36:53,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:36:58,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:00,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:00,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:00,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:37:00,487 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:37:00,488 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:37:00,569 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:37:01,568 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:37:01,568 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:37:03,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:04,731 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:37:05,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:08,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:10,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:13,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:18,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:18,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:18,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:37:18,199 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:37:18,200 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:37:18,281 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:37:18,577 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:37:19,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:19,578 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:37:23,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:24,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:28,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:29,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:34,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:34,737 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:37:35,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:35,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:37:35,356 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:37:35,356 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:37:35,438 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:37:35,588 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:37:37,589 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:37:39,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:40,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:44,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:45,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:49,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:53,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:53,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:53,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:37:53,773 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:37:53,775 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:37:53,857 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:37:54,599 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:37:54,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:37:55,600 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:37:58,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:37:59,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:03,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:04,744 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:38:05,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:10,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:10,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:38:10,681 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:38:10,682 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:38:10,763 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:38:10,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:11,610 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:38:11,611 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:38:15,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:15,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:20,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:20,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:25,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:28,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:28,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:38:28,270 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:38:28,270 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:38:28,351 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:38:28,621 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:38:29,621 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:38:31,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:33,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:34,749 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:38:36,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:38,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:41,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:44,989 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:44,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:38:44,997 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:38:44,997 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:38:45,078 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:38:45,632 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:38:45,632 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:38:47,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:47,633 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:38:49,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:52,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:38:54,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:38:57,081 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:02,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:02,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:39:02,065 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:39:02,065 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:39:02,146 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:39:02,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:02,643 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:39:03,643 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:39:04,755 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:39:07,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:07,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:12,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:12,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:17,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:20,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:20,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:20,375 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:39:20,379 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:39:20,379 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:39:20,460 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:39:20,654 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:39:21,655 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:39:23,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:25,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:28,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:30,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:33,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:34,761 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:39:37,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:37,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:37,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:39:37,543 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:39:37,544 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:39:37,627 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:39:37,665 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:39:38,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:39,666 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:39:42,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:43,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:47,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:48,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:53,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:39:55,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:55,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:39:55,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:39:55,857 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:39:55,857 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:39:55,938 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:39:56,677 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:39:57,678 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:39:58,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:00,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:03,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:04,767 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:40:05,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:09,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:12,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:12,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:40:12,804 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:40:12,804 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:40:12,888 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:40:13,718 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:40:13,719 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:40:14,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:17,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:19,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:22,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:24,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:29,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:29,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:40:29,594 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:40:29,594 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:40:29,675 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:40:29,743 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:40:30,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:31,744 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:40:34,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:34,773 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:40:35,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:39,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:40,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:45,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:47,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:47,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:40:47,136 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:40:47,137 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:40:47,219 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:40:47,754 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:40:47,755 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:40:49,755 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:40:51,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:52,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:40:56,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:40:57,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:01,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:04,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:04,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:41:04,312 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:41:04,313 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:41:04,394 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:41:04,765 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:41:04,778 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:41:05,766 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:41:06,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:09,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:11,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:14,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:16,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:21,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:22,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:22,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:22,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:41:22,604 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:41:22,605 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:41:22,686 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:41:22,777 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:41:23,777 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:41:27,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:27,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:32,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:32,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:34,785 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:41:37,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:39,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:39,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:39,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:41:39,999 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:41:39,999 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:41:40,079 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:41:40,802 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:41:41,803 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:41:43,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:44,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:48,081 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:49,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:53,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:57,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:41:57,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:41:57,741 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:41:57,741 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:41:57,822 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:41:57,823 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:41:58,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:41:59,823 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:42:02,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:03,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:04,790 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:42:07,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:09,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:14,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:14,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:14,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:14,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:14,813 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:42:14,815 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:42:14,896 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:42:15,895 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:42:15,896 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:42:19,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:19,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:24,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:24,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:29,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:31,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:31,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:31,640 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:42:31,641 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:42:31,723 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:42:31,906 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:42:33,907 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:42:34,797 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:42:35,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:36,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:40,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:41,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:45,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:46,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:48,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:48,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:49,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:49,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:49,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:50,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:51,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:52,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:54,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:54,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:54,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:55,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:55,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:55,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:55,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:42:55,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:42:56,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:42:59,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:00,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:00,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:00,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:00,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:01,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:01,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:01,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:01,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:01,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:04,802 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:43:05,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:05,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:06,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:07,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:07,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:07,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:07,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:07,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:07,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:08,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:10,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:11,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:11,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:13,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:13,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:13,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:13,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:13,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:14,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:14,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:15,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:16,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:17,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:19,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:19,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:19,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:20,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:20,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:20,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:20,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:20,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:21,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:24,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:25,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:25,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:25,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:25,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:26,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:26,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:26,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:26,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:27,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:30,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:30,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:31,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:32,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:32,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:32,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:32,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:32,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:32,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:33,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:34,808 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:43:35,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:36,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:36,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:38,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:38,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:38,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:38,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:39,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:39,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:39,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:40,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:41,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:42,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:44,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:44,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:44,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:45,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:45,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:45,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:45,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:45,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:46,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:49,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:50,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:51,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:51,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:51,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:51,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:51,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:51,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:51,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:52,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:55,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:55,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:43:56,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:43:57,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:57,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:57,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:57,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:57,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:57,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:43:58,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:00,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:01,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:01,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:03,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:03,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:03,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:04,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:04,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:04,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:04,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:04,813 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:44:05,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:06,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:08,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:09,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:09,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:09,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:10,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:10,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:10,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:10,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:10,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:11,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:14,436 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:15,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:16,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:16,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:16,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:16,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:16,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:16,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:16,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:17,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:20,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:20,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:21,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:22,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:22,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:22,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:22,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:23,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:23,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:23,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:25,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:26,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:26,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:28,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:28,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:28,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:29,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:29,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:29,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:29,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:30,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:31,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:33,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:34,819 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:44:35,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:35,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:35,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:35,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:35,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:35,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:35,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:36,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:36,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:39,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:40,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:41,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:41,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:41,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:41,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:41,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:41,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:41,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:42,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:45,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:45,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:46,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:47,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:47,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:47,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:47,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:48,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:48,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:48,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:50,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:51,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:52,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:53,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:53,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:53,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:54,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:54,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:54,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:54,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:44:55,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:44:56,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:44:58,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:00,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:00,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:00,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:00,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:00,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:00,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:00,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:01,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:01,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:04,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:04,825 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:45:05,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:06,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:06,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:06,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:06,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:06,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:06,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:06,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:07,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:10,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:10,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:11,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:12,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:12,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:12,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:13,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:13,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:13,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:13,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:15,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:16,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:17,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:18,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:18,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:18,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:19,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:19,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:19,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:19,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:20,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:21,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:23,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:25,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:25,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:25,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:25,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:25,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:25,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:25,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:26,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:26,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:29,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:30,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:31,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:31,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:31,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:31,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:31,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:31,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:32,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:32,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:34,831 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:45:35,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:35,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:36,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:37,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:37,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:37,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:38,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:38,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:38,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:38,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:40,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:41,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:42,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:43,989 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:44,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:44,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:44,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:44,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:44,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:45,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:45,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:46,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:48,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:50,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:50,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:50,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:50,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:50,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:50,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:50,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:51,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:51,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:54,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:55,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:45:56,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:56,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:56,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:56,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:45:57,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:57,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:57,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:45:57,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:00,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:01,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:01,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:02,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:02,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:02,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:03,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:03,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:03,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:03,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:04,837 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:46:05,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:06,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:07,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:09,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:09,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:09,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:09,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:10,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:11,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:15,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:16,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:20,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:21,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:24,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:25,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:26,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:30,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:31,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:34,842 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:46:35,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:36,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:39,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:40,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:40,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:41,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:41,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:45,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:45,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:45,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:45,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:45,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:45,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:46,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:46,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:46,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:48,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:50,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:51,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:51,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:51,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:51,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:52,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:52,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:52,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:52,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:54,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:55,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:46:56,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:46:57,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:57,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:58,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:58,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:58,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:58,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:46:58,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:00,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:00,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:01,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:03,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:03,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:04,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:04,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:04,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:04,848 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:47:05,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:05,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:05,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:06,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:06,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:10,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:10,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:10,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:10,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:10,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:10,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:11,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:11,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:11,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:13,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:15,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:16,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:16,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:16,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:16,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:16,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:17,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:17,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:17,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:19,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:20,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:21,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:22,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:22,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:23,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:23,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:23,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:23,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:23,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:25,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:25,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:26,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:28,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:28,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:29,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:29,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:29,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:30,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:30,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:30,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:31,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:31,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:34,854 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:47:35,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:35,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:35,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:35,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:35,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:35,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:36,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:36,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:36,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:38,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:40,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:41,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:41,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:41,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:41,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:41,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:42,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:42,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:42,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:44,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:45,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:46,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:47,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:47,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:48,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:48,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:48,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:48,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:48,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:50,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:50,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:51,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:53,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:53,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:54,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:54,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:54,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:54,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:54,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:47:55,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:47:56,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:47:56,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:00,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:00,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:00,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:00,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:00,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:00,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:01,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:01,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:01,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:02,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:04,860 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:48:05,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:06,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:06,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:06,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:06,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:06,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:07,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:07,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:07,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:09,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:10,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:11,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:12,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:12,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:13,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:13,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:13,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:13,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:13,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:15,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:15,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:16,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:18,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:18,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:19,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:19,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:19,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:19,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:19,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:20,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:21,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:21,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:25,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:25,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:25,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:25,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:25,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:25,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:26,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:26,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:26,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:28,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:30,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:31,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:31,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:31,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:31,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:31,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:32,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:32,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:32,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:34,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:34,866 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:48:35,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:36,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:37,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:37,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:38,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:38,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:38,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:38,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:38,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:40,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:40,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:41,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:43,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:43,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:44,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:44,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:44,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:45,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:45,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:45,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:46,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:46,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:50,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:50,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:50,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:50,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:50,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:50,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:51,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:51,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:51,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:53,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:55,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:48:56,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:56,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:56,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:48:56,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:57,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:57,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:57,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:57,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:48:59,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:00,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:01,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:02,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:02,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:03,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:03,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:03,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:04,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:04,094 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:04,872 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:49:05,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:05,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:06,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:09,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:09,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:09,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:09,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:09,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:10,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:10,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:10,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:11,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:12,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:15,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:15,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:15,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:16,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:16,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:16,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:16,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:16,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:16,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:18,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:20,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:21,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:21,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:21,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:22,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:22,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:22,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:22,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:22,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:24,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:25,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:26,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:28,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:28,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:28,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:28,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:28,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:29,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:29,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:30,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:31,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:31,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:34,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:34,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:34,877 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:49:34,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:35,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:35,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:35,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:35,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:35,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:36,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:37,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:40,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:40,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:40,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:41,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:41,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:41,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:41,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:41,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:41,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:43,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:45,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:46,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:47,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:47,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:47,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:47,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:47,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:48,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:48,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:49,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:50,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:51,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:53,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:53,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:53,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:53,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:54,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:54,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:54,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:55,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:49:56,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:56,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:49:59,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:49:59,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:00,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:00,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:00,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:00,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:00,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:00,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:01,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:04,883 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:50:05,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:06,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:10,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:11,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:15,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:15,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:16,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:20,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:23,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:23,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:23,424 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:50:23,538 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:50:23,600 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:50:24,275 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:50:25,276 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:50:26,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:28,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:31,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:33,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:34,889 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:50:36,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:41,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:41,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:41,418 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:50:41,418 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:50:41,499 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:50:42,313 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:50:42,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:43,318 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:50:46,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:47,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:51,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:52,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:57,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:50:57,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:57,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:50:57,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:50:57,934 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:50:57,935 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:50:58,016 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:50:58,345 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:50:59,346 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:51:01,347 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:51:02,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:03,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:04,894 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:51:07,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:08,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:13,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:15,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:15,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:51:15,184 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:51:15,184 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:51:15,264 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:51:15,356 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:51:17,357 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:51:19,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:20,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:24,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:25,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:29,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:33,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:33,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:33,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:51:33,030 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:51:33,030 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:51:33,113 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:51:33,367 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:51:34,900 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:51:34,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:35,369 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:51:38,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:39,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:43,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:44,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:49,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:51:50,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:50,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:50,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:51:50,024 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:51:50,025 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:51:50,105 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:51:50,378 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:51:51,379 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:51:53,380 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:51:55,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:51:55,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:00,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:00,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:04,907 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:52:05,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:07,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:08,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:52:08,007 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:52:08,008 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:52:08,090 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:52:08,390 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:52:09,390 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:52:11,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:13,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:16,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:18,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:21,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:24,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:24,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:52:24,179 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:52:24,179 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:52:24,260 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:52:24,400 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:52:26,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:27,402 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:52:29,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:31,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:34,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:34,912 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:52:36,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:41,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:42,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:42,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:52:42,283 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:52:42,283 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:52:42,365 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:52:42,411 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:52:45,413 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:52:47,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:47,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:52,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:52,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:57,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:52:59,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:52:59,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:52:59,456 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:52:59,457 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:52:59,539 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:53:00,422 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:53:01,423 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:53:02,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:04,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:04,918 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:53:07,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:09,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:12,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:16,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:16,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:53:16,391 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:53:16,391 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:53:16,471 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:53:17,471 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:53:18,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:19,473 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:53:21,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:23,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:26,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:28,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:31,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:33,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:34,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:53:34,760 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:53:34,761 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:53:34,842 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:53:34,923 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:53:35,483 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:53:36,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:36,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:37,484 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:53:38,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:41,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:43,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:48,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:48,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:51,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:53:51,638 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:53:51,638 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:53:51,720 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:53:52,494 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:53:53,495 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:53:53,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:54,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:53:58,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:53:59,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:03,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:04,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:04,930 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:54:09,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:09,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:09,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:54:09,565 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:54:09,565 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:54:09,647 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:54:10,505 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:54:10,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:11,506 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:54:14,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:15,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:19,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:20,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:25,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:26,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:26,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:26,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:54:26,106 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:54:26,107 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:54:26,188 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:54:26,515 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:54:27,516 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:54:31,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:31,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:34,936 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:54:36,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:36,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:41,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:43,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:43,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:54:43,851 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:54:43,851 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:54:43,936 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:54:44,526 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:54:45,527 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:54:47,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:48,844 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:52,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:54:53,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:54:57,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:01,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:55:01,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:55:01,370 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:55:01,371 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:55:01,451 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:55:01,537 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:55:03,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:03,538 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:55:04,941 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:55:06,363 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:55:08,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:11,363 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:55:13,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:18,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:55:18,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:55:18,699 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:55:18,700 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:55:18,780 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:55:19,549 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:55:19,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:21,551 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:55:23,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:55:24,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:29,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:31,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:55:34,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:34,947 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:55:36,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:55:39,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:45,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:45,566 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:55:49,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:55:50,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:55:55,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:00,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:02,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:02,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:04,953 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:56:05,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:09,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:10,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:15,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:15,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:20,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:21,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:21,588 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:56:25,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:26,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:30,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:31,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:34,959 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:56:36,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:37,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:41,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:56:41,039 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:56:41,039 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:56:41,124 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:56:41,601 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:56:42,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:43,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:43,603 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:56:47,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:48,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:52,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:53,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:57,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:56:58,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:56:58,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:56:58,733 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:56:58,734 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:56:58,815 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:56:59,614 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:57:01,615 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:57:02,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:03,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:04,965 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:57:07,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:08,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:12,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:15,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:15,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:57:15,660 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:57:15,661 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:57:15,741 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:57:16,626 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:57:17,626 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:57:18,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:20,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:23,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:25,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:28,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:31,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:33,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:33,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:57:33,916 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:57:33,917 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:57:33,997 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:57:34,638 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:57:34,970 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:57:35,639 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:57:36,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:38,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:41,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:43,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:47,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:48,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:50,552 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:57:50,555 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:57:50,555 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:57:50,637 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:57:50,649 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:57:51,650 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:57:52,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:52,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:54,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:57:57,550 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:57:59,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:02,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:04,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:04,976 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:58:08,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:08,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:08,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:58:08,348 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:58:08,349 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:58:08,430 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:58:08,664 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:58:09,664 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:58:10,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:13,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:15,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:18,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:20,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:24,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:24,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:25,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:27,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:58:27,051 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:58:27,052 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:58:27,133 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:58:27,677 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:58:29,678 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:58:30,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:31,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:34,982 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:58:35,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:36,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:41,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:41,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:44,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:58:44,207 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:58:44,210 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:58:44,292 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:58:44,689 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:58:45,689 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:58:47,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:47,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:52,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:58:52,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:57,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:58:58,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:02,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:59:02,026 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:59:02,027 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:59:02,107 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:59:02,701 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:59:03,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:03,702 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:59:04,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:04,988 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:59:08,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:09,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:13,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:15,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:18,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:59:18,560 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:59:18,562 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:59:18,647 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:59:18,711 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:59:19,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:19,712 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:59:20,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:24,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:25,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:29,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:30,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:34,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:34,993 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 15:59:36,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:36,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:36,456 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:59:36,459 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:59:36,460 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:59:36,540 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:59:36,723 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:59:37,724 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:59:40,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:42,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:45,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:47,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:50,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:53,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 15:59:53,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 15:59:53,564 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 15:59:53,565 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 15:59:53,644 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 15:59:53,734 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 15:59:55,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 15:59:55,739 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 15:59:58,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:00,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:03,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:04,997 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:00:05,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:10,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:10,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:11,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:00:11,005 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:00:11,006 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:00:11,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:11,098 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:00:11,749 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:00:13,751 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:00:15,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:16,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:20,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:21,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:26,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:28,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:28,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:28,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:00:28,834 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:00:28,835 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:00:28,928 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:00:29,761 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:00:29,762 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:00:31,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:34,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:35,003 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:00:37,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:39,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:42,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:45,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:45,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:45,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:00:45,832 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:00:45,833 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:00:45,926 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:00:46,772 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:00:47,773 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:00:47,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:50,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:52,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:00:55,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:00:57,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:02,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:03,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:03,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:01:03,361 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:01:03,362 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:01:03,456 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:01:03,783 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:01:05,009 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:01:05,785 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:01:08,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:08,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:13,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:13,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:18,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:20,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:20,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:01:20,122 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:01:20,123 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:01:20,214 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:01:20,794 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:01:21,795 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:01:23,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:25,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:28,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:30,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:33,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:35,015 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:01:37,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:37,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:01:37,857 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:01:37,857 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:01:37,948 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:01:38,806 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:01:38,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:39,807 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:01:42,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:43,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:47,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:48,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:53,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:01:55,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:01:55,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:01:55,602 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:01:55,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:01:55,683 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:01:55,818 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:01:57,819 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:01:59,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:00,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:04,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:05,021 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:02:05,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:10,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:12,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:12,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:12,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:02:12,889 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:02:12,890 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:02:12,970 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:02:13,830 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:02:13,830 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:02:15,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:17,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:20,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:22,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:25,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:30,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:30,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:30,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:02:30,272 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:02:30,274 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:02:30,364 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:02:30,841 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:02:31,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:31,841 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:02:35,026 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:02:35,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:37,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:40,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:42,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:47,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:47,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:47,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:47,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:02:47,037 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:02:47,039 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:02:47,132 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:02:47,852 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:02:47,852 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:02:52,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:53,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:02:57,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:02:58,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:02,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:05,032 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:03:05,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:05,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:05,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:03:05,168 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:03:05,168 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:03:05,263 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:03:05,864 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:03:05,864 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:03:07,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:07,865 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:03:10,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:12,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:15,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:17,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:21,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:21,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:21,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:03:21,920 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:03:21,922 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:03:21,959 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:03:22,875 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:03:22,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:23,876 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:03:27,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:27,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:32,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:32,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:35,039 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:03:38,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:39,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:39,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:03:39,068 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:03:39,068 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:03:39,107 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:03:39,883 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:03:39,883 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:03:43,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:44,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:48,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:49,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:53,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:03:57,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:03:57,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:03:57,058 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:03:57,058 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:03:57,097 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:03:57,892 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:03:57,892 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:03:59,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:02,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:04,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:05,044 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:04:07,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:10,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:13,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:13,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:04:13,653 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:04:13,654 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:04:13,691 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:04:13,897 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:04:15,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:15,897 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:04:18,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:20,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:23,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:25,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:30,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:31,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:31,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:04:31,752 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:04:31,753 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:04:31,791 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:04:31,904 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:04:33,905 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:04:35,050 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:04:36,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:36,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:41,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:41,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:46,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:48,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:48,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:48,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:04:48,238 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:04:48,239 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:04:48,276 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:04:48,912 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:04:49,912 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:04:51,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:53,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:04:56,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:04:58,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:01,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:05,056 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:05:06,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:06,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:06,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:05:06,639 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:05:06,640 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:05:06,677 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:05:06,919 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:05:07,920 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:05:11,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:11,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:16,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:16,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:21,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:23,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:23,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:05:23,394 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:05:23,396 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:05:23,488 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:05:23,926 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:05:23,927 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:05:25,927 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:05:27,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:28,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:32,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:33,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:35,061 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:05:38,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:41,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:41,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:05:41,020 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:05:41,022 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:05:41,102 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:05:41,934 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:05:41,935 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:05:43,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:46,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:48,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:51,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:53,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:58,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:05:58,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:05:58,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:05:58,593 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:05:58,593 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:05:58,673 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:05:58,941 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:06:00,942 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:06:03,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:03,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:05,067 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:06:08,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:09,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:14,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:15,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:15,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:06:15,837 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:06:15,838 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:06:15,928 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:06:15,949 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:06:18,951 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:06:19,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:20,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:24,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:25,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:29,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:31,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:33,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:06:33,729 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:06:33,729 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:06:33,814 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:06:33,957 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:06:35,073 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:06:35,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:36,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:36,959 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:06:40,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:41,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:45,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:46,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:50,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:50,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:06:50,690 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:06:50,691 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:06:50,771 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:06:50,968 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:06:52,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:06:52,970 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:06:55,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:06:57,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:00,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:02,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:05,079 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:07:06,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:08,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:08,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:08,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:07:08,683 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:07:08,684 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:07:08,764 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:07:08,980 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:07:10,982 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:07:11,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:13,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:16,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:18,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:21,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:26,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:26,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:07:26,247 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:07:26,248 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:07:26,329 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:07:26,992 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:07:27,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:28,994 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:07:31,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:32,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:35,085 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:07:36,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:38,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:43,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:43,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:43,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:07:43,333 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:07:43,333 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:07:43,415 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:07:44,003 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:07:45,004 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:07:48,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:48,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:53,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:07:53,417 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:58,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:07:58,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:01,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:08:01,274 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:08:01,275 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:08:01,356 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:08:02,018 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:08:03,019 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:08:04,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:04,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:05,091 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:08:09,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:10,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:14,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:15,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:17,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:08:17,798 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:08:17,799 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:08:17,880 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:08:18,029 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:08:19,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:20,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:21,031 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:08:24,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:25,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:29,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:30,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:35,096 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:08:36,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:36,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:36,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:36,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:08:36,418 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:08:36,419 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:08:36,499 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:08:37,042 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:08:39,043 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:08:41,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:42,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:46,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:47,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:51,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:52,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:52,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:08:53,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:08:53,006 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:08:53,008 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:08:53,089 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:08:54,089 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:08:55,090 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:08:57,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:08:58,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:02,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:03,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:05,102 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:09:07,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:11,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:11,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:11,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:09:11,129 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:09:11,130 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:09:11,212 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:09:12,107 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:09:12,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:13,107 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:09:16,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:17,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:21,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:22,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:27,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:28,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:28,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:09:28,631 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:09:28,632 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:09:28,712 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:09:29,118 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:09:31,119 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:09:32,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:33,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:35,108 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:09:38,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:38,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:43,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:45,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:45,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:09:45,715 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:09:45,716 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:09:45,796 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:09:46,125 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:09:47,125 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:09:48,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:50,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:53,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:09:55,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:09:58,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:03,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:03,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:10:03,391 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:10:03,392 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:10:03,474 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:10:04,133 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:10:04,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:05,113 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:10:05,133 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:10:08,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:10,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:13,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:15,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:20,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:20,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:20,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:10:20,209 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:10:20,210 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:10:20,296 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:10:21,141 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:10:23,142 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:10:25,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:25,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:30,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:30,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:35,120 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:10:36,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:38,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:38,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:38,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:10:38,092 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:10:38,093 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:10:38,175 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:10:39,174 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:10:41,175 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:10:41,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:43,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:46,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:48,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:51,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:55,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:10:55,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:10:55,311 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:10:55,313 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:10:55,395 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:10:56,182 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:10:56,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:10:57,183 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:11:00,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:01,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:05,125 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:11:05,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:07,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:12,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:12,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:12,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:11:12,785 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:11:12,786 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:11:12,868 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:11:13,190 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:11:15,191 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:11:17,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:17,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:22,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:22,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:27,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:30,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:30,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:30,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:11:30,294 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:11:30,296 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:11:30,376 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:11:31,198 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:11:33,198 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:11:33,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:35,131 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:11:35,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:39,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:40,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:44,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:46,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:46,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:11:46,818 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:11:46,819 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:11:46,895 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:11:47,204 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:11:49,205 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:11:49,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:51,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:54,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:11:56,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:11:59,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:02,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:04,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:05,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:12:05,045 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:12:05,046 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:12:05,127 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:12:05,137 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:12:05,212 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:12:07,213 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:12:08,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:10,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:13,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:15,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:18,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:20,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:21,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:12:21,648 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:12:21,649 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:12:21,730 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:12:22,219 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:12:23,219 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:12:23,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:25,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:28,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:30,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:33,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:35,144 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:12:36,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:39,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:39,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:12:39,645 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:12:39,646 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:12:39,727 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:12:40,224 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:12:41,225 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:12:41,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:44,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:46,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:49,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:51,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:54,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:56,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:12:56,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:12:56,992 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:12:56,992 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:12:57,072 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:12:57,230 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:12:59,231 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:12:59,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:12:59,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:02,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:04,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:05,149 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:13:07,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:10,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:12,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:14,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:13:14,507 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:13:14,508 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:13:14,587 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:13:15,237 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:13:16,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:17,238 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:13:17,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:21,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:22,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:26,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:27,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:31,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:31,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:31,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:13:31,940 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:13:31,941 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:13:32,020 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:13:32,244 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:13:33,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:33,244 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:13:35,155 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:13:37,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:38,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:42,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:43,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:48,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:48,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:48,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:48,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:13:48,737 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:13:48,738 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:13:48,816 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:13:49,252 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:13:51,253 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:13:53,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:54,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:13:58,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:13:59,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:03,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:05,160 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:14:06,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:06,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:06,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:14:06,704 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:14:06,706 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:14:06,789 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:14:07,259 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:14:09,259 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:14:09,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:12,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:14,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:17,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:19,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:23,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:23,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:23,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:14:23,381 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:14:23,382 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:14:23,462 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:14:24,266 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:14:25,266 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:14:25,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:28,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:30,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:33,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:35,167 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:14:36,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:41,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:41,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:14:41,156 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:14:41,156 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:14:41,236 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:14:41,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:41,273 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:14:43,274 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:14:46,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:46,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:51,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:51,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:56,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:14:58,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:14:58,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:14:58,652 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:14:58,653 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:14:58,732 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:14:59,281 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:14:59,281 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:15:01,282 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:15:01,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:03,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:05,172 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:15:07,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:08,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:12,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:15,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:15,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:15:15,664 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:15:15,665 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:15:15,744 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:15:16,286 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:15:17,287 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:15:17,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:20,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:22,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:25,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:27,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:32,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:33,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:33,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:15:33,182 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:15:33,183 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:15:33,263 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:15:33,293 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:15:35,178 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:15:35,293 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:15:38,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:38,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:43,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:43,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:48,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:49,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:49,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:15:49,634 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:15:49,635 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:15:49,715 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:15:50,298 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:15:51,299 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:15:53,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:54,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:15:58,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:15:59,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:03,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:05,184 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:16:07,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:07,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:07,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:16:07,979 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:16:07,979 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:16:08,061 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:16:08,304 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:16:09,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:09,305 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:16:12,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:14,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:17,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:19,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:24,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:24,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:24,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:24,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:16:24,945 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:16:24,946 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:16:25,028 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:16:25,310 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:16:27,311 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:16:30,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:30,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:35,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:35,190 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:16:35,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:40,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:42,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:42,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:42,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:16:42,441 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:16:42,442 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:16:42,523 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:16:43,317 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:16:43,318 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:16:45,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:47,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:50,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:52,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:55,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:16:59,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:59,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:16:59,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:16:59,983 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:16:59,984 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:17:00,064 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:17:00,324 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:17:01,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:01,325 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:17:05,196 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:17:05,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:06,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:10,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:11,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:16,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:16,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:16,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:16,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:17:16,286 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:17:16,287 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:17:16,367 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:17:16,368 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:17:17,368 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:17:21,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:21,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:26,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:26,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:31,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:34,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:34,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:34,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:17:34,150 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:17:34,151 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:17:34,232 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:17:34,376 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:17:35,201 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:17:35,377 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:17:37,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:40,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:42,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:45,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:47,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:50,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:50,844 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:50,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:17:50,858 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:17:50,859 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:17:50,940 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:17:51,382 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:17:51,383 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:17:52,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:17:53,383 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:17:56,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:17:57,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:01,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:02,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:05,207 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:18:08,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:08,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:08,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:08,989 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:18:08,993 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:18:08,993 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:18:09,074 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:18:09,389 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:18:11,390 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:18:13,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:14,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:18,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:19,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:24,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:26,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:26,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:18:26,661 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:18:26,661 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:18:26,742 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:18:27,395 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:18:27,396 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:18:29,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:31,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:34,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:35,213 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:18:36,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:40,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:43,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:43,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:18:43,818 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:18:43,819 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:18:43,899 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:18:44,400 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:18:45,401 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:18:45,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:48,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:50,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:18:53,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:18:55,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:00,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:01,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:01,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:19:01,080 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:19:01,081 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:19:01,160 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:19:01,406 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:19:03,406 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:19:05,219 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:19:06,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:06,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:11,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:11,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:16,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:17,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:17,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:19:17,554 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:19:17,554 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:19:17,635 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:19:18,411 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:19:19,411 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:19:21,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:22,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:26,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:27,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:31,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:35,226 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:19:35,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:35,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:19:35,520 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:19:35,521 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:19:35,602 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:19:36,416 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:19:37,416 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:19:37,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:40,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:42,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:45,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:47,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:52,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:19:52,626 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:52,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:52,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:19:52,634 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:19:52,635 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:19:52,715 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:19:53,421 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:19:53,421 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:19:57,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:19:57,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:02,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:02,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:05,232 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:20:08,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:10,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:10,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:10,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:10,105 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:20:10,106 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:20:10,186 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:20:10,426 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:20:11,427 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:20:14,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:15,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:19,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:20,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:24,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:25,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:26,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:27,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:27,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:27,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:28,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:29,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:30,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:31,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:33,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:33,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:33,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:33,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:33,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:33,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:34,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:34,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:35,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:35,238 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:20:37,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:39,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:39,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:39,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:39,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:39,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:39,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:40,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:40,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:40,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:43,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:44,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:45,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:45,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:45,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:45,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:46,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:46,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:46,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:46,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:49,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:50,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:50,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:51,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:51,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:51,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:52,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:52,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:52,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:52,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:54,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:20:55,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:20:56,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:58,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:58,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:58,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:58,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:58,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:58,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:59,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:20:59,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:00,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:02,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:04,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:04,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:04,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:04,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:04,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:04,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:04,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:05,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:05,243 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:21:05,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:08,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:09,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:10,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:10,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:10,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:10,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:10,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:11,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:11,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:11,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:14,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:15,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:15,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:16,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:16,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:16,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:17,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:17,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:17,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:17,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:19,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:20,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:21,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:22,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:22,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:22,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:23,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:23,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:23,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:24,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:24,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:25,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:27,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:29,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:29,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:29,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:29,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:29,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:29,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:29,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:30,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:30,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:33,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:34,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:35,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:35,249 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:21:35,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:35,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:35,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:35,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:35,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:36,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:36,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:39,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:39,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:40,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:41,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:41,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:41,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:42,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:42,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:42,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:42,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:44,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:45,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:46,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:47,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:47,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:47,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:48,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:48,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:48,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:48,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:49,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:50,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:52,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:54,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:54,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:54,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:54,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:21:54,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:54,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:54,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:55,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:21:55,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:58,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:21:59,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:00,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:00,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:00,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:00,375 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:00,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:00,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:01,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:01,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:04,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:04,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:05,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:05,255 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:22:06,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:06,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:06,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:07,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:07,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:07,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:07,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:09,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:10,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:11,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:12,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:12,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:12,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:13,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:13,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:13,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:13,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:14,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:15,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:17,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:19,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:19,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:19,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:19,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:19,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:19,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:19,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:20,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:20,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:23,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:24,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:25,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:25,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:25,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:25,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:25,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:25,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:26,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:26,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:29,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:29,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:30,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:31,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:31,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:31,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:32,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:32,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:32,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:32,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:34,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:35,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:35,260 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:22:36,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:37,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:37,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:37,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:38,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:38,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:38,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:38,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:39,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:40,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:42,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:44,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:44,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:44,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:44,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:44,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:44,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:44,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:45,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:45,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:48,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:49,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:50,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:50,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:50,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:50,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:50,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:50,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:51,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:51,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:54,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:22:54,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:55,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:22:56,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:56,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:56,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:57,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:57,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:57,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:57,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:22:59,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:00,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:01,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:02,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:02,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:02,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:03,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:03,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:03,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:03,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:04,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:05,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:05,266 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:23:07,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:09,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:09,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:09,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:09,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:09,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:09,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:09,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:10,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:10,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:13,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:14,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:15,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:15,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:15,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:15,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:15,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:15,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:15,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:16,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:19,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:19,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:20,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:21,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:21,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:21,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:22,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:22,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:22,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:22,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:24,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:25,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:26,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:27,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:27,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:27,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:28,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:28,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:28,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:28,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:29,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:30,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:32,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:34,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:34,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:34,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:34,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:34,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:34,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:34,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:35,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:35,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:35,272 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:23:38,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:40,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:40,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:40,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:40,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:40,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:40,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:40,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:40,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:41,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:44,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:45,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:45,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:46,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:46,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:46,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:46,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:23:50,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:50,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:23:55,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:23:55,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:00,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:00,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:02,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:05,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:05,278 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:24:05,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:10,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:10,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:15,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:15,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:17,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:17,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:19,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:20,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:20,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:22,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:22,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:23,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:23,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:23,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:23,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:23,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:25,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:25,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:25,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:28,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:28,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:29,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:29,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:29,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:29,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:29,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:30,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:30,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:31,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:35,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:35,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:35,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:35,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:35,284 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:24:35,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:35,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:35,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:36,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:36,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:38,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:40,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:40,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:41,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:41,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:41,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:41,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:42,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:42,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:42,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:44,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:45,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:45,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:47,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:47,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:48,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:48,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:48,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:48,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:48,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:50,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:50,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:50,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:53,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:53,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:54,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:54,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:54,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:55,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:55,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:24:55,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:24:55,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:24:56,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:00,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:00,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:00,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:00,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:00,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:00,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:00,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:01,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:01,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:03,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:05,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:05,290 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:25:05,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:06,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:06,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:06,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:07,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:07,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:07,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:07,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:09,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:10,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:10,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:12,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:12,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:13,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:13,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:13,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:13,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:13,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:15,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:15,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:15,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:19,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:19,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:19,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:19,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:19,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:20,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:20,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:20,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:20,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:21,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:25,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:25,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:25,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:25,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:25,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:25,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:26,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:26,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:26,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:28,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:30,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:30,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:31,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:31,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:31,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:32,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:32,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:32,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:32,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:34,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:35,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:35,295 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:25:35,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:37,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:37,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:38,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:38,375 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:38,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:38,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:38,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:40,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:40,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:40,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:44,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:44,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:44,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:44,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:44,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:45,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:45,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:45,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:45,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:46,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:50,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:50,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:50,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:50,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:50,786 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:50,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:50,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:51,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:51,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:53,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:55,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:25:55,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:25:56,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:56,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:57,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:57,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:57,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:57,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:57,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:25:59,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:00,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:00,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:02,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:02,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:03,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:03,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:03,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:03,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:03,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:05,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:05,301 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:26:05,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:05,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:09,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:09,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:09,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:09,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:09,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:10,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:10,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:10,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:10,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:11,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:15,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:15,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:15,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:15,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:15,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:15,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:16,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:16,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:16,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:18,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:20,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:20,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:21,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:21,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:22,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:22,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:22,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:22,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:22,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:24,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:25,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:25,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:27,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:27,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:28,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:28,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:28,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:29,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:29,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:30,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:30,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:30,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:34,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:34,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:34,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:34,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:34,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:35,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:35,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:35,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:35,307 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:26:35,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:36,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:40,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:40,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:40,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:40,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:40,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:40,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:41,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:41,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:41,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:43,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:45,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:45,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:46,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:46,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:47,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:47,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:47,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:47,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:47,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:49,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:50,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:50,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:52,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:52,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:53,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:53,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:53,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:53,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:53,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:55,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:26:55,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:26:55,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:59,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:59,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:59,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:59,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:26:59,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:00,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:00,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:00,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:00,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:01,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:05,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:05,312 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:27:05,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:05,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:05,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:05,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:05,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:06,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:06,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:06,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:08,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:10,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:10,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:11,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:11,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:12,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:12,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:12,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:12,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:12,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:14,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:15,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:15,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:17,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:17,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:18,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:18,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:18,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:18,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:18,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:20,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:20,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:20,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:24,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:24,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:24,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:24,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:24,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:25,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:25,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:25,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:25,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:26,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:30,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:30,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:30,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:30,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:30,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:30,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:31,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:31,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:31,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:33,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:35,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:35,318 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:27:35,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:36,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:36,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:37,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:37,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:37,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:37,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:37,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:40,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:40,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:45,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:45,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:50,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:50,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:52,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:27:55,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:27:56,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:56,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:27:59,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:28:00,137 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:28:00,238 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:28:00,305 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:28:00,759 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:28:00,760 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:28:01,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:01,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:02,760 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:28:05,324 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:28:06,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:06,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:11,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:13,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:16,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:16,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:28:16,519 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:28:16,519 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:28:16,601 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:28:16,769 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:28:18,770 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:28:19,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:21,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:24,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:26,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:31,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:33,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:33,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:28:33,417 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:28:33,419 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:28:33,499 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:28:33,780 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:28:34,781 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:28:35,330 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:28:37,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:38,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:42,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:43,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:47,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:48,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:50,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:28:50,739 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:28:50,739 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:28:50,824 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:28:51,824 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:28:52,824 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:28:52,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:53,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:28:57,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:28:58,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:02,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:03,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:05,336 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:29:07,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:29:07,811 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:29:07,811 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:29:07,892 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:29:07,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:08,892 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:29:09,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:10,893 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:29:12,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:14,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:17,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:19,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:22,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:26,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:26,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:26,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:29:26,337 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:29:26,338 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:29:26,419 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:29:26,904 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:29:28,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:28,905 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:29:31,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:33,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:35,341 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:29:36,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:39,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:43,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:43,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:29:43,008 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:29:43,010 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:29:43,090 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:29:43,915 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:29:44,916 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:29:45,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:48,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:50,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:29:53,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:29:55,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:00,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:00,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:00,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:30:00,538 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:30:00,538 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:30:00,618 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:30:00,926 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:30:02,927 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:30:05,347 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:30:05,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:05,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:10,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:10,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:15,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:17,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:17,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:30:17,362 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:30:17,363 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:30:17,443 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:30:17,937 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:30:18,938 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:30:20,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:22,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:25,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:27,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:30,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:34,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:34,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:30:34,340 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:30:34,340 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:30:34,420 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:30:34,948 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:30:35,354 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:30:36,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:36,949 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:30:39,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:41,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:44,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:46,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:51,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:52,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:30:52,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:30:52,391 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:30:52,392 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:30:52,471 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:30:52,959 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:30:54,961 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:30:56,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:30:57,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:01,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:02,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:05,359 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:31:07,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:09,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:09,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:31:09,062 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:31:09,062 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:31:09,143 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:31:09,970 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:31:10,971 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:31:13,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:14,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:18,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:19,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:23,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:27,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:27,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:27,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:31:27,596 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:31:27,598 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:31:27,678 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:31:27,981 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:31:28,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:28,982 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:31:32,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:33,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:35,364 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:31:37,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:39,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:44,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:44,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:31:44,367 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:31:44,368 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:31:44,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:44,448 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:31:44,992 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:31:46,994 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:31:49,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:49,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:54,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:31:54,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:31:59,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:02,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:02,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:32:02,109 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:32:02,109 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:32:02,191 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:32:03,004 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:32:05,006 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:32:05,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:05,370 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:32:07,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:10,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:12,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:15,375 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:18,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:18,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:32:18,815 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:32:18,815 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:32:18,896 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:32:19,015 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:32:20,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:21,016 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:32:23,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:25,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:28,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:30,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:35,376 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:32:35,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:35,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:32:35,493 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:32:35,493 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:32:35,573 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:32:36,025 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:32:36,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:37,026 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:32:40,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:41,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:45,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:46,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:51,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:54,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:54,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:32:54,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:32:54,080 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:32:54,081 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:32:54,161 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:32:55,037 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:32:55,038 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:32:57,039 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:32:57,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:32:59,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:02,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:04,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:05,382 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:33:07,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:10,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:10,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:10,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:33:10,616 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:33:10,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:33:10,696 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:33:11,048 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:33:12,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:13,049 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:33:15,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:17,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:20,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:22,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:27,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:29,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:29,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:29,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:33:29,209 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:33:29,209 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:33:29,288 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:33:30,060 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:33:31,060 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:33:33,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:34,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:35,388 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:33:38,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:39,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:43,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:46,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:46,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:33:46,125 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:33:46,126 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:33:46,206 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:33:47,071 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:33:47,071 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:33:49,072 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:33:49,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:51,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:54,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:33:56,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:33:59,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:03,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:03,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:34:03,341 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:34:03,341 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:34:03,421 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:34:04,082 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:34:04,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:05,082 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:34:05,394 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:34:08,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:10,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:13,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:15,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:20,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:20,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:34:20,091 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:34:20,091 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:34:20,175 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:34:21,174 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:34:21,175 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:34:21,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:25,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:26,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:30,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:31,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:35,400 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:34:36,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:36,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:36,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:34:36,647 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:34:36,648 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:34:36,730 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:34:37,185 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:34:39,186 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:34:41,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:41,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:46,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:46,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:51,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:51,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:55,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:34:55,742 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:34:55,743 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:34:55,826 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:34:56,197 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:34:56,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:34:57,198 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:34:57,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:34:57,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:01,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:02,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:05,406 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:35:07,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:09,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:12,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:12,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:35:12,712 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:35:12,713 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:35:12,794 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:35:13,208 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:35:14,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:15,209 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:35:17,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:19,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:22,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:24,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:27,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:30,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:30,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:30,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:35:30,922 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:35:30,923 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:35:31,004 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:35:31,219 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:35:33,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:33,221 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:35:35,411 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:35:35,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:38,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:40,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:43,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:48,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:48,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:35:48,113 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:35:48,114 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:35:48,198 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:35:48,230 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:35:49,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:49,231 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:35:53,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:54,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:35:58,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:35:59,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:04,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:04,832 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:04,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:36:04,840 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:36:04,840 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:36:04,923 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:36:05,242 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:36:05,417 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:36:07,245 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:36:09,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:09,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:14,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:14,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:19,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:21,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:21,598 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:36:21,602 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:36:21,602 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:36:21,682 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:36:22,283 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:36:23,284 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:36:24,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:26,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:29,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:31,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:34,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:35,423 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:36:38,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:38,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:36:38,391 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:36:38,391 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:36:38,469 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:36:39,309 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:36:39,310 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:36:40,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:43,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:45,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:48,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:50,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:53,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:55,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:36:57,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:36:57,166 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:36:57,167 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:36:57,247 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:36:57,321 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:36:59,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:36:59,322 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:37:01,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:04,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:05,429 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:37:06,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:09,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:11,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:14,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:14,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:14,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:37:14,424 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:37:14,425 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:37:14,505 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:37:15,332 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:37:15,333 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:37:16,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:19,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:21,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:24,417 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:26,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:30,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:31,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:32,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:37:32,405 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:37:32,406 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:37:32,492 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:37:33,344 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:37:33,344 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:37:35,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:35,435 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:37:37,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:40,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:42,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:45,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:45,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:47,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:49,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:37:49,756 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:37:49,757 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:37:49,838 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:37:50,355 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:37:51,355 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:37:51,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:52,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:37:56,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:37:57,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:02,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:03,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:03,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:05,440 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:38:06,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:38:06,681 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:38:06,682 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:38:06,763 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:38:07,365 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:38:08,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:08,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:09,367 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:38:13,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:13,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:18,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:19,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:23,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:38:23,332 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:38:23,333 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:38:23,414 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:38:24,414 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:38:24,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:25,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:25,414 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:38:29,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:30,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:34,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:35,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:35,444 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:38:39,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:40,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:40,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:38:40,522 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:38:40,523 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:38:40,605 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:38:41,425 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:38:41,425 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:38:44,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:45,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:49,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:50,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:54,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:38:55,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:38:59,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:38:59,132 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:38:59,133 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:38:59,216 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:38:59,436 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:39:00,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:01,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:01,438 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:39:05,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:05,450 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:39:06,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:10,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:12,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:15,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:16,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:39:16,108 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:39:16,109 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:39:16,190 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:39:16,447 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:39:17,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:17,448 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:39:21,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:22,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:26,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:27,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:31,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:33,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:33,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:33,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:39:33,949 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:39:33,951 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:39:34,032 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:39:34,459 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:39:35,456 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:39:35,472 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:39:36,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:38,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:41,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:43,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:46,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:50,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:50,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:50,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:39:50,689 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:39:50,691 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:39:50,772 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:39:51,482 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:39:51,483 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:39:51,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:39:55,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:39:56,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:00,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:01,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:05,462 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:40:07,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:08,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:08,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:40:08,049 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:40:08,051 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:40:08,131 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:40:08,493 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:40:09,494 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:40:13,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:13,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:18,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:18,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:23,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:24,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:24,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:40:24,780 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:40:24,781 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:40:24,862 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:40:25,504 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:40:25,505 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:40:28,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:29,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:33,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:34,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:35,468 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:40:39,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:42,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:42,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:40:42,187 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:40:42,188 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:40:42,268 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:40:42,515 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:40:43,516 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:40:45,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:47,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:50,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:52,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:40:55,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:40:57,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:00,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:00,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:41:00,707 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:41:00,707 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:41:00,787 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:41:01,527 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:41:01,527 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:41:02,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:05,473 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:41:05,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:07,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:10,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:13,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:15,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:17,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:41:17,432 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:41:17,434 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:41:17,516 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:41:17,537 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:41:19,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:19,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:19,539 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:41:20,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:24,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:25,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:29,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:30,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:35,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:35,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:35,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:41:35,242 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:41:35,243 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:41:35,325 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:41:35,479 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:41:35,549 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:41:36,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:37,550 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:41:40,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:41,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:45,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:46,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:51,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:41:51,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:51,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:51,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:41:51,616 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:41:51,617 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:41:51,697 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:41:52,560 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:41:53,560 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:41:56,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:41:56,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:01,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:01,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:05,485 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:42:07,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:09,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:09,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:42:09,340 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:42:09,342 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:42:09,424 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:42:09,571 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:42:11,572 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:42:13,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:14,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:18,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:19,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:23,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:26,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:26,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:42:26,697 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:42:26,698 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:42:26,779 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:42:27,582 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:42:27,583 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:42:28,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:31,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:33,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:35,491 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:42:36,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:39,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:43,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:43,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:42:43,869 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:42:43,870 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:42:43,951 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:42:44,593 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:42:44,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:45,594 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:42:48,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:49,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:53,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:42:54,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:42:59,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:02,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:02,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:43:02,255 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:43:02,256 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:43:02,336 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:43:02,604 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:43:03,605 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:43:05,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:05,497 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:43:07,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:10,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:12,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:15,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:18,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:18,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:18,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:43:18,610 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:43:18,610 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:43:18,693 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:43:19,693 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:43:19,694 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:43:20,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:23,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:25,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:28,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:30,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:35,502 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:43:36,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:36,832 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:36,832 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:36,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:43:36,840 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:43:36,841 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:43:36,923 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:43:37,705 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:43:37,705 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:43:41,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:41,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:46,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:46,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:51,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:53,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:53,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:43:53,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:43:53,184 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:43:53,185 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:43:53,265 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:43:53,715 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:43:55,717 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:43:57,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:43:58,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:02,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:03,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:05,509 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:44:07,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:10,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:10,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:44:10,725 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:44:10,725 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:44:10,807 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:44:11,808 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:44:11,808 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:44:12,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:15,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:17,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:20,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:22,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:27,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:28,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:28,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:44:28,159 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:44:28,161 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:44:28,243 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:44:28,818 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:44:29,819 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:44:33,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:33,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:35,514 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:44:38,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:38,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:43,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:44,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:44,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:44:44,868 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:44:44,868 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:44:44,948 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:44:45,829 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:44:45,830 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:44:48,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:49,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:53,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:44:54,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:44:58,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:03,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:03,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:45:03,583 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:45:03,584 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:45:03,668 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:45:03,841 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:45:04,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:05,520 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:45:05,842 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:45:08,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:10,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:13,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:15,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:19,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:19,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:45:19,834 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:45:19,835 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:45:19,915 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:45:20,915 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:45:20,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:21,916 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:45:24,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:25,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:29,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:30,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:34,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:35,526 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:45:36,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:38,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:45:38,070 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:45:38,070 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:45:38,150 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:45:38,927 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:45:39,928 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:45:40,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:42,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:45,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:47,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:50,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:50,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:52,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:45:54,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:45:54,594 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:45:54,595 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:45:54,676 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:45:54,938 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:45:55,938 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:45:56,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:45:57,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:01,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:02,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:05,532 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:46:06,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:08,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:11,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:11,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:46:11,773 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:46:11,773 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:46:11,861 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:46:11,948 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:46:13,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:13,950 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:46:16,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:18,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:21,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:23,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:26,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:28,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:29,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:46:29,998 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:46:29,999 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:46:30,080 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:46:30,961 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:46:31,961 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:46:31,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:34,081 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:35,537 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:46:36,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:39,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:43,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:44,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:46,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:46:46,457 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:46:46,457 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:46:46,538 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:46:46,971 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:46:47,972 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:46:48,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:50,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:53,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:46:55,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:46:58,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:00,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:05,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:05,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:47:05,261 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:47:05,262 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:47:05,342 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:47:05,543 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:47:05,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:05,983 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:47:05,983 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:47:10,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:10,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:15,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:15,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:20,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:21,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:21,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:21,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:47:21,417 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:47:21,417 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:47:21,500 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:47:21,993 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:47:21,994 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:47:26,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:26,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:31,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:31,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:35,549 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:47:36,552 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:39,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:39,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:39,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:47:39,119 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:47:39,119 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:47:39,200 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:47:40,005 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:47:40,006 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:47:42,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:44,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:47,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:49,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:52,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:56,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:56,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:47:56,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:47:56,271 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:47:56,271 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:47:56,352 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:47:57,030 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:47:57,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:47:58,031 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:48:01,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:02,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:05,555 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:48:06,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:07,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:12,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:13,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:13,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:48:13,201 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:48:13,201 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:48:13,287 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:48:14,059 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:48:14,060 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:48:18,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:18,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:23,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:23,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:28,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:31,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:31,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:48:31,128 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:48:31,129 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:48:31,211 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:48:32,088 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:48:32,089 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:48:34,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:35,561 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:48:36,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:39,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:41,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:44,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:47,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:47,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:48:47,544 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:48:47,544 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:48:47,626 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:48:48,114 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:48:48,115 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:48:49,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:52,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:54,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:48:57,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:48:59,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:02,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:04,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:05,567 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:49:06,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:49:06,204 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:49:06,205 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:49:06,289 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:49:07,143 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:49:08,144 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:49:08,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:10,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:13,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:15,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:18,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:20,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:22,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:49:22,938 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:49:22,939 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:49:23,020 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:49:23,169 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:49:23,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:23,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:24,172 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:49:26,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:28,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:31,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:33,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:35,573 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:49:36,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:40,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:40,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:40,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:49:40,406 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:49:40,407 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:49:40,488 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:49:41,198 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:49:42,200 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:49:42,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:45,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:47,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:50,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:52,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:57,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:49:57,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:57,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:49:57,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:49:57,702 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:49:57,704 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:49:57,786 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:49:58,225 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:49:58,226 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:50:02,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:02,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:05,579 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:50:07,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:08,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:13,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:14,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:14,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:50:14,281 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:50:14,282 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:50:14,364 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:50:15,251 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:50:16,252 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:50:19,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:19,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:24,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:24,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:29,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:32,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:32,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:50:32,361 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:50:32,361 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:50:32,444 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:50:33,278 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:50:34,279 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:50:34,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:35,585 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:50:37,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:39,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:42,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:44,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:49,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:49,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:50:49,296 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:50:49,296 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:50:49,379 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:50:49,379 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:50:50,379 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:50:50,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:54,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:50:55,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:50:59,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:00,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:04,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:05,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:05,590 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:51:07,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:51:07,632 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:51:07,633 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:51:07,713 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:51:08,391 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:51:08,391 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:51:09,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:10,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:14,626 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:15,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:20,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:20,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:24,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:51:24,500 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:51:24,500 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:51:24,582 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:51:25,402 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:51:26,402 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:51:26,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:26,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:26,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:31,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:31,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:35,596 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:51:36,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:36,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:41,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:41,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:41,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:41,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:51:41,952 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:51:41,953 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:51:42,034 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:51:42,412 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:51:42,413 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:51:46,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:47,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:51,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:52,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:57,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:51:58,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:58,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:51:58,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:51:58,790 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:51:58,791 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:51:58,872 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:51:59,423 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:52:00,424 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:52:02,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:03,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:05,602 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:52:08,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:08,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:13,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:15,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:15,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:15,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:52:15,951 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:52:15,952 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:52:16,032 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:52:16,440 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:52:16,440 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:52:19,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:20,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:24,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:25,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:29,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:33,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:33,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:33,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:52:33,951 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:52:33,953 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:52:34,034 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:52:34,464 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:52:34,468 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:52:35,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:35,608 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:52:38,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:40,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:43,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:45,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:50,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:52:50,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:50,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:50,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:52:50,721 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:52:50,723 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:52:50,803 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:52:51,475 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:52:52,475 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:52:53,476 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:52:55,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:52:55,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:00,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:00,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:05,614 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:53:06,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:09,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:09,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:53:09,154 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:53:09,155 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:53:09,237 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:53:09,486 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:53:10,487 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:53:11,488 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:53:12,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:14,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:17,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:19,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:22,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:25,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:25,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:53:25,616 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:53:25,617 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:53:25,699 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:53:26,497 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:53:26,498 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:53:27,498 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:53:27,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:30,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:32,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:35,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:35,619 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:53:38,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:43,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:43,436 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:53:43,439 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:53:43,440 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:53:43,521 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:53:43,521 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:53:44,521 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:53:44,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:45,522 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:53:48,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:49,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:53,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:53:54,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:53:59,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:00,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:00,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:54:00,302 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:54:00,303 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:54:00,386 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:54:00,531 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:54:01,532 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:54:02,533 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:54:05,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:05,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:05,625 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:54:10,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:10,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:15,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:17,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:17,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:54:17,283 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:54:17,285 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:54:17,365 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:54:17,542 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:54:18,543 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:54:19,544 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:54:21,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:22,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:26,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:27,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:31,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:35,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:35,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:35,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:54:35,457 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:54:35,459 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:54:35,539 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:54:35,554 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:54:35,631 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:54:36,554 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:54:36,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:37,555 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:54:40,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:41,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:45,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:46,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:51,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:54:51,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:51,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:51,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:54:51,786 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:54:51,788 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:54:51,868 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:54:52,564 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:54:52,565 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:54:53,565 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:54:56,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:54:56,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:01,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:01,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:05,638 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:55:07,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:10,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:10,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:55:10,602 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:55:10,604 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:55:10,684 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:55:11,576 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:55:12,577 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:55:12,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:13,577 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:55:15,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:17,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:20,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:22,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:27,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:27,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:55:27,030 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:55:27,032 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:55:27,112 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:55:27,586 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:55:28,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:28,587 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:55:29,587 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:55:32,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:33,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:35,643 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:55:37,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:38,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:43,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:44,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:44,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:55:44,804 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:55:44,804 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:55:44,886 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:55:45,597 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:55:46,598 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:55:47,599 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:55:48,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:49,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:53,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:55:54,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:55:58,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:01,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:01,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:56:01,742 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:56:01,743 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:56:01,824 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:56:02,608 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:56:02,609 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:56:03,609 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:56:04,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:05,648 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:56:06,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:10,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:11,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:15,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:18,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:18,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:56:18,253 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:56:18,254 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:56:18,339 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:56:18,619 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:56:18,619 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:56:19,619 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:56:21,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:23,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:26,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:28,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:31,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:33,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:35,654 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:56:36,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:37,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:56:37,156 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 16:56:37,156 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 16:56:37,236 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 16:56:37,631 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 16:56:39,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:39,632 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 16:56:42,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:44,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:47,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:49,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:52,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:53,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:56:54,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:56:54,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:56:54,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:56:54,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:56:55,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:56:57,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:56:58,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:56:59,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:00,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:00,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:00,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:00,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:00,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:00,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:01,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:02,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:04,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:04,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:05,660 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:57:06,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:06,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:06,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:06,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:06,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:07,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:07,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:07,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:09,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:10,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:12,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:12,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:12,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:12,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:13,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:13,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:13,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:13,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:14,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:17,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:17,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:18,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:18,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:18,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:19,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:19,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:19,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:19,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:19,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:22,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:23,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:24,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:25,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:25,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:25,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:25,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:25,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:25,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:26,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:27,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:29,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:29,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:31,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:31,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:31,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:31,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:31,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:32,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:32,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:32,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:34,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:35,666 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:57:35,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:37,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:37,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:37,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:37,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:38,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:38,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:38,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:38,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:39,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:42,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:42,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:43,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:43,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:43,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:44,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:44,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:44,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:44,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:44,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:47,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:48,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:49,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:50,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:50,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:50,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:50,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:50,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:50,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:51,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:52,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:54,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:57:54,620 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:56,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:56,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:56,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:56,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:56,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:57,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:57,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:57:57,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:57:59,167 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:00,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:02,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:02,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:02,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:02,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:03,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:03,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:03,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:03,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:04,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:05,672 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:58:07,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:07,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:09,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:09,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:09,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:09,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:09,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:09,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:09,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:10,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:12,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:13,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:14,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:15,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:15,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:15,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:15,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:15,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:15,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:16,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:17,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:19,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:19,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:21,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:21,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:21,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:22,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:22,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:22,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:22,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:22,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:24,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:26,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:27,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:27,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:27,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:27,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:28,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:28,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:28,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:28,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:29,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:32,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:32,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:34,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:34,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:34,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:34,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:34,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:34,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:34,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:35,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:35,677 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:58:37,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:38,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:39,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:40,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:40,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:40,404 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:40,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:40,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:41,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:41,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:42,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:44,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:44,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:46,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:46,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:46,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:47,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:47,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:47,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:47,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:47,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:49,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:51,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:52,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:52,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:52,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:52,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:53,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:53,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:53,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:53,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:54,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:57,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:57,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:58:59,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:58:59,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:59,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:59,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:59,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:59,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:58:59,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:00,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:02,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:03,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:04,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:05,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:05,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:05,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:05,683 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:59:05,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:05,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:06,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:06,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:07,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:09,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:09,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:11,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:11,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:11,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:12,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:12,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:12,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:12,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:12,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:14,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:16,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:17,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:17,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:17,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:17,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:18,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:18,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:18,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:19,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:19,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:22,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:22,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:24,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:24,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:24,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:24,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:24,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:24,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:24,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:25,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:27,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:28,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:29,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:30,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:30,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:30,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:30,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:31,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:31,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:31,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:32,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:34,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:34,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:35,689 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 16:59:36,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:36,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:36,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:37,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:37,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:37,363 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:37,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:37,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:39,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:41,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:42,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:42,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:42,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:42,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:43,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:43,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:43,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:44,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:44,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:47,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:47,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:49,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:49,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:49,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:49,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:49,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:49,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:49,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:50,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:52,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:53,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:54,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 16:59:55,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:55,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:55,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:55,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:56,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:56,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:56,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 16:59:57,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 16:59:59,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:00,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:01,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:01,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:01,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:02,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:02,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:02,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:02,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:02,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:04,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:05,694 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:00:06,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:07,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:08,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:08,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:08,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:08,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:08,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:08,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:09,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:09,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:12,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:12,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:14,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:14,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:14,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:14,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:14,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:17,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:19,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:22,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:24,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:27,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:29,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:29,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:32,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:34,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:35,701 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:00:37,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:39,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:42,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:44,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:45,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:45,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:47,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:47,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:49,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:50,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:50,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:50,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:51,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:51,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:51,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:51,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:52,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:53,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:54,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:56,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:56,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:57,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:57,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:57,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:57,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:00:57,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:57,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:00:59,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:00:59,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:02,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:02,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:02,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:03,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:03,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:03,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:04,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:04,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:04,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:05,706 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:01:05,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:07,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:09,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:09,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:09,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:09,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:09,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:09,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:10,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:10,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:12,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:12,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:14,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:15,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:15,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:15,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:16,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:16,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:16,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:16,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:17,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:18,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:19,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:21,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:21,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:22,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:22,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:22,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:22,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:22,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:22,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:24,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:24,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:27,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:27,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:27,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:28,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:28,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:28,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:29,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:29,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:29,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:30,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:32,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:34,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:34,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:34,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:34,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:34,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:34,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:35,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:35,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:35,712 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:01:36,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:37,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:39,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:40,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:40,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:40,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:40,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:41,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:41,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:41,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:42,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:43,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:44,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:46,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:46,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:47,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:47,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:47,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:47,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:47,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:47,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:49,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:49,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:52,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:52,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:52,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:53,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:53,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:53,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:53,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:53,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:54,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:55,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:57,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:01:59,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:59,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:59,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:01:59,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:59,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:01:59,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:00,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:00,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:01,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:02,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:04,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:05,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:05,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:05,717 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:02:05,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:05,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:06,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:06,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:06,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:07,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:08,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:09,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:11,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:11,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:12,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:12,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:12,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:12,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:12,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:12,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:14,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:14,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:17,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:17,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:17,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:18,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:18,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:18,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:18,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:18,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:19,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:20,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:22,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:24,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:24,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:24,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:24,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:24,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:24,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:25,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:25,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:26,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:27,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:29,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:30,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:30,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:30,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:30,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:30,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:31,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:31,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:32,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:33,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:34,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:35,723 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:02:36,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:36,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:36,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:37,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:37,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:37,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:37,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:37,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:39,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:39,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:42,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:42,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:42,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:43,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:43,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:43,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:43,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:43,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:44,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:45,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:47,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:49,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:49,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:49,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:49,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:49,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:49,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:50,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:50,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:51,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:52,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:54,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:02:55,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:55,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:55,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:55,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:55,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:56,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:56,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:57,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:02:58,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:02:59,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:01,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:01,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:01,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:02,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:02,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:02,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:02,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:02,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:04,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:04,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:05,729 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:03:07,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:07,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:07,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:08,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:08,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:08,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:09,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:09,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:09,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:10,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:12,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:14,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:14,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:14,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:14,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:14,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:14,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:15,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:15,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:17,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:17,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:19,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:20,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:20,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:21,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:21,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:21,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:21,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:21,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:22,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:23,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:24,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:26,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:26,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:27,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:27,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:27,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:27,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:27,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:27,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:29,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:29,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:32,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:33,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:33,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:33,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:33,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:33,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:34,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:34,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:34,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:35,735 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:03:36,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:37,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:39,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:39,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:39,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:39,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:39,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:40,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:40,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:40,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:42,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:42,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:44,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:45,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:45,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:46,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:46,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:46,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:46,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:46,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:47,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:48,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:49,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:51,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:51,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:52,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:52,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:52,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:52,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:52,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:53,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:54,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:54,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:57,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:03:58,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:58,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:58,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:58,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:58,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:59,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:03:59,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:03:59,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:01,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:02,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:04,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:04,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:04,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:04,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:05,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:05,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:05,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:05,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:05,741 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:04:07,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:09,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:12,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:14,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:17,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:19,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:20,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:22,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:25,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:27,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:27,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:27,994 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:04:28,096 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:04:28,156 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:04:28,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:04:29,991 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:04:30,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:33,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:35,747 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:04:35,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:38,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:40,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:43,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:46,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:46,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:04:46,351 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:04:46,352 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:04:46,432 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:04:47,001 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:04:48,002 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:04:49,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:51,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:54,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:04:56,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:04:59,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:02,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:02,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:05:02,705 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:05:02,705 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:05:02,788 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:05:03,011 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:05:04,012 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:05:04,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:05,752 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:05:07,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:10,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:12,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:15,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:19,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:19,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:19,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:05:19,698 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:05:19,698 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:05:19,780 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:05:20,022 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:05:20,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:22,023 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:05:24,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:25,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:29,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:30,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:35,758 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:05:36,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:36,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:36,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:36,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:05:36,609 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:05:36,610 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:05:36,693 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:05:37,033 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:05:38,033 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:05:41,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:41,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:46,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:46,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:51,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:05:51,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:54,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:05:54,009 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:05:54,009 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:05:54,090 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:05:55,090 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:05:56,091 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:05:56,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:05:57,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:02,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:02,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:05,764 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:06:07,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:07,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:07,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:11,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:06:11,696 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:06:11,696 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:06:11,778 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:06:12,101 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:06:12,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:12,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:14,103 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:06:17,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:17,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:22,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:22,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:27,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:28,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:28,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:06:28,847 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:06:28,849 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:06:28,932 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:06:29,112 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:06:30,113 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:06:32,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:33,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:35,769 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:06:38,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:38,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:43,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:46,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:46,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:06:46,965 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:06:46,966 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:06:47,046 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:06:47,123 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:06:48,124 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:06:49,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:51,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:54,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:06:56,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:06:59,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:03,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:03,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:07:03,864 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:07:03,865 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:07:03,946 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:07:04,134 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:07:04,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:05,775 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:07:06,135 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:07:08,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:10,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:13,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:15,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:20,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:20,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:07:20,359 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:07:20,359 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:07:20,440 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:07:21,145 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:07:21,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:22,146 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:07:25,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:26,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:30,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:31,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:35,781 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:07:36,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:37,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:37,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:37,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:07:37,490 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:07:37,490 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:07:37,572 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:07:38,156 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:07:38,156 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:07:40,157 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:07:42,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:42,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:47,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:47,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:52,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:07:52,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:55,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:07:55,076 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:07:55,076 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:07:55,158 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:07:55,166 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:07:56,167 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:07:58,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:07:58,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:03,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:03,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:05,787 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:08:08,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:09,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:09,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:12,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:08:12,811 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:08:12,812 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:08:12,892 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:08:13,178 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:08:13,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:14,178 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:08:14,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:18,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:19,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:23,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:24,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:28,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:29,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:30,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:08:30,004 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:08:30,005 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:08:30,085 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:08:30,188 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:08:32,190 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:08:34,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:34,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:35,790 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:08:39,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:39,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:44,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:47,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:47,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:08:47,977 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:08:47,978 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:08:48,059 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:08:48,200 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:08:50,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:50,201 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:08:52,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:08:55,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:08:57,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:00,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:04,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:04,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:09:04,687 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:09:04,688 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:09:04,770 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:09:05,211 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:09:05,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:05,796 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:09:06,211 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:09:09,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:10,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:14,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:15,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:19,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:20,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:21,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:09:21,742 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:09:21,742 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:09:21,825 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:09:22,222 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:09:22,222 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:09:24,223 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:09:24,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:25,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:29,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:30,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:35,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:35,801 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:09:36,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:38,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:09:38,991 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:09:38,992 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:09:39,072 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:09:39,233 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:09:40,233 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:09:40,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:42,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:45,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:47,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:50,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:52,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:56,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:56,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:09:56,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:09:56,710 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:09:56,710 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:09:56,791 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:09:57,244 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:09:57,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:09:58,245 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:10:01,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:02,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:05,807 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:10:06,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:07,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:12,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:13,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:13,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:10:13,872 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:10:13,873 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:10:13,957 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:10:14,255 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:10:14,256 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:10:16,256 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:10:17,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:18,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:22,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:23,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:27,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:31,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:31,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:31,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:10:31,145 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:10:31,147 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:10:31,227 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:10:31,266 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:10:32,267 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:10:33,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:35,813 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:10:36,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:38,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:41,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:43,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:47,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:48,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:49,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:10:49,052 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:10:49,053 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:10:49,132 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:10:49,278 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:10:50,278 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:10:52,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:54,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:10:57,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:10:59,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:04,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:05,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:05,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:05,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:11:05,804 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:11:05,806 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:11:05,904 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:11:05,905 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:11:06,288 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:11:06,289 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:11:08,289 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:11:09,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:10,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:14,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:15,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:19,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:23,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:23,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:11:23,159 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:11:23,160 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:11:23,240 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:11:23,299 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:11:24,300 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:11:25,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:28,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:30,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:33,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:35,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:35,898 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:11:39,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:39,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:11:39,966 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:11:39,967 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:11:40,049 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:11:40,310 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:11:40,310 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:11:41,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:42,311 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:11:44,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:46,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:49,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:51,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:56,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:11:57,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:11:57,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:11:57,859 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:11:57,860 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:11:57,940 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:11:58,338 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:11:58,339 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:12:00,343 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:12:01,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:02,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:05,904 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:12:07,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:07,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:12,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:12,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:15,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:12:15,092 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:12:15,092 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:12:15,174 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:12:15,360 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:12:16,361 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:12:17,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:18,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:23,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:23,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:28,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:28,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:32,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:12:32,376 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:12:32,377 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:12:32,461 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:12:33,372 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:12:33,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:34,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:34,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:34,372 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:12:35,909 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:12:38,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:39,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:43,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:44,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:48,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:50,404 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:50,404 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:50,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:12:50,411 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:12:50,412 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:12:50,492 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:12:51,383 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:12:52,384 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:12:54,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:12:55,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:12:59,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:00,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:04,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:05,915 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:13:06,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:06,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:06,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:13:06,592 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:13:06,593 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:13:06,673 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:13:07,394 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:13:08,394 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:13:09,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:11,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:14,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:16,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:19,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:24,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:24,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:13:24,296 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:13:24,297 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:13:24,379 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:13:24,404 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:13:25,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:26,406 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:13:29,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:30,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:34,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:35,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:35,921 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:13:40,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:41,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:41,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:13:41,191 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:13:41,192 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:13:41,272 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:13:41,416 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:13:42,416 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:13:46,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:46,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:51,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:51,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:56,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:13:59,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:13:59,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:13:59,029 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:13:59,029 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:13:59,111 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:13:59,428 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:14:00,428 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:14:01,429 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:14:02,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:04,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:05,927 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:14:07,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:09,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:12,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:16,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:16,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:14:16,624 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:14:16,625 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:14:16,706 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:14:17,439 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:14:18,440 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:14:18,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:19,441 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:14:21,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:23,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:26,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:28,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:33,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:33,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:33,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:14:33,333 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:14:33,333 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:14:33,417 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:14:33,450 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:14:34,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:34,450 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:14:35,451 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:14:35,932 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:14:38,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:39,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:43,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:44,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:49,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:51,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:51,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:14:51,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:14:51,736 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:14:51,737 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:14:51,819 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:14:52,462 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:14:52,462 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:14:53,462 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:14:55,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:14:56,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:00,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:01,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:05,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:05,938 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:15:07,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:07,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:07,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:15:07,892 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:15:07,892 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:15:07,974 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:15:08,472 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:15:08,472 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:15:09,473 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:15:10,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:12,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:15,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:17,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:20,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:25,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:25,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:15:25,700 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:15:25,701 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:15:25,782 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:15:26,483 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:15:26,484 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:15:26,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:27,484 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:15:29,250 WARNING FileStreamThread:21069 [file_stream.py:request_with_retry():674] requests_with_retry encountered retryable exception: HTTPSConnectionPool(host='api.wandb.ai', port=443): Read timed out. (read timeout=180.0). func: functools.partial(>, timeout=180.0), args: ('https://api.wandb.ai/files/autometa/pyg-1-m-2.3/ky6h7dv9/file_stream',), kwargs: {'json': {'files': {'output.log': {'offset': 431, 'content': ['ERROR 2024-05-15T17:11:58.246762 3%|██ | 626/18756 [3:48:53<87:32:18, 17.38s/it]\n']}}, 'dropped': 0}}
+2024-05-15 17:15:30,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:31,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:35,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:35,944 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:15:36,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:41,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:42,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:42,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:15:42,647 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:15:42,648 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:15:42,729 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:15:43,495 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:15:44,496 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:15:45,496 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:15:47,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:47,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:52,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:52,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:57,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:15:59,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:15:59,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:15:59,974 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:15:59,975 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:16:00,060 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:16:00,506 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:16:00,506 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:16:01,506 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:16:03,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:04,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:05,950 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:16:08,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:09,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:13,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:17,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:17,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:16:17,672 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:16:17,673 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:16:17,755 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:16:18,517 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:16:19,518 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:16:19,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:22,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:24,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:27,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:29,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:34,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:34,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:16:34,331 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:16:34,331 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:16:34,412 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:16:34,527 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:16:35,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:35,528 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:16:35,956 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:16:37,529 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:16:39,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:40,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:44,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:45,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:49,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:50,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:52,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:16:52,655 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:16:52,655 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:16:52,736 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:16:53,539 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:16:54,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:16:55,541 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:16:56,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:16:59,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:01,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:05,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:05,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:05,961 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:17:06,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:08,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:17:08,947 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:17:08,947 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:17:09,028 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:17:09,549 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:17:10,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:11,551 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:17:12,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:15,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:17,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:20,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:22,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:26,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:26,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:17:26,317 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:17:26,318 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:17:26,398 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:17:26,560 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:17:27,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:27,561 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:17:29,562 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:17:31,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:32,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:35,965 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:17:36,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:37,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:42,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:43,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:43,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:17:43,717 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:17:43,718 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:17:43,798 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:17:44,572 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:17:45,573 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:17:48,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:48,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:53,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:17:53,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:17:58,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:00,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:00,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:18:00,786 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:18:00,788 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:18:00,870 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:18:01,582 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:18:03,584 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:18:03,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:05,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:05,972 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:18:08,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:10,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:13,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:18,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:18,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:18,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:18:18,987 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:18:18,987 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:18:19,068 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:18:19,594 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:18:21,595 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:18:23,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:24,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:28,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:29,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:34,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:35,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:35,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:35,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:18:35,533 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:18:35,534 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:18:35,615 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:18:35,977 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:18:36,615 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:18:37,616 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:18:39,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:40,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:44,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:45,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:49,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:53,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:53,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:18:53,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:18:53,429 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:18:53,430 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:18:53,510 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:18:53,626 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:18:55,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:18:55,627 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:18:58,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:00,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:03,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:05,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:05,983 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:19:10,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:10,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:10,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:19:10,195 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:19:10,196 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:19:10,276 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:19:10,636 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:19:11,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:11,637 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:19:15,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:16,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:20,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:21,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:26,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:27,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:27,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:19:27,301 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:19:27,302 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:19:27,384 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:19:27,647 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:19:29,648 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:19:31,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:32,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:35,989 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:19:36,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:37,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:41,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:45,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:45,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:19:45,121 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:19:45,121 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:19:45,202 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:19:45,658 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:19:47,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:47,660 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:19:50,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:52,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:19:55,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:19:57,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:02,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:02,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:20:02,181 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:20:02,182 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:20:02,263 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:20:02,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:02,669 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:20:03,670 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:20:05,994 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:20:07,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:07,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:12,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:12,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:17,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:19,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:19,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:20:19,988 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:20:19,988 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:20:20,070 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:20:20,681 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:20:21,681 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:20:23,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:24,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:28,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:29,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:33,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:36,000 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:20:36,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:36,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:20:36,579 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:20:36,580 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:20:36,660 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:20:36,691 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:20:37,692 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:20:38,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:39,693 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:20:41,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:43,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:46,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:48,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:52,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:53,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:20:54,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:20:54,112 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:20:54,113 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:20:54,192 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:20:54,703 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:20:55,703 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:20:57,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:20:59,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:02,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:04,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:06,006 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:21:07,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:07,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:10,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:11,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:21:11,472 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:21:11,472 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:21:11,555 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:21:11,723 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:21:12,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:13,728 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:21:15,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:17,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:20,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:22,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:25,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:28,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:28,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:21:28,610 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:21:28,611 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:21:28,691 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:21:28,764 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:21:29,768 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:21:30,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:31,773 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:21:33,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:35,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:36,012 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:21:38,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:41,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:46,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:46,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:46,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:21:46,102 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:21:46,103 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:21:46,183 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:21:46,791 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:21:47,791 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:21:51,094 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:51,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:21:56,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:21:56,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:01,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:02,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:02,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:22:02,884 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:22:02,885 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:22:02,966 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:22:03,801 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:22:03,802 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:22:05,803 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:22:06,018 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:22:07,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:07,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:12,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:12,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:17,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:20,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:20,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:22:20,357 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:22:20,358 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:22:20,438 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:22:20,812 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:22:21,813 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:22:22,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:25,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:27,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:30,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:32,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:36,024 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:22:37,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:37,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:37,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:37,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:22:37,547 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:22:37,548 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:22:37,628 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:22:37,823 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:22:39,825 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:22:42,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:42,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:47,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:47,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:52,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:22:52,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:55,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:22:55,178 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:22:55,179 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:22:55,261 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:22:55,835 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:22:55,835 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:22:57,836 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:22:58,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:22:58,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:03,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:03,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:06,029 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:23:09,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:09,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:09,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:12,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:23:12,743 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:23:12,743 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:23:12,825 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:23:12,846 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:23:13,847 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:23:14,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:14,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:15,848 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:23:19,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:19,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:24,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:24,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:29,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:29,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:29,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:23:29,956 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:23:29,957 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:23:30,038 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:23:30,858 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:23:31,858 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:23:34,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:35,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:36,035 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:23:39,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:41,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:46,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:47,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:47,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:23:47,008 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:23:47,008 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:23:47,088 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:23:47,869 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:23:47,869 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:23:49,870 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:23:51,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:52,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:23:56,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:23:57,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:01,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:04,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:04,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:24:04,397 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:24:04,397 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:24:04,478 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:24:04,880 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:24:05,880 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:24:06,041 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:24:07,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:09,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:12,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:14,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:17,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:21,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:21,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:24:21,645 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:24:21,646 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:24:21,730 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:24:21,891 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:24:22,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:23,892 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:24:26,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:27,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:31,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:32,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:36,047 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:24:38,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:38,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:38,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:38,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:24:38,890 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:24:38,891 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:24:38,973 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:24:38,973 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:24:39,973 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:24:41,975 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:24:43,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:43,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:48,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:48,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:53,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:24:56,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:56,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:24:56,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:24:56,539 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:24:56,540 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:24:56,622 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:24:56,985 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:24:57,986 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:24:59,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:01,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:04,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:06,053 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:25:06,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:10,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:13,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:13,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:13,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:25:13,552 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:25:13,553 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:25:13,634 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:25:13,996 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:25:15,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:15,997 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:25:18,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:20,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:23,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:25,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:30,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:31,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:31,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:25:31,094 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:25:31,095 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:25:31,175 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:25:32,007 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:25:32,008 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:25:34,009 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:25:36,058 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:25:36,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:36,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:41,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:41,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:46,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:47,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:47,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:25:47,926 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:25:47,928 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:25:48,008 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:25:48,018 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:25:50,019 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:25:52,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:52,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:25:57,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:25:57,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:02,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:05,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:05,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:26:05,253 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:26:05,254 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:26:05,334 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:26:06,029 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:26:06,030 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:26:06,064 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:26:07,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:08,031 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:26:10,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:12,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:15,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:17,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:22,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:22,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:22,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:26:22,694 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:26:22,695 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:26:22,775 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:26:23,040 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:26:24,041 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:26:27,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:27,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:32,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:32,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:36,070 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:26:38,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:39,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:39,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:26:39,947 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:26:39,948 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:26:40,028 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:26:40,051 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:26:42,053 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:26:44,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:44,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:49,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:49,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:54,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:26:57,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:57,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:26:58,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:26:58,004 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:26:58,005 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:26:58,085 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:26:59,086 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:26:59,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:00,086 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:27:02,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:04,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:06,077 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:27:07,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:10,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:15,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:15,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:15,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:27:15,008 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:27:15,009 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:27:15,090 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:27:15,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:15,098 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:27:16,099 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:27:20,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:20,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:25,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:25,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:30,094 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:32,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:32,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:32,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:27:32,375 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:27:32,376 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:27:32,456 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:27:33,110 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:27:34,110 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:27:35,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:36,082 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:27:37,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:41,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:42,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:46,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:49,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:49,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:49,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:27:49,212 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:27:49,212 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:27:49,294 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:27:50,121 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:27:50,121 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:27:51,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:52,122 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:27:54,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:27:56,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:27:59,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:01,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:06,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:06,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:06,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:28:06,091 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:28:06,094 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:28:06,104 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:28:06,185 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:28:06,186 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:28:07,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:08,187 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:28:11,081 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:12,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:16,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:17,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:22,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:24,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:24,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:24,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:28:24,336 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:28:24,336 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:28:24,418 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:28:25,198 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:28:26,198 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:28:27,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:29,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:32,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:34,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:36,094 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:28:38,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:41,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:41,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:28:41,545 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:28:41,546 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:28:41,626 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:28:42,209 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:28:43,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:44,210 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:28:46,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:48,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:51,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:53,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:58,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:28:59,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:28:59,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:28:59,762 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:28:59,763 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:28:59,844 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:29:00,220 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:29:02,222 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:29:03,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:04,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:29:06,100 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:29:09,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:09,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:29:14,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:16,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:29:16,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:29:16,878 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:29:16,879 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:29:16,963 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:29:17,231 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:29:18,232 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:29:19,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:21,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:29:24,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:26,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:29:29,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:34,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:29:34,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:29:34,012 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:29:34,012 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:29:34,098 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:29:34,242 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:29:35,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:36,105 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:29:36,244 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:29:39,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:29:40,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:44,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:29:45,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:50,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:51,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:29:51,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:29:51,074 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:29:51,074 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:29:51,156 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:29:51,253 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:29:52,254 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:29:55,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:29:56,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:00,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:01,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:05,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:06,111 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:30:07,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:07,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:07,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:30:07,913 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:30:07,913 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:30:07,994 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:30:08,264 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:30:10,266 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:30:10,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:12,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:15,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:17,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:20,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:25,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:26,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:26,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:26,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:30:26,073 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:30:26,074 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:30:26,156 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:30:26,276 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:30:28,277 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:30:31,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:31,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:36,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:36,117 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:30:37,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:42,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:43,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:43,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:30:43,345 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:30:43,345 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:30:43,425 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:30:44,288 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:30:44,288 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:30:47,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:48,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:52,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:30:53,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:30:57,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:01,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:01,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:31:01,244 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:31:01,245 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:31:01,327 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:31:02,326 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:31:02,326 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:31:03,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:06,123 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:31:06,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:09,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:11,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:14,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:18,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:18,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:31:18,363 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:31:18,364 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:31:18,444 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:31:19,336 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:31:19,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:20,337 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:31:23,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:24,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:28,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:29,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:34,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:35,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:35,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:31:35,293 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:31:35,294 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:31:35,378 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:31:35,380 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:31:36,129 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:31:36,381 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:31:40,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:40,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:45,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:45,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:50,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:52,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:31:52,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:31:52,827 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:31:52,827 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:31:52,908 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:31:53,395 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:31:54,395 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:31:55,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:31:57,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:00,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:02,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:05,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:06,135 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:32:09,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:09,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:09,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:32:09,754 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:32:09,754 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:32:09,837 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:32:10,406 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:32:10,406 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:32:11,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:12,407 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:32:14,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:16,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:19,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:21,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:26,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:27,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:27,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:27,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:32:27,693 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:32:27,694 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:32:27,774 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:32:28,417 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:32:28,418 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:32:32,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:32,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:36,141 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:32:37,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:38,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:43,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:45,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:45,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:32:45,316 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:32:45,317 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:32:45,400 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:32:45,428 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:32:46,429 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:32:48,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:50,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:53,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:32:55,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:32:58,404 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:00,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:02,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:02,977 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:33:02,977 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:33:03,056 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:33:03,440 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:33:04,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:04,440 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:33:05,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:06,146 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:33:09,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:10,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:14,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:15,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:19,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:19,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:20,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:20,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:20,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:20,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:20,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:24,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:24,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:25,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:25,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:26,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:26,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:26,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:26,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:26,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:27,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:29,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:30,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:30,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:32,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:32,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:32,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:32,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:32,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:32,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:33,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:34,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:35,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:36,153 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:33:36,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:38,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:38,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:38,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:38,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:38,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:39,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:39,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:39,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:40,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:43,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:44,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:44,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:44,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:44,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:45,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:45,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:45,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:45,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:45,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:49,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:49,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:50,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:50,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:50,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:50,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:51,404 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:51,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:51,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:52,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:54,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:33:55,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:55,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:33:57,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:57,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:57,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:57,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:57,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:57,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:58,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:33:59,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:00,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:01,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:03,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:03,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:03,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:03,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:04,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:04,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:04,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:04,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:05,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:06,159 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:34:08,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:09,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:09,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:09,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:09,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:10,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:10,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:10,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:10,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:10,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:14,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:14,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:15,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:16,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:16,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:16,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:16,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:16,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:16,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:17,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:19,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:20,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:20,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:22,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:22,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:22,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:22,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:22,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:23,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:23,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:24,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:25,989 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:26,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:28,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:28,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:28,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:29,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:29,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:29,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:29,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:29,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:30,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:33,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:34,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:34,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:35,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:35,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:35,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:35,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:35,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:35,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:36,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:36,164 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:34:39,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:39,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:40,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:41,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:41,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:41,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:41,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:41,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:41,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:42,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:44,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:45,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:45,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:47,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:47,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:47,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:48,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:48,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:48,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:48,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:49,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:50,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:52,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:53,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:53,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:53,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:54,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:34:54,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:54,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:54,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:54,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:55,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:34:58,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:34:59,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:00,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:00,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:00,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:00,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:00,620 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:00,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:00,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:01,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:04,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:04,552 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:06,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:06,170 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:35:06,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:06,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:06,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:06,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:06,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:06,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:07,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:09,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:10,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:11,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:12,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:12,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:12,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:13,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:13,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:13,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:13,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:14,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:16,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:17,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:18,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:18,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:18,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:19,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:19,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:19,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:19,467 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:19,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:21,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:23,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:24,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:25,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:25,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:25,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:25,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:25,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:25,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:26,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:26,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:29,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:29,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:31,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:31,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:31,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:31,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:31,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:31,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:32,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:32,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:34,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:35,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:36,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:36,176 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:35:37,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:37,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:37,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:38,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:38,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:38,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:38,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:39,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:41,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:42,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:43,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:43,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:43,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:44,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:44,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:44,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:44,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:45,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:46,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:48,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:49,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:50,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:50,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:50,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:50,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:50,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:50,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:51,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:51,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:54,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:35:54,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:56,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:35:56,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:56,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:56,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:56,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:57,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:57,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:57,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:35:59,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:01,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:01,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:02,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:02,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:02,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:03,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:03,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:03,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:03,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:04,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:06,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:06,182 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:36:07,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:08,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:08,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:08,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:09,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:09,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:09,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:09,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:10,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:11,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:13,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:14,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:15,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:15,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:15,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:15,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:15,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:15,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:16,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:16,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:19,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:19,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:21,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:21,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:21,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:21,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:21,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:22,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:22,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:22,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:24,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:26,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:26,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:27,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:27,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:27,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:28,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:28,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:28,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:28,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:29,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:31,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:32,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:33,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:34,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:34,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:34,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:34,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:34,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:34,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:35,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:36,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:36,188 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:36:38,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:39,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:40,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:40,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:40,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:40,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:41,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:44,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:46,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:49,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:51,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:54,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:36:55,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:36:56,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:36:59,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:01,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:04,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:06,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:06,195 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:37:09,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:11,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:12,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:12,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:13,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:14,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:16,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:17,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:17,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:17,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:17,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:18,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:18,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:18,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:19,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:20,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:21,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:23,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:23,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:23,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:24,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:24,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:24,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:24,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:24,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:26,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:26,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:29,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:29,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:29,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:30,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:30,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:30,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:30,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:30,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:31,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:32,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:34,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:36,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:36,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:36,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:36,199 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:37:36,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:36,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:36,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:37,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:37,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:38,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:40,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:41,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:42,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:42,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:42,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:42,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:42,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:43,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:43,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:45,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:45,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:46,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:48,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:48,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:48,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:49,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:49,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:49,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:49,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:50,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:51,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:51,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:54,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:54,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:55,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:55,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:37:55,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:55,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:55,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:55,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:37:56,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:37:57,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:00,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:00,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:00,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:01,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:01,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:01,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:01,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:02,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:02,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:03,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:05,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:06,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:06,205 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:38:07,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:07,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:07,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:07,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:07,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:08,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:08,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:10,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:10,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:11,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:13,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:13,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:13,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:13,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:14,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:14,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:14,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:15,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:16,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:16,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:19,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:19,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:20,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:20,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:20,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:20,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:20,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:20,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:21,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:22,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:25,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:25,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:25,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:26,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:26,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:26,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:26,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:26,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:26,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:28,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:30,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:31,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:32,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:32,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:32,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:32,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:32,786 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:33,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:33,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:34,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:35,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:36,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:36,211 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:38:38,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:38,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:38,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:38,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:38,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:39,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:39,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:40,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:41,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:41,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:44,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:44,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:45,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:45,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:45,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:45,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:45,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:45,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:46,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:47,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:50,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:50,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:50,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:51,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:51,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:51,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:51,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:51,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:51,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:53,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:55,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:38:56,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:38:57,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:57,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:57,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:57,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:57,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:58,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:58,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:38:59,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:00,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:01,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:03,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:03,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:03,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:03,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:03,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:04,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:04,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:05,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:06,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:06,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:06,217 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:39:09,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:09,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:09,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:10,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:10,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:10,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:10,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:10,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:11,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:12,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:15,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:15,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:15,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:16,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:16,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:16,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:16,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:16,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:16,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:18,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:20,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:21,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:21,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:21,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:22,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:22,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:22,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:23,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:23,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:24,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:25,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:26,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:28,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:28,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:28,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:28,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:28,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:29,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:29,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:30,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:31,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:31,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:34,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:34,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:34,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:34,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:35,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:35,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:35,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:35,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:36,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:36,222 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:39:37,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:40,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:40,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:40,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:41,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:41,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:41,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:41,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:41,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:41,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:43,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:45,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:46,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:46,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:46,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:47,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:47,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:47,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:47,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:47,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:49,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:50,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:51,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:53,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:53,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:53,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:53,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:54,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:54,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:54,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:55,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:39:56,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:39:56,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:59,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:39:59,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:00,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:00,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:00,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:00,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:00,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:00,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:01,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:02,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:05,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:05,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:05,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:06,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:06,229 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:40:06,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:06,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:06,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:07,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:07,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:08,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:10,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:11,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:12,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:12,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:12,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:12,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:12,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:13,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:13,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:15,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:15,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:16,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:18,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:18,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:18,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:19,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:19,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:19,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:19,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:20,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:21,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:21,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:24,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:24,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:25,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:25,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:25,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:25,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:25,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:25,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:26,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:27,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:30,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:31,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:31,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:31,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:31,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:31,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:31,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:32,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:32,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:35,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:36,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:36,234 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:40:40,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:41,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:45,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:46,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:47,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:50,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:54,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:40:54,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:40:54,885 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:40:54,940 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:40:54,994 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:40:55,815 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:40:55,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:40:56,816 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:40:57,816 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:40:59,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:00,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:04,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:05,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:06,240 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:41:09,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:11,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:12,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:41:12,308 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:41:12,308 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:41:12,346 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:41:12,826 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:41:12,826 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:41:13,826 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:41:15,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:16,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:20,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:21,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:25,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:25,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:26,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:29,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:41:29,205 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:41:29,205 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:41:29,242 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:41:29,837 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:41:31,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:31,838 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:41:32,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:36,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:36,246 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:41:37,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:41,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:42,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:46,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:46,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:41:46,210 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:41:46,212 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:41:46,292 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:41:46,848 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:41:47,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:47,848 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:41:51,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:52,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:41:56,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:41:57,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:02,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:02,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:02,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:42:02,574 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:42:02,574 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:42:02,661 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:42:02,858 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:42:03,859 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:42:05,860 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:42:06,251 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:42:07,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:08,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:12,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:13,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:18,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:20,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:20,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:42:20,956 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:42:20,956 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:42:21,036 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:42:21,870 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:42:23,872 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:42:24,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:25,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:29,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:30,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:34,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:36,257 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:42:37,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:37,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:42:37,909 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:42:37,909 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:42:37,990 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:42:38,881 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:42:39,882 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:42:39,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:42,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:44,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:47,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:49,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:54,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:42:55,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:55,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:42:55,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:42:55,679 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:42:55,680 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:42:55,760 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:42:55,892 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:42:57,893 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:43:00,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:00,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:05,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:05,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:06,263 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:43:11,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:13,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:13,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:13,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:43:13,696 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:43:13,697 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:43:13,777 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:43:13,904 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:43:15,905 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:43:16,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:18,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:21,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:23,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:26,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:30,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:30,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:30,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:43:30,084 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:43:30,084 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:43:30,164 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:43:30,915 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:43:31,915 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:43:32,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:35,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:36,269 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:43:37,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:40,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:42,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:47,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:47,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:47,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:43:47,575 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:43:47,575 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:43:47,656 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:43:47,926 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:43:49,927 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:43:52,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:52,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:43:57,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:43:57,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:02,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:03,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:03,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:44:03,858 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:44:03,859 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:44:03,941 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:44:04,941 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:44:05,942 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:44:06,275 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:44:08,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:08,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:13,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:13,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:18,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:22,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:22,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:44:22,199 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:44:22,199 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:44:22,280 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:44:22,953 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:44:23,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:23,954 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:44:27,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:28,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:32,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:33,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:36,281 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:44:38,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:39,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:39,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:44:39,203 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:44:39,203 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:44:39,284 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:44:39,964 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:44:41,965 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:44:43,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:44,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:48,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:49,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:53,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:44:56,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:44:56,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:44:56,239 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:44:56,239 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:44:56,320 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:44:56,976 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:44:57,976 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:44:58,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:01,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:03,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:06,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:06,287 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:45:09,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:14,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:14,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:14,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:14,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:45:14,400 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:45:14,400 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:45:14,481 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:45:15,004 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:45:16,005 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:45:19,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:19,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:24,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:24,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:29,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:30,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:30,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:30,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:45:30,584 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:45:30,585 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:45:30,666 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:45:31,031 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:45:32,031 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:45:34,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:35,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:36,292 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:45:40,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:40,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:45,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:48,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:48,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:48,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:45:48,347 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:45:48,348 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:45:48,428 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:45:49,044 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:45:50,045 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:45:50,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:53,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:45:55,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:45:58,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:00,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:04,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:04,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:04,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:46:04,798 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:46:04,798 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:46:04,879 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:46:05,055 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:46:05,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:06,055 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:46:06,298 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:46:09,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:11,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:14,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:16,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:21,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:22,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:22,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:46:22,599 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:46:22,600 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:46:22,682 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:46:23,066 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:46:24,066 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:46:26,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:27,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:31,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:32,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:36,304 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:46:37,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:40,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:40,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:46:40,389 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:46:40,389 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:46:40,470 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:46:41,077 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:46:42,078 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:46:42,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:45,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:47,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:50,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:52,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:46:57,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:46:57,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:46:57,193 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:46:57,194 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:46:57,276 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:46:58,088 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:46:58,088 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:46:58,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:00,089 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:47:02,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:03,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:06,310 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:47:07,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:08,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:13,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:15,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:15,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:47:15,405 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:47:15,406 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:47:15,487 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:47:16,099 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:47:18,101 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:47:18,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:20,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:23,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:25,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:28,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:31,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:31,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:31,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:47:31,638 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:47:31,638 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:47:31,719 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:47:32,110 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:47:33,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:34,111 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:47:36,316 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:47:36,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:39,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:41,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:44,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:49,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:49,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:49,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:47:49,108 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:47:49,109 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:47:49,191 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:47:50,191 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:47:50,191 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:47:50,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:54,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:47:55,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:47:59,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:00,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:05,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:06,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:06,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:06,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:48:06,084 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:48:06,084 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:48:06,164 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:48:06,202 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:48:06,321 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:48:08,203 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:48:10,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:11,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:15,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:16,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:20,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:21,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:21,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:23,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:48:23,933 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:48:23,933 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:48:24,014 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:48:24,213 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:48:26,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:26,214 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:48:26,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:31,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:31,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:36,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:36,327 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:48:37,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:41,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:41,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:48:41,942 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:48:41,942 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:48:42,023 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:48:42,224 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:48:43,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:44,226 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:48:47,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:48,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:52,026 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:53,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:57,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:48:59,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:48:59,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:48:59,133 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:48:59,133 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:48:59,215 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:48:59,235 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:49:00,236 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:49:02,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:04,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:06,333 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:49:07,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:09,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:12,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:17,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:17,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:17,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:17,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:49:17,394 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:49:17,396 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:49:17,477 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:49:18,247 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:49:18,248 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:49:20,248 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:49:22,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:22,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:27,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:27,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:32,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:33,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:33,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:33,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:49:33,898 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:49:33,898 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:49:33,979 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:49:34,257 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:49:36,259 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:49:36,338 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:49:38,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:38,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:43,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:43,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:48,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:48,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:50,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:49:51,001 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:49:51,002 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:49:51,085 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:49:51,268 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:49:52,269 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:49:53,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:54,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:49:58,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:49:59,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:04,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:04,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:04,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:06,344 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:50:08,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:50:08,387 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:50:08,387 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:50:08,468 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:50:09,280 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:50:09,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:10,280 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:50:10,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:14,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:15,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:19,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:20,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:24,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:25,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:25,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:50:25,950 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:50:25,951 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:50:26,032 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:50:26,291 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:50:28,292 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:50:30,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:30,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:35,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:35,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:36,350 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:50:40,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:43,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:43,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:50:43,240 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:50:43,241 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:50:43,322 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:50:43,323 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:50:44,323 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:50:46,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:48,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:51,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:50:53,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:50:56,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:00,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:00,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:51:00,485 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:51:00,486 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:51:00,566 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:51:01,334 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:51:01,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:02,335 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:51:05,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:06,356 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:51:07,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:10,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:12,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:17,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:17,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:17,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:51:17,982 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:51:17,983 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:51:18,063 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:51:18,345 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:51:20,346 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:51:22,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:23,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:27,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:28,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:33,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:35,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:35,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:35,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:51:35,239 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:51:35,241 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:51:35,322 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:51:35,356 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:51:36,362 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:51:36,363 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:51:38,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:40,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:43,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:45,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:48,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:52,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:52,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:52,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:51:52,402 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:51:52,403 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:51:52,483 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:51:53,374 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:51:53,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:51:54,375 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:51:57,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:51:58,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:02,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:03,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:06,368 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:52:09,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:09,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:09,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:09,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:52:09,609 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:52:09,609 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:52:09,694 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:52:10,386 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:52:10,387 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:52:12,388 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:52:14,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:14,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:19,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:19,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:24,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:27,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:27,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:52:27,140 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:52:27,141 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:52:27,223 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:52:27,397 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:52:28,398 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:52:30,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:32,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:35,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:36,373 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:52:37,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:40,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:44,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:44,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:52:44,062 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:52:44,064 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:52:44,145 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:52:44,408 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:52:46,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:46,410 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:52:49,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:51,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:52:54,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:52:56,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:01,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:01,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:01,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:53:01,642 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:53:01,643 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:53:01,724 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:53:02,420 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:53:02,421 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:53:06,380 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:53:06,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:06,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:11,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:11,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:16,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:19,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:19,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:53:19,116 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:53:19,118 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:53:19,199 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:53:19,431 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:53:20,432 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:53:22,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:24,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:27,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:29,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:32,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:36,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:36,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:36,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:53:36,318 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:53:36,319 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:53:36,425 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:53:36,428 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:53:36,442 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:53:37,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:38,443 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:53:41,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:42,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:46,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:47,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:52,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:53,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:53,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:53:53,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:53:53,750 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:53:53,751 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:53:53,836 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:53:54,453 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:53:54,454 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:53:56,455 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:53:57,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:53:58,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:02,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:03,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:06,428 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:54:08,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:10,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:10,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:10,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:54:10,650 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:54:10,652 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:54:10,732 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:54:11,464 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:54:12,465 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:54:13,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:15,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:18,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:20,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:23,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:28,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:28,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:28,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:54:28,644 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:54:28,646 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:54:28,727 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:54:29,476 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:54:29,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:30,477 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:54:33,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:34,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:36,434 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:54:38,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:40,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:45,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:45,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:45,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:54:45,440 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:54:45,442 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:54:45,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:45,524 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:54:45,525 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:54:46,525 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:54:50,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:50,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:54:55,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:54:55,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:00,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:03,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:03,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:03,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:55:03,095 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:55:03,097 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:55:03,178 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:55:03,535 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:55:04,536 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:55:06,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:06,439 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:55:08,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:11,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:13,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:16,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:20,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:20,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:55:20,515 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:55:20,516 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:55:20,598 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:55:21,598 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:55:21,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:22,599 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:55:25,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:26,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:30,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:31,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:36,445 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:55:37,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:37,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:37,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:55:37,303 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:55:37,304 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:55:37,390 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:55:37,608 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:55:38,609 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:55:42,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:42,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:47,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:47,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:52,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:55:55,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:55:55,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:55:55,434 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:55:55,436 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:55:55,517 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:55:55,619 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:55:56,620 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:55:57,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:00,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:02,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:05,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:06,450 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:56:08,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:12,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:12,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:56:12,215 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:56:12,217 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:56:12,301 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:56:12,630 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:56:14,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:14,632 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:56:17,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:19,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:22,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:24,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:29,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:30,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:30,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:56:30,181 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:56:30,181 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:56:30,265 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:56:30,642 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:56:30,642 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:56:32,643 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:56:35,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:35,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:36,455 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:56:40,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:40,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:45,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:46,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:46,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:46,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:56:46,958 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:56:46,959 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:56:47,040 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:56:47,653 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:56:48,653 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:56:51,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:51,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:56:56,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:56:56,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:01,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:04,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:04,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:04,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:57:04,290 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:57:04,291 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:57:04,374 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:57:04,664 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:57:04,664 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:57:06,375 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:06,460 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:57:06,665 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:57:09,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:11,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:14,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:16,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:21,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:21,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:22,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:57:22,008 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:57:22,008 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:57:22,088 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:57:22,675 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:57:22,676 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:57:27,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:27,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:32,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:32,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:36,466 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:57:37,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:38,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:38,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:57:38,751 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:57:38,752 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:57:38,832 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:57:39,686 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:57:40,687 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:57:42,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:43,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:47,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:48,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:52,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:56,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:57:56,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:57:56,926 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:57:56,927 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:57:57,009 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:57:57,698 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:57:58,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:57:58,698 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:58:01,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:03,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:06,472 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:58:06,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:08,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:13,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:13,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:13,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:58:13,651 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:58:13,652 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:58:13,732 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:58:14,732 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:58:14,733 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:58:18,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:18,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:23,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:23,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:28,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:30,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:30,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:58:30,968 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:58:30,968 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:58:31,050 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:58:31,744 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:58:32,744 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:58:34,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:35,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:36,478 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:58:39,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:40,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:44,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:48,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:48,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:58:48,171 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:58:48,172 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:58:48,253 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:58:48,755 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:58:48,755 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:58:50,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:53,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:58:55,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:58:58,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:00,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:03,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:05,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:05,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:59:05,795 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:59:05,796 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:59:05,877 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:59:06,484 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:59:06,766 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:59:06,766 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:59:08,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:10,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:13,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:15,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:19,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:20,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:23,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:59:23,544 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:59:23,545 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:59:23,628 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:59:23,777 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:59:24,777 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:59:25,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:25,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:30,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:30,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:35,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:36,490 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 17:59:36,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:36,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:40,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:59:40,401 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:59:40,402 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:59:40,483 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:59:40,788 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:59:41,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:41,788 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:59:42,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:42,789 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:59:46,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:47,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:51,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:52,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:56,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 17:59:58,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 17:59:58,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 17:59:58,231 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 17:59:58,232 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 17:59:58,312 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 17:59:58,799 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 17:59:58,800 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 17:59:59,800 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:00:02,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:03,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:06,496 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:00:07,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:08,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:12,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:14,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:14,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:00:14,937 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:00:14,938 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:00:15,019 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:00:15,810 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:00:16,811 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:00:17,811 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:00:18,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:19,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:23,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:24,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:28,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:30,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:32,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:00:32,291 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:00:32,291 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:00:32,372 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:00:32,821 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:00:32,821 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:00:33,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:33,821 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:00:35,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:36,502 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:00:38,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:40,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:43,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:45,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:48,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:49,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:00:49,652 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:00:49,653 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:00:49,734 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:00:49,832 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:00:50,832 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:00:51,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:51,833 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:00:53,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:00:56,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:00:58,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:01,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:03,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:06,507 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:01:07,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:07,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:07,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:01:07,247 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:01:07,248 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:01:07,330 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:01:07,843 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:01:08,844 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:01:09,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:09,844 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:01:12,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:14,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:17,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:19,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:24,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:24,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:24,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:24,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:01:24,393 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:01:24,395 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:01:24,476 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:01:24,862 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:01:25,862 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:01:26,862 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:01:29,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:29,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:34,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:34,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:36,513 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:01:39,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:41,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:41,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:01:41,427 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:01:41,427 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:01:41,508 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:01:41,872 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:01:42,872 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:01:43,873 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:01:45,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:46,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:50,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:51,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:55,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:01:56,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:01:59,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:01:59,098 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:01:59,099 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:01:59,180 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:01:59,883 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:02:00,884 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:02:01,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:01,884 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:02:02,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:06,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:06,519 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:02:07,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:11,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:12,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:16,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:02:16,261 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:02:16,262 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:02:16,342 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:02:16,894 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:02:16,894 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:02:17,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:17,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:17,895 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:02:22,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:22,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:27,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:27,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:32,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:33,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:33,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:02:33,550 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:02:33,550 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:02:33,632 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:02:33,905 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:02:35,906 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:02:36,525 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:02:37,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:38,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:42,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:43,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:47,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:50,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:50,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:02:50,798 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:02:50,799 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:02:50,882 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:02:50,915 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:02:51,916 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:02:52,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:02:53,917 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:02:55,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:02:57,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:00,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:02,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:06,530 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:03:08,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:08,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:08,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:03:08,192 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:03:08,193 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:03:08,274 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:03:08,927 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:03:09,927 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:03:13,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:13,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:18,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:18,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:23,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:25,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:25,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:25,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:03:25,072 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:03:25,074 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:03:25,156 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:03:25,938 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:03:27,939 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:03:29,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:30,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:34,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:35,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:36,536 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:03:39,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:42,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:42,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:42,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:03:42,395 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:03:42,395 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:03:42,477 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:03:42,948 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:03:43,949 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:03:45,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:47,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:50,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:52,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:55,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:03:59,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:59,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:03:59,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:03:59,978 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:03:59,979 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:04:00,060 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:04:00,960 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:04:01,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:01,960 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:04:04,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:06,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:06,542 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:04:09,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:11,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:16,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:16,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:16,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:16,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:04:16,933 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:04:16,934 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:04:17,015 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:04:17,015 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:04:18,015 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:04:20,017 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:04:21,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:22,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:26,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:27,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:32,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:34,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:34,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:04:34,566 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:04:34,567 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:04:34,649 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:04:35,027 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:04:36,027 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:04:36,548 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:04:37,550 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:39,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:42,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:44,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:47,552 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:51,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:51,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:04:51,551 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:04:51,553 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:04:51,634 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:04:52,038 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:04:52,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:04:54,039 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:04:56,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:04:57,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:01,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:02,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:06,554 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:05:08,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:09,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:09,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:05:09,574 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:05:09,575 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:05:09,655 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:05:10,049 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:05:12,050 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:05:13,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:14,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:18,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:19,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:23,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:26,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:26,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:05:26,404 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:05:26,405 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:05:26,486 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:05:27,060 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:05:28,061 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:05:29,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:31,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:34,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:36,404 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:36,559 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:05:39,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:43,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:43,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:43,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:05:43,595 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:05:43,595 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:05:43,676 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:05:44,071 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:05:44,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:46,072 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:05:48,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:49,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:53,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:05:54,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:05:59,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:01,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:01,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:01,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:06:01,303 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:06:01,304 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:06:01,384 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:06:02,082 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:06:04,084 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:06:05,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:06,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:06,565 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:06:10,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:11,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:15,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:18,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:18,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:18,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:06:18,015 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:06:18,017 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:06:18,098 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:06:19,098 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:06:20,099 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:06:21,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:23,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:26,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:28,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:31,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:36,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:36,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:36,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:36,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:06:36,472 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:06:36,473 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:06:36,554 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:06:36,571 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:06:37,110 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:06:38,110 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:06:41,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:41,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:46,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:46,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:51,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:53,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:06:53,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:06:53,450 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:06:53,450 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:06:53,532 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:06:54,121 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:06:56,122 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:06:57,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:06:58,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:02,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:03,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:06,577 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:07:07,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:11,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:11,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:07:11,335 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:07:11,335 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:07:11,418 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:07:12,133 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:07:12,133 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:07:13,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:14,134 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:07:16,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:18,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:21,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:23,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:28,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:28,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:07:28,146 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:07:28,148 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:07:28,228 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:07:29,144 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:07:29,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:30,144 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:07:33,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:34,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:36,583 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:07:38,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:39,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:44,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:45,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:45,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:07:45,198 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:07:45,199 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:07:45,280 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:07:46,155 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:07:46,156 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:07:48,156 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:07:50,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:50,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:07:55,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:07:55,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:00,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:02,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:02,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:08:02,953 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:08:02,954 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:08:03,041 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:08:03,166 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:08:04,166 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:08:06,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:06,589 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:08:07,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:11,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:12,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:16,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:19,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:19,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:19,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:08:19,791 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:08:19,792 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:08:19,874 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:08:20,177 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:08:21,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:22,178 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:08:24,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:26,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:29,786 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:31,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:36,595 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:08:37,598 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:38,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:38,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:38,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:08:38,455 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:08:38,456 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:08:38,536 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:08:39,189 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:08:40,190 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:08:43,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:43,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:48,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:48,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:53,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:08:55,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:08:55,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:08:55,424 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:08:55,424 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:08:55,505 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:08:56,200 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:08:56,200 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:08:58,201 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:08:59,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:00,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:04,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:05,417 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:06,601 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:09:09,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:13,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:13,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:13,057 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:09:13,057 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:09:13,140 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:09:13,211 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:09:14,211 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:09:15,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:18,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:20,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:23,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:25,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:29,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:29,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:29,841 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:09:29,842 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:09:29,923 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:09:30,221 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:09:30,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:32,223 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:09:34,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:35,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:36,607 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:09:39,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:41,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:44,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:46,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:46,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:47,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:47,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:47,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:47,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:49,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:51,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:51,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:52,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:52,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:52,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:53,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:53,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:53,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:53,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:54,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:09:56,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:09:57,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:59,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:59,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:59,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:59,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:59,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:59,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:09:59,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:00,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:01,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:03,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:04,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:05,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:05,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:05,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:05,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:05,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:06,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:06,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:06,612 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:10:06,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:09,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:09,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:11,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:11,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:11,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:11,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:12,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:12,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:12,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:12,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:14,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:16,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:16,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:17,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:17,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:17,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:18,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:18,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:18,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:18,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:19,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:21,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:22,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:24,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:24,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:24,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:24,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:24,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:24,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:24,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:25,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:26,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:28,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:29,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:30,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:30,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:30,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:30,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:30,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:31,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:31,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:31,620 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:34,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:34,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:36,618 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:10:36,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:36,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:36,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:36,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:37,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:37,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:37,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:37,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:39,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:41,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:41,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:42,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:42,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:42,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:43,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:43,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:43,550 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:43,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:44,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:46,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:47,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:49,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:49,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:49,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:49,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:49,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:49,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:49,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:50,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:51,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:53,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:54,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:55,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:55,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:55,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:55,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:55,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:56,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:56,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:10:56,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:10:59,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:10:59,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:01,626 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:01,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:01,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:01,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:02,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:02,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:02,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:02,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:04,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:06,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:06,624 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:11:06,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:07,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:07,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:07,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:08,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:08,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:08,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:08,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:09,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:11,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:12,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:14,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:14,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:14,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:14,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:14,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:14,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:14,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:15,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:16,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:18,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:19,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:20,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:20,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:20,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:20,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:20,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:21,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:21,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:21,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:24,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:24,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:26,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:26,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:26,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:26,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:27,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:27,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:27,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:27,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:29,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:31,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:31,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:32,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:32,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:32,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:33,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:33,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:33,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:33,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:34,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:36,630 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:11:36,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:37,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:39,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:39,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:39,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:39,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:39,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:39,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:39,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:40,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:41,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:43,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:44,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:45,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:45,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:45,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:45,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:45,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:46,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:46,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:46,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:49,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:49,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:51,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:51,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:51,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:51,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:52,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:52,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:52,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:52,786 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:54,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:11:56,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:56,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:11:57,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:57,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:57,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:58,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:58,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:58,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:59,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:11:59,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:01,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:02,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:04,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:04,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:04,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:04,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:04,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:04,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:04,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:05,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:06,635 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:12:06,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:08,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:09,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:10,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:10,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:10,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:10,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:11,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:11,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:11,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:11,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:14,874 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:15,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:16,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:16,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:16,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:16,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:17,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:17,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:17,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:17,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:19,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:21,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:21,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:22,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:22,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:22,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:23,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:23,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:23,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:24,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:24,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:26,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:27,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:29,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:29,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:29,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:29,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:29,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:29,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:29,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:30,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:31,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:33,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:34,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:35,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:35,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:35,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:35,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:36,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:36,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:36,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:36,642 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:12:36,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:39,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:40,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:41,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:41,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:41,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:41,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:42,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:42,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:42,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:42,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:44,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:46,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:46,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:47,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:47,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:47,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:48,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:48,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:48,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:49,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:49,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:51,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:52,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:54,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:54,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:54,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:54,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:54,874 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:54,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:12:54,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:55,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:56,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:12:58,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:12:59,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:00,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:00,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:00,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:01,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:01,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:01,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:01,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:01,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:04,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:05,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:06,647 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:13:06,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:06,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:06,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:06,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:06,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:09,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:11,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:14,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:16,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:19,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:21,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:22,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:24,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:26,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:29,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:31,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:34,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:36,653 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:13:36,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:37,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:37,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:39,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:39,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:41,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:42,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:42,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:43,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:43,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:43,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:43,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:43,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:44,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:45,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:46,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:49,094 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:49,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:49,552 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:49,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:49,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:49,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:50,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:50,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:51,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:51,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:54,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:13:55,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:55,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:55,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:55,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:56,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:56,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:56,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:56,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:13:58,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:13:59,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:01,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:01,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:01,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:02,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:02,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:02,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:02,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:02,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:04,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:04,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:06,658 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:14:06,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:07,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:07,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:08,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:08,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:08,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:09,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:09,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:09,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:10,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:11,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:14,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:14,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:14,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:14,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:14,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:14,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:15,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:15,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:16,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:17,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:19,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:20,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:20,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:20,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:21,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:21,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:21,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:21,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:21,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:23,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:24,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:26,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:26,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:26,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:27,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:27,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:27,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:27,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:27,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:29,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:29,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:31,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:32,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:32,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:33,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:33,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:33,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:34,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:34,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:34,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:35,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:36,664 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:14:36,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:39,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:39,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:39,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:39,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:39,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:39,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:40,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:40,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:41,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:42,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:44,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:45,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:45,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:46,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:46,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:46,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:46,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:46,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:46,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:48,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:49,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:51,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:51,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:51,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:52,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:52,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:52,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:52,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:52,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:54,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:54,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:14:56,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:14:58,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:58,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:58,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:58,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:58,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:59,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:59,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:14:59,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:01,026 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:01,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:04,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:04,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:04,844 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:04,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:04,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:05,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:05,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:05,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:06,670 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:15:06,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:07,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:09,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:10,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:10,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:11,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:11,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:11,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:11,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:11,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:11,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:13,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:14,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:16,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:16,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:16,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:17,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:17,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:17,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:18,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:18,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:19,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:19,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:21,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:23,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:23,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:23,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:23,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:23,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:24,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:24,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:24,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:26,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:26,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:29,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:29,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:29,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:29,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:29,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:30,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:30,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:30,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:31,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:32,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:34,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:35,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:35,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:36,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:36,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:36,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:36,677 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:15:36,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:36,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:36,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:38,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:39,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:41,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:41,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:41,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:42,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:42,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:42,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:42,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:42,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:44,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:44,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:46,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:48,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:48,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:48,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:48,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:48,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:49,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:49,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:49,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:51,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:51,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:54,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:54,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:54,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:54,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:15:54,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:55,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:55,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:55,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:56,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:15:57,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:15:59,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:00,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:00,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:01,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:01,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:01,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:01,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:01,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:01,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:03,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:04,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:06,682 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:16:06,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:06,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:06,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:07,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:07,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:07,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:08,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:08,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:09,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:09,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:11,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:13,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:13,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:13,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:13,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:13,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:14,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:14,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:14,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:16,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:16,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:19,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:19,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:19,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:19,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:19,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:20,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:20,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:20,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:21,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:22,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:24,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:25,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:25,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:26,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:26,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:26,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:26,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:26,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:26,832 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:28,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:29,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:31,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:31,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:31,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:32,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:32,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:32,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:33,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:33,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:34,844 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:34,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:36,687 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:16:36,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:38,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:38,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:38,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:38,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:38,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:39,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:39,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:39,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:41,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:41,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:44,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:44,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:44,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:44,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:45,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:45,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:45,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:45,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:46,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:47,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:49,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:50,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:50,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:51,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:51,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:51,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:51,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:51,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:51,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:53,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:54,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:16:56,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:16:56,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:56,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:57,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:57,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:57,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:58,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:58,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:16:59,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:01,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:04,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:06,693 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:17:06,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:09,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:11,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:13,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:17:16,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:16,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:20,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:17:20,737 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:17:20,845 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:17:20,900 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:17:21,636 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:17:21,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:22,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:22,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:22,637 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:17:26,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:27,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:31,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:33,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:36,699 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:17:37,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:37,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:17:37,317 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:17:37,319 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:17:37,399 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:17:37,646 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:17:38,647 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:17:39,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:39,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:42,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:44,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:47,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:50,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:52,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:17:54,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:17:54,126 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:17:54,128 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:17:54,206 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:17:54,657 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:17:54,657 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:17:56,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:56,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:17:58,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:01,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:03,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:06,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:06,705 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:18:08,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:11,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:11,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:11,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:18:11,525 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:18:11,526 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:18:11,609 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:18:11,668 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:18:12,668 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:18:14,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:16,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:19,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:21,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:24,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:27,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:27,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:27,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:18:27,863 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:18:27,864 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:18:27,945 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:18:28,679 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:18:28,679 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:18:29,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:32,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:34,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:36,711 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:18:37,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:40,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:42,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:42,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:45,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:46,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:18:46,867 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:18:46,867 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:18:46,948 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:18:47,691 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:18:48,691 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:18:48,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:50,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:53,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:55,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:18:59,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:18:59,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:00,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:03,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:19:03,419 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:19:03,420 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:19:03,503 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:19:03,701 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:19:04,702 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:19:05,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:06,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:06,716 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:19:10,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:11,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:15,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:16,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:21,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:21,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:22,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:19:22,004 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:19:22,005 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:19:22,087 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:19:22,713 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:19:22,714 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:19:24,714 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:19:26,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:27,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:31,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:32,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:36,723 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:19:37,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:39,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:39,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:19:39,043 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:19:39,044 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:19:39,124 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:19:39,728 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:19:40,729 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:19:43,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:44,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:48,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:49,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:53,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:19:55,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:19:55,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:19:55,934 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:19:55,935 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:19:56,018 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:19:56,739 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:19:56,740 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:19:59,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:00,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:04,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:05,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:06,728 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:20:09,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:13,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:13,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:20:13,236 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:20:13,236 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:20:13,318 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:20:13,750 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:20:14,751 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:20:15,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:18,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:20,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:23,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:25,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:29,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:29,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:29,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:20:29,454 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:20:29,454 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:20:29,535 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:20:29,760 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:20:30,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:30,761 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:20:34,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:35,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:36,734 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:20:39,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:40,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:45,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:45,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:45,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:48,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:20:48,341 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:20:48,342 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:20:48,423 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:20:48,772 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:20:49,773 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:20:50,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:50,774 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:20:51,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:20:55,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:20:56,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:01,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:01,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:01,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:04,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:21:05,001 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:21:05,002 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:21:05,083 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:21:05,807 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:21:06,740 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:21:06,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:06,807 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:21:06,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:11,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:11,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:16,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:16,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:21,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:23,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:23,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:21:23,137 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:21:23,139 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:21:23,219 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:21:23,833 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:21:24,834 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:21:27,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:28,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:32,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:33,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:36,746 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:21:37,749 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:40,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:40,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:21:40,509 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:21:40,510 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:21:40,591 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:21:40,844 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:21:41,845 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:21:42,846 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:21:43,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:45,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:48,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:50,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:53,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:57,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:21:57,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:21:57,404 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:21:57,405 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:21:57,487 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:21:57,855 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:21:58,856 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:21:59,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:21:59,857 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:22:02,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:04,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:06,751 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:22:07,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:09,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:14,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:14,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:14,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:22:14,979 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:22:14,980 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:22:15,060 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:22:15,867 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:22:16,868 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:22:17,868 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:22:19,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:20,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:24,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:25,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:30,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:31,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:31,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:31,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:22:31,210 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:22:31,210 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:22:31,293 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:22:31,877 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:22:32,878 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:22:33,878 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:22:35,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:36,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:36,757 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:22:40,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:41,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:45,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:49,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:49,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:49,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:22:49,008 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:22:49,009 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:22:49,090 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:22:49,889 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:22:50,889 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:22:51,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:51,890 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:22:54,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:22:56,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:22:59,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:01,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:06,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:06,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:06,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:23:06,319 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:23:06,320 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:23:06,402 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:23:06,764 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:23:06,899 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:23:07,900 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:23:08,901 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:23:11,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:11,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:16,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:16,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:21,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:22,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:24,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:23:24,534 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:23:24,534 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:23:24,615 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:23:24,911 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:23:25,912 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:23:26,912 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:23:27,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:27,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:32,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:32,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:36,769 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:23:37,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:38,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:42,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:23:42,462 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:23:42,462 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:23:42,544 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:23:42,923 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:23:43,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:43,923 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:23:44,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:45,925 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:23:48,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:49,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:53,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:54,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:58,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:23:59,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:23:59,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:23:59,939 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:23:59,941 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:24:00,024 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:24:00,934 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:24:01,935 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:24:04,026 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:04,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:06,775 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:24:09,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:09,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:14,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:17,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:17,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:17,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:24:17,403 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:24:17,404 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:24:17,484 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:24:17,945 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:24:19,947 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:24:20,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:22,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:25,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:27,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:30,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:34,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:34,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:34,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:24:34,167 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:24:34,168 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:24:34,248 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:24:34,956 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:24:35,957 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:24:36,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:36,780 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:24:39,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:41,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:44,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:46,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:51,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:51,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:51,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:24:51,104 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:24:51,105 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:24:51,187 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:24:51,967 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:24:52,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:24:53,969 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:24:56,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:24:57,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:01,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:02,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:06,786 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:25:07,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:08,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:08,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:08,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:25:08,473 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:25:08,473 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:25:08,554 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:25:08,978 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:25:09,979 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:25:13,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:14,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:18,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:19,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:23,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:26,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:26,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:26,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:25:26,776 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:25:26,776 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:25:26,856 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:25:26,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:25:27,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:25:28,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:29,992 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:25:31,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:33,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:36,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:36,793 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:25:39,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:44,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:44,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:25:44,153 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:25:44,153 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:25:44,236 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:25:45,001 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:25:45,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:46,002 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:25:49,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:50,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:25:54,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:25:55,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:00,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:02,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:02,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:26:02,204 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:26:02,205 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:26:02,288 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:26:03,013 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:26:04,014 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:26:05,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:06,798 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:26:07,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:10,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:12,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:15,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:19,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:19,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:26:19,451 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:26:19,451 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:26:19,532 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:26:20,024 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:26:21,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:22,025 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:26:24,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:26,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:29,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:31,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:36,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:36,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:36,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:26:36,794 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:26:36,795 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:26:36,894 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:26:36,894 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:26:37,035 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:26:38,035 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:26:40,037 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:26:41,786 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:41,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:46,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:46,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:51,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:53,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:26:53,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:26:53,438 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:26:53,439 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:26:53,520 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:26:54,046 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:26:56,047 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:26:57,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:26:58,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:02,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:03,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:06,898 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:27:07,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:10,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:10,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:10,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:27:10,645 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:27:10,646 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:27:10,730 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:27:11,057 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:27:12,057 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:27:13,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:15,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:18,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:23,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:23,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:28,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:29,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:29,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:33,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:36,904 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:27:38,075 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:27:38,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:42,456 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:43,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:48,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:53,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:27:55,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:27:58,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:01,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:28:03,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:06,909 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:28:08,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:11,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:28:13,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:18,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:24,102 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:28:24,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:25,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:28:29,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:30,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:28:34,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:35,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:28:36,915 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:28:39,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:41,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:28:41,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:28:41,961 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:28:41,963 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:28:42,044 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:28:42,114 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:28:44,116 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:28:45,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:46,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:28:50,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:51,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:28:55,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:28:57,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:28:59,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:28:59,975 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:28:59,975 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:29:00,056 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:29:00,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:00,126 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:29:02,128 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:29:02,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:05,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:06,921 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:29:07,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:10,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:14,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:15,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:17,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:29:17,091 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:29:17,091 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:29:17,174 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:29:18,174 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:29:18,174 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:29:20,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:21,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:25,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:26,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:30,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:31,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:35,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:35,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:29:35,387 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:29:35,387 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:29:35,469 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:29:36,186 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:29:36,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:36,927 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:29:38,187 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:29:40,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:41,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:45,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:46,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:50,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:51,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:29:53,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:29:53,177 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:29:53,178 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:29:53,262 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:29:53,263 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:29:54,263 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:29:56,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:29:57,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:01,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:02,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:06,934 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:30:07,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:07,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:07,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:10,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:30:10,991 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:30:10,991 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:30:11,072 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:30:11,275 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:30:12,276 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:30:12,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:13,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:14,277 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:30:17,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:18,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:22,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:23,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:28,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:29,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:29,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:30:29,052 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:30:29,054 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:30:29,135 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:30:29,287 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:30:30,288 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:30:32,289 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:30:33,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:34,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:36,939 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:30:38,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:39,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:43,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:46,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:46,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:30:46,710 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:30:46,711 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:30:46,791 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:30:47,299 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:30:48,300 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:30:49,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:51,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:54,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:30:56,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:30:59,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:03,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:03,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:31:03,889 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:31:03,890 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:31:03,971 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:31:04,310 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:31:04,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:06,313 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:31:06,945 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:31:08,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:10,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:13,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:15,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:20,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:20,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:31:20,809 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:31:20,810 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:31:20,894 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:31:21,341 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:31:21,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:22,342 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:31:25,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:26,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:30,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:31,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:36,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:36,950 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:31:38,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:38,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:38,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:31:38,271 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:31:38,272 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:31:38,352 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:31:38,353 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:31:40,354 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:31:42,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:43,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:47,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:48,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:52,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:55,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:55,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:31:55,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:31:55,693 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:31:55,694 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:31:55,777 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:31:56,365 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:31:57,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:31:58,366 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:32:01,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:02,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:06,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:06,956 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:32:07,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:12,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:13,598 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:13,620 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:32:13,622 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:32:13,624 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:32:13,705 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:32:14,377 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:32:16,378 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:32:18,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:18,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:23,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:23,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:28,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:30,388 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:32:30,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:30,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:32:30,991 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:32:30,992 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:32:31,074 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:32:31,388 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:32:32,389 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:32:34,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:35,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:36,962 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:32:39,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:40,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:44,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:48,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:48,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:32:48,785 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:32:48,786 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:32:48,867 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:32:49,400 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:32:50,401 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:32:50,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:53,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:32:55,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:32:58,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:00,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:04,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:05,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:06,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:33:06,505 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:33:06,505 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:33:06,586 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:33:06,968 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:33:07,412 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:33:08,413 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:33:09,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:10,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:14,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:15,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:19,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:20,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:23,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:33:23,824 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:33:23,825 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:33:23,908 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:33:24,424 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:33:25,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:26,425 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:33:26,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:30,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:31,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:35,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:36,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:36,974 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:33:40,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:41,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:33:41,025 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:33:41,026 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:33:41,108 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:33:41,435 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:33:42,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:42,436 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:33:45,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:47,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:50,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:52,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:57,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:33:58,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:33:58,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:33:58,260 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:33:58,261 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:33:58,347 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:33:58,447 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:34:00,448 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:34:02,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:03,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:06,980 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:34:07,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:08,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:12,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:13,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:16,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:34:16,275 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:34:16,275 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:34:16,357 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:34:16,458 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:34:18,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:18,460 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:34:19,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:23,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:24,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:28,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:30,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:33,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:34,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:34:34,188 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:34:34,189 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:34:34,273 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:34:34,470 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:34:36,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:36,472 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:34:36,985 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:34:38,989 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:41,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:43,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:46,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:48,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:51,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:51,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:34:51,688 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:34:51,689 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:34:51,770 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:34:52,482 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:34:52,483 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:34:54,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:34:56,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:34:59,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:01,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:04,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:06,989 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:35:09,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:09,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:09,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:35:09,721 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:35:09,722 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:35:09,804 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:35:09,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:10,494 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:35:10,495 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:35:14,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:14,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:19,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:19,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:24,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:26,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:26,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:26,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:35:26,725 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:35:26,726 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:35:26,809 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:35:27,505 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:35:28,506 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:35:29,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:31,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:34,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:36,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:36,995 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:35:39,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:41,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:44,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:35:44,692 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:35:44,694 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:35:44,775 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:35:45,517 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:35:45,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:46,518 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:35:47,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:50,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:52,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:55,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:35:58,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:35:58,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:00,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:01,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:36:01,323 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:36:01,324 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:36:01,409 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:36:01,528 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:36:02,528 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:36:04,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:06,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:07,001 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:36:09,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:12,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:14,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:17,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:19,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:19,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:36:19,927 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:36:19,928 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:36:20,014 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:36:20,540 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:36:20,541 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:36:22,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:24,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:27,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:29,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:32,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:36,960 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:36,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:36:36,998 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:36:37,000 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:36:37,099 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:36:37,100 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:36:37,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:37,551 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:36:38,552 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:36:41,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:42,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:46,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:47,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:52,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:54,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:36:54,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:36:54,936 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:36:54,936 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:36:55,018 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:36:55,563 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:36:56,563 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:36:58,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:36:59,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:03,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:04,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:07,093 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:37:08,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:12,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:12,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:37:12,683 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:37:12,685 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:37:12,767 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:37:13,574 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:37:13,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:14,575 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:37:17,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:18,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:22,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:23,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:28,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:29,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:29,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:37:29,919 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:37:29,919 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:37:30,000 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:37:30,585 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:37:30,586 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:37:34,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:34,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:37,099 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:37:39,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:39,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:44,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:44,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:48,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:37:48,125 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:37:48,126 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:37:48,214 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:37:48,597 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:37:48,598 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:37:49,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:50,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:54,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:37:55,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:37:59,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:01,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:04,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:04,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:38:04,971 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:38:04,972 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:38:05,053 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:38:05,608 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:38:06,609 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:38:06,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:07,104 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:38:10,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:11,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:15,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:16,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:20,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:22,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:22,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:38:22,833 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:38:22,833 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:38:22,916 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:38:23,619 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:38:24,620 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:38:25,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:27,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:30,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:32,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:35,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:37,110 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:38:39,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:39,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:38:39,735 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:38:39,736 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:38:39,819 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:38:40,631 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:38:40,631 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:38:41,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:44,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:46,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:49,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:51,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:56,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:38:58,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:38:58,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:38:58,247 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:38:58,248 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:38:58,332 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:38:58,642 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:38:58,643 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:39:00,643 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:39:02,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:03,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:07,117 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:39:08,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:08,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:13,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:15,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:15,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:39:15,132 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:39:15,132 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:39:15,213 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:39:15,653 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:39:16,654 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:39:18,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:20,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:23,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:25,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:28,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:30,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:33,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:33,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:39:33,454 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:39:33,455 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:39:33,536 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:39:33,665 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:39:34,666 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:39:35,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:37,122 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:39:39,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:40,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:44,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:45,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:49,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:50,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:50,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:39:50,831 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:39:50,832 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:39:50,913 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:39:51,677 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:39:52,678 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:39:54,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:39:55,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:39:59,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:00,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:04,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:07,128 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:40:08,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:08,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:40:08,501 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:40:08,502 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:40:08,584 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:40:08,688 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:40:10,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:10,690 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:40:13,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:15,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:18,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:20,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:25,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:25,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:25,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:40:25,955 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:40:25,955 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:40:26,038 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:40:26,700 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:40:26,701 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:40:30,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:31,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:35,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:36,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:37,133 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:40:41,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:42,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:43,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:40:43,017 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:40:43,018 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:40:43,101 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:40:43,711 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:40:44,712 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:40:47,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:47,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:52,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:40:52,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:40:57,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:00,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:00,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:41:00,531 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:41:00,531 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:41:00,612 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:41:00,723 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:41:00,723 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:41:02,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:05,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:07,140 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:41:08,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:10,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:13,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:17,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:17,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:17,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:41:17,445 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:41:17,446 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:41:17,528 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:41:17,734 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:41:18,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:18,734 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:41:23,404 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:23,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:28,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:28,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:33,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:35,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:35,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:35,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:41:35,349 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:41:35,350 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:41:35,432 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:41:35,745 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:41:36,745 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:41:37,145 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:41:37,746 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:41:39,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:40,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:44,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:45,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:49,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:50,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:53,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:41:53,282 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:41:53,283 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:41:53,364 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:41:53,756 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:41:54,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:41:54,757 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:41:55,758 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:41:56,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:41:59,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:01,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:04,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:07,151 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:42:07,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:10,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:10,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:42:10,833 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:42:10,833 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:42:10,915 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:42:11,769 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:42:13,770 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:42:13,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:15,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:18,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:20,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:24,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:25,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:28,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:42:28,449 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:42:28,450 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:42:28,531 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:42:28,780 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:42:29,781 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:42:30,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:31,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:35,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:36,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:37,157 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:42:40,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:42,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:46,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:46,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:42:46,182 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:42:46,183 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:42:46,264 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:42:46,791 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:42:47,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:47,792 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:42:51,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:52,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:42:56,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:42:57,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:02,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:03,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:03,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:43:03,046 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:43:03,046 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:43:03,128 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:43:03,802 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:43:05,804 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:43:07,163 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:43:08,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:08,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:13,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:13,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:18,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:20,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:20,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:20,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:43:20,222 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:43:20,223 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:43:20,304 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:43:20,814 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:43:21,814 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:43:23,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:26,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:28,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:31,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:33,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:37,168 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:43:38,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:38,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:38,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:43:38,383 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:43:38,384 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:43:38,467 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:43:38,827 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:43:39,827 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:43:43,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:43,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:48,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:48,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:53,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:43:55,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:43:55,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:43:55,506 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:43:55,506 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:43:55,588 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:43:55,838 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:43:57,839 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:43:58,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:00,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:03,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:05,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:07,174 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:44:09,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:11,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:13,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:44:13,224 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:44:13,225 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:44:13,306 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:44:13,849 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:44:14,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:15,851 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:44:16,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:19,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:21,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:24,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:27,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:29,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:30,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:44:30,939 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:44:30,941 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:44:31,025 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:44:31,868 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:44:32,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:33,870 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:44:35,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:37,180 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:44:37,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:40,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:42,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:45,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:48,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:48,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:48,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:44:48,265 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:44:48,266 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:44:48,347 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:44:48,898 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:44:49,898 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:44:50,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:53,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:44:55,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:44:58,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:00,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:05,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:05,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:05,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:45:05,616 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:45:05,616 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:45:05,697 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:45:05,908 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:45:07,186 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:45:07,910 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:45:10,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:11,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:15,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:16,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:21,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:22,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:22,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:22,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:45:22,433 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:45:22,434 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:45:22,518 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:45:22,920 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:45:23,920 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:45:26,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:27,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:31,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:32,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:36,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:37,192 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:45:40,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:40,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:40,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:45:40,814 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:45:40,815 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:45:40,896 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:45:40,932 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:45:41,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:41,932 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:45:46,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:46,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:51,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:51,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:56,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:45:57,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:57,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:45:57,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:45:57,272 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:45:57,273 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:45:57,354 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:45:57,943 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:45:59,944 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:46:02,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:02,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:07,227 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:46:07,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:07,949 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:46:08,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:12,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:12,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:13,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:15,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:46:15,856 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:46:15,856 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:46:15,938 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:46:15,955 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:46:17,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:17,956 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:46:18,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:22,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:23,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:27,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:28,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:32,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:46:32,816 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:46:32,816 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:46:32,897 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:46:32,966 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:46:33,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:33,967 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:46:34,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:37,210 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:46:38,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:40,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:43,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:45,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:50,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:50,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:46:50,142 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:46:50,142 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:46:50,225 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:46:50,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:46:50,978 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:46:51,979 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:46:55,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:46:55,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:00,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:00,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:05,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:07,216 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:47:07,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:07,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:07,511 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:47:07,512 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:47:07,592 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:47:07,989 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:47:09,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:47:10,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:12,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:15,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:17,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:20,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:24,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:24,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:24,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:24,446 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:47:24,446 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:47:24,528 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:47:25,000 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:47:26,001 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:47:26,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:29,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:31,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:34,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:36,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:37,222 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:47:39,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:41,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:41,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:41,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:41,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:42,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:42,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:44,436 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:45,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:47,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:47,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:47,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:47,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:47,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:48,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:48,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:48,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:49,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:51,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:52,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:53,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:53,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:53,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:54,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:54,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:54,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:54,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:54,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:57,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:47:58,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:59,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:47:59,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:59,953 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:47:59,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:00,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:00,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:00,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:01,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:02,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:04,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:04,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:06,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:06,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:06,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:06,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:06,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:06,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:07,227 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:48:07,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:07,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:09,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:10,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:12,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:12,456 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:12,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:12,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:12,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:13,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:13,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:13,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:14,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:16,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:17,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:18,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:18,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:18,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:19,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:19,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:19,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:19,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:19,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:22,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:23,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:24,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:24,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:24,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:24,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:25,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:25,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:25,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:26,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:27,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:29,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:29,560 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:31,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:31,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:31,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:31,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:31,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:31,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:32,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:32,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:34,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:35,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:37,233 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:48:37,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:37,560 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:37,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:37,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:38,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:38,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:38,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:38,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:39,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:42,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:42,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:43,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:43,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:43,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:44,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:44,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:44,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:44,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:44,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:47,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:48,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:49,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:50,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:50,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:50,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:50,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:50,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:50,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:51,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:52,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:54,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:48:54,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:56,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:56,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:56,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:56,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:56,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:57,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:57,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:48:57,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:48:59,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:00,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:02,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:02,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:02,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:02,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:03,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:03,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:03,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:03,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:04,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:07,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:07,239 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:49:07,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:08,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:08,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:08,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:09,363 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:09,456 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:09,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:09,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:10,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:12,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:13,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:14,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:15,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:15,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:15,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:15,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:15,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:15,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:16,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:17,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:19,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:19,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:21,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:21,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:21,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:21,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:21,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:22,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:22,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:22,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:24,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:25,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:27,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:27,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:27,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:27,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:28,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:28,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:28,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:28,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:29,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:32,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:32,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:33,927 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:33,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:33,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:34,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:34,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:34,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:34,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:35,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:37,245 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:49:37,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:38,417 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:39,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:40,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:40,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:40,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:40,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:40,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:40,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:41,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:42,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:44,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:44,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:46,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:46,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:46,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:46,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:46,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:47,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:47,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:47,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:49,467 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:50,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:52,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:52,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:52,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:52,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:53,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:53,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:53,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:53,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:54,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:57,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:57,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:49:58,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:58,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:58,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:59,375 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:59,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:49:59,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:59,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:49:59,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:02,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:03,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:04,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:05,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:05,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:05,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:05,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:05,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:05,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:06,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:07,250 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:50:07,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:09,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:09,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:11,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:11,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:11,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:11,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:11,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:12,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:12,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:12,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:14,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:15,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:17,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:17,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:17,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:17,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:18,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:18,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:18,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:18,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:19,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:22,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:22,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:23,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:23,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:23,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:24,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:24,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:24,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:24,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:25,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:27,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:28,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:29,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:30,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:30,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:30,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:30,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:30,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:30,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:31,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:32,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:34,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:34,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:36,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:36,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:36,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:36,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:36,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:37,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:37,256 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:50:37,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:37,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:39,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:40,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:42,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:42,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:42,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:42,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:43,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:43,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:43,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:43,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:44,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:47,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:47,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:48,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:48,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:48,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:49,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:49,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:49,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:49,620 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:50,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:52,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:53,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:54,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:55,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:55,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:55,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:55,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:55,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:55,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:56,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:50:57,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:50:59,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:50:59,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:01,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:01,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:01,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:01,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:02,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:04,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:07,263 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:51:07,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:09,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:12,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:14,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:17,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:17,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:19,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:22,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:24,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:27,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:29,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:32,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:32,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:32,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:34,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:34,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:37,268 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:51:37,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:37,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:37,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:38,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:38,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:38,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:38,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:38,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:39,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:40,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:42,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:43,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:43,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:44,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:44,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:44,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:44,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:44,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:44,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:46,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:47,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:49,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:50,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:50,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:50,555 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:50,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:50,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:51,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:51,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:52,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:52,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:54,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:51:56,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:56,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:56,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:56,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:56,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:57,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:51:57,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:57,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:59,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:51:59,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:02,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:02,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:02,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:03,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:03,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:03,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:03,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:03,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:04,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:05,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:07,274 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:52:07,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:08,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:08,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:09,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:09,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:09,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:09,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:09,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:09,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:11,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:12,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:14,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:14,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:15,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:15,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:15,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:15,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:16,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:16,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:17,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:17,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:19,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:21,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:21,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:21,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:21,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:21,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:22,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:22,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:22,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:24,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:24,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:27,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:27,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:27,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:27,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:28,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:28,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:28,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:28,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:29,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:30,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:32,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:33,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:33,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:34,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:34,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:34,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:34,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:34,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:34,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:36,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:37,279 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:52:37,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:39,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:40,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:40,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:40,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:40,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:40,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:41,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:41,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:42,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:42,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:44,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:46,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:46,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:46,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:46,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:46,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:47,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:47,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:47,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:49,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:49,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:52,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:52,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:52,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:52,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:53,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:53,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:53,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:53,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:54,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:55,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:57,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:52:58,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:58,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:59,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:59,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:59,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:59,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:52:59,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:52:59,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:01,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:02,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:04,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:04,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:05,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:05,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:05,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:05,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:06,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:06,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:07,285 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:53:07,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:07,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:09,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:11,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:11,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:11,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:11,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:11,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:12,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:12,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:12,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:14,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:14,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:17,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:17,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:17,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:17,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:18,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:18,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:18,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:18,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:19,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:20,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:22,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:23,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:23,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:24,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:24,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:24,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:24,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:24,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:24,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:26,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:27,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:29,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:29,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:30,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:30,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:30,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:30,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:31,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:31,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:32,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:32,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:34,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:36,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:36,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:36,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:36,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:36,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:37,291 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:53:37,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:37,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:37,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:39,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:39,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:42,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:42,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:42,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:42,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:43,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:43,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:43,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:43,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:44,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:45,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:47,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:48,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:48,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:49,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:49,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:49,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:49,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:49,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:49,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:51,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:52,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:54,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:53:54,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:54,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:55,404 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:55,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:55,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:56,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:56,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:57,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:53:57,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:53:59,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:01,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:01,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:01,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:01,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:01,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:02,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:02,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:02,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:04,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:04,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:07,297 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:54:07,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:07,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:07,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:07,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:08,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:08,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:08,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:08,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:09,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:10,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:12,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:13,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:13,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:14,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:14,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:14,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:14,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:14,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:14,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:16,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:17,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:19,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:19,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:19,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:20,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:20,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:20,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:21,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:21,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:22,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:22,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:24,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:26,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:26,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:26,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:26,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:26,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:27,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:27,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:27,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:29,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:29,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:32,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:32,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:32,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:33,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:33,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:33,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:33,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:33,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:34,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:35,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:37,302 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:54:37,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:38,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:38,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:39,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:39,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:39,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:39,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:39,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:39,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:41,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:42,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:44,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:45,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:45,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:45,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:45,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:45,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:46,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:46,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:47,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:48,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:49,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:51,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:51,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:51,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:51,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:52,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:52,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:52,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:52,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:54:54,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:54:57,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:54:59,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:02,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:04,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:07,308 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:55:07,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:07,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:55:11,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:11,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:12,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:14,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:55:15,027 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:55:15,127 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:55:15,186 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:55:15,317 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:55:16,317 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:55:16,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:18,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:21,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:23,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:26,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:28,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:32,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:32,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:55:32,750 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:55:32,751 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:55:32,833 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:55:33,328 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:55:33,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:34,328 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:55:37,314 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:55:37,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:39,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:42,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:44,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:49,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:49,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:49,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:55:49,576 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:55:49,576 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:55:49,658 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:55:50,350 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:55:50,350 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:55:54,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:54,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:55:59,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:55:59,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:04,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:07,320 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:56:07,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:07,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:56:07,602 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:56:07,603 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:56:07,685 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:56:08,361 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:56:08,362 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:56:09,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:12,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:14,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:17,598 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:19,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:24,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:24,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:56:24,393 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:56:24,393 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:56:24,477 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:56:25,386 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:56:25,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:26,387 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:56:29,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:30,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:34,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:35,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:37,326 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:56:41,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:42,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:42,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:42,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:56:42,283 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:56:42,283 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:56:42,364 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:56:42,411 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:56:44,413 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:56:46,366 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:47,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:51,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:52,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:56,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:56:59,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:59,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:56:59,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:56:59,850 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:56:59,850 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:56:59,932 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:57:00,440 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:57:00,441 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:57:01,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:02,442 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:57:04,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:06,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:07,331 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:57:09,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:12,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:17,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:17,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:17,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:57:17,094 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:57:17,095 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:57:17,176 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:57:17,467 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:57:18,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:18,471 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:57:22,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:23,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:27,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:28,179 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:33,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:34,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:34,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:34,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:57:34,630 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:57:34,631 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:57:34,712 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:57:35,498 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:57:36,499 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:57:37,337 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:57:38,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:39,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:43,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:44,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:48,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:51,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:51,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:57:51,157 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:57:51,158 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:57:51,240 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:57:51,518 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:57:52,519 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:57:54,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:57:56,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:57:59,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:01,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:04,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:07,343 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:58:08,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:08,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:58:08,796 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:58:08,797 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:58:08,879 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:58:09,529 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:58:09,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:10,530 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:58:13,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:14,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:18,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:19,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:24,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:25,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:25,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:58:25,893 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:58:25,894 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:58:25,976 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:58:26,540 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:58:26,540 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:58:29,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:30,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:34,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:35,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:37,349 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:58:40,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:41,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:43,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:58:43,814 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:58:43,814 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:58:43,896 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:58:44,551 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:58:44,552 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:58:45,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:46,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:50,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:51,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:58:55,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:58:57,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:00,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:01,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:59:01,444 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:59:01,444 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:59:01,526 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:59:01,562 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:59:02,563 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:59:03,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:03,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:06,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:07,355 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:59:08,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:12,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:13,436 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:17,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:18,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:18,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:18,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:59:18,992 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:59:18,993 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:59:19,075 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:59:19,573 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:59:20,574 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:59:23,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:23,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:28,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:28,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:33,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:35,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:35,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:35,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:59:35,781 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:59:35,783 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:59:35,864 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:59:36,584 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:59:36,585 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:59:37,360 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 18:59:38,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:40,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:43,363 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:45,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:48,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:52,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:52,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 18:59:52,511 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 18:59:52,512 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 18:59:52,593 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 18:59:52,595 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 18:59:53,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 18:59:54,596 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 18:59:57,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 18:59:58,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:02,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:03,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:07,367 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:00:09,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:09,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:09,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:00:09,978 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:00:09,978 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:00:10,060 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:00:10,606 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:00:10,606 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:00:14,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:15,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:19,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:20,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:25,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:27,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:27,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:00:27,468 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:00:27,469 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:00:27,552 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:00:27,617 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:00:28,617 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:00:30,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:32,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:35,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:37,372 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:00:37,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:41,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:43,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:45,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:00:45,670 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:00:45,670 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:00:45,752 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:00:46,629 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:00:46,629 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:00:46,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:48,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:51,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:53,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:00:56,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:00:59,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:01,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:03,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:01:03,280 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:01:03,280 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:01:03,361 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:01:03,640 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:01:04,640 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:01:05,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:05,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:07,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:07,378 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:01:10,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:12,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:15,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:17,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:20,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:20,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:20,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:01:20,859 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:01:20,859 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:01:20,942 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:01:21,651 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:01:22,652 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:01:22,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:25,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:27,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:30,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:32,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:37,384 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:01:37,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:37,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:37,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:01:37,644 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:01:37,645 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:01:37,728 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:01:38,727 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:01:38,728 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:01:38,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:42,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:43,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:47,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:48,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:53,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:01:54,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:54,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:54,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:01:54,599 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:01:54,599 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:01:54,682 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:01:54,738 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:01:56,739 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:01:59,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:01:59,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:04,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:04,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:07,390 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:02:09,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:10,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:11,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:02:11,771 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:02:11,772 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:02:11,853 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:02:12,749 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:02:12,750 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:02:14,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:15,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:19,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:20,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:25,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:25,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:25,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:29,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:02:29,149 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:02:29,149 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:02:29,231 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:02:29,760 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:02:30,761 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:02:31,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:31,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:36,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:36,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:37,395 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:02:41,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:41,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:46,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:47,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:47,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:02:47,532 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:02:47,533 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:02:47,616 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:02:47,771 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:02:48,774 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:02:51,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:52,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:02:56,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:02:57,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:01,620 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:04,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:04,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:03:04,855 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:03:04,856 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:03:04,937 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:03:05,785 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:03:06,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:07,401 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:03:07,787 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:03:09,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:12,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:14,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:17,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:22,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:22,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:22,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:03:22,588 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:03:22,588 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:03:22,671 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:03:22,796 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:03:23,797 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:03:25,798 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:03:27,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:27,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:32,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:32,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:37,407 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:03:38,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:39,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:39,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:03:39,458 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:03:39,459 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:03:39,540 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:03:39,807 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:03:41,808 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:03:43,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:44,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:48,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:49,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:53,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:56,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:03:56,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:03:56,596 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:03:56,597 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:03:56,679 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:03:56,818 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:03:57,818 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:03:58,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:03:59,820 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:04:01,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:03,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:06,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:07,413 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:04:09,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:13,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:13,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:04:13,820 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:04:13,821 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:04:13,903 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:04:14,902 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:04:14,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:15,903 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:04:18,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:19,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:23,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:24,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:29,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:31,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:31,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:31,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:04:31,245 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:04:31,245 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:04:31,330 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:04:31,913 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:04:33,915 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:04:35,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:36,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:37,418 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:04:40,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:41,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:45,422 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:49,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:49,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:49,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:04:49,479 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:04:49,479 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:04:49,560 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:04:49,924 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:04:50,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:51,926 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:04:54,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:04:55,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:04:59,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:00,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:05,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:06,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:06,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:06,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:05:06,215 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:05:06,216 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:05:06,297 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:05:06,935 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:05:07,424 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:05:07,936 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:05:11,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:11,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:16,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:16,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:21,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:24,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:24,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:05:24,521 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:05:24,521 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:05:24,603 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:05:24,946 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:05:25,947 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:05:26,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:29,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:31,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:34,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:36,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:37,431 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:05:40,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:40,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:05:40,995 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:05:40,996 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:05:41,079 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:05:41,957 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:05:42,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:43,958 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:05:45,989 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:47,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:50,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:52,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:57,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:05:58,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:05:58,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:05:58,479 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:05:58,480 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:05:58,561 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:05:58,968 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:05:59,969 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:06:02,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:03,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:07,436 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:06:08,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:08,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:13,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:15,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:15,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:06:15,498 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:06:15,498 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:06:15,580 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:06:15,979 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:06:17,980 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:06:18,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:20,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:23,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:25,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:28,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:32,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:32,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:32,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:06:32,656 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:06:32,656 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:06:32,737 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:06:32,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:06:33,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:33,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:06:37,442 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:06:37,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:39,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:42,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:44,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:49,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:51,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:51,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:06:51,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:06:51,122 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:06:51,123 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:06:51,206 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:06:52,002 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:06:54,003 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:06:55,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:06:56,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:00,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:01,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:05,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:07,448 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:07:08,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:08,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:07:08,013 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:07:08,013 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:07:08,096 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:07:08,097 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:07:10,097 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:07:11,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:13,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:16,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:18,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:21,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:26,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:26,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:26,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:07:26,451 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:07:26,452 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:07:26,533 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:07:27,108 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:07:28,108 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:07:31,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:31,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:36,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:36,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:37,453 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:07:42,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:42,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:42,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:07:42,971 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:07:42,972 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:07:43,053 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:07:43,118 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:07:44,119 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:07:46,120 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:07:47,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:48,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:52,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:07:53,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:07:58,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:00,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:00,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:08:00,685 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:08:00,686 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:08:00,767 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:08:01,129 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:08:02,130 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:08:03,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:05,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:07,460 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:08:09,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:10,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:14,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:17,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:17,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:08:17,577 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:08:17,577 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:08:17,663 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:08:18,140 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:08:19,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:20,141 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:08:22,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:24,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:27,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:29,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:34,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:34,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:08:34,613 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:08:34,614 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:08:34,696 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:08:34,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:35,151 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:08:36,152 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:08:37,465 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:08:39,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:40,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:44,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:45,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:50,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:52,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:52,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:08:52,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:08:52,777 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:08:52,778 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:08:52,860 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:08:53,163 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:08:54,163 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:08:55,861 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:08:57,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:00,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:02,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:05,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:07,467 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:09:09,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:09,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:09,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:09:09,552 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:09:09,553 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:09:09,635 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:09:10,173 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:09:11,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:12,175 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:09:14,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:16,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:19,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:21,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:24,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:24,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:26,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:28,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:09:28,181 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:09:28,181 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:09:28,262 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:09:28,264 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:09:30,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:30,263 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:09:32,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:35,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:37,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:37,473 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:09:40,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:40,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:42,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:44,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:09:44,638 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:09:44,638 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:09:44,720 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:09:45,272 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:09:46,273 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:09:46,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:47,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:51,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:52,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:09:56,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:09:57,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:02,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:02,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:10:02,264 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:10:02,264 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:10:02,346 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:10:02,347 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:10:03,347 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:10:03,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:04,347 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:10:07,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:07,479 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:10:08,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:12,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:13,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:18,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:19,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:19,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:10:19,352 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:10:19,353 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:10:19,435 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:10:20,435 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:10:20,435 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:10:24,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:24,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:29,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:29,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:34,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:36,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:36,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:36,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:10:36,406 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:10:36,406 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:10:36,489 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:10:37,485 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:10:37,501 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:10:38,501 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:10:39,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:41,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:44,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:46,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:49,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:54,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:54,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:10:54,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:10:54,385 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:10:54,386 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:10:54,468 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:10:54,511 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:10:55,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:10:56,513 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:10:59,378 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:00,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:04,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:05,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:07,491 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:11:10,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:11,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:11,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:11,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:11:11,209 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:11:11,209 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:11:11,291 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:11:11,522 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:11:12,523 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:11:16,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:16,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:21,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:21,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:26,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:26,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:26,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:29,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:11:29,621 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:11:29,622 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:11:29,705 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:11:30,534 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:11:31,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:31,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:32,536 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:11:36,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:36,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:37,496 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:11:41,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:42,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:46,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:46,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:11:46,747 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:11:46,747 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:11:46,829 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:11:47,545 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:11:47,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:48,546 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:11:51,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:52,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:11:56,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:11:57,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:02,598 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:02,832 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:04,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:12:04,750 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:12:04,751 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:12:04,831 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:12:05,557 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:12:06,557 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:12:07,502 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:12:07,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:08,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:12,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:13,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:18,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:18,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:21,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:12:21,857 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:12:21,857 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:12:21,940 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:12:22,567 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:12:23,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:23,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:24,569 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:12:28,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:28,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:33,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:33,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:37,508 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:12:39,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:39,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:39,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:39,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:12:39,036 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:12:39,037 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:12:39,117 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:12:39,578 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:12:40,579 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:12:44,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:44,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:49,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:49,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:54,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:12:56,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:56,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:12:56,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:12:56,135 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:12:56,135 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:12:56,173 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:12:56,589 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:12:58,590 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:12:59,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:01,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:04,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:06,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:07,514 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:13:09,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:13,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:13,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:13,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:13:13,527 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:13:13,527 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:13:13,609 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:13:14,609 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:13:14,609 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:13:14,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:18,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:19,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:23,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:24,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:28,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:28,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:29,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:31,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:13:31,700 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:13:31,701 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:13:31,783 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:13:32,621 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:13:32,621 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:13:33,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:34,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:37,520 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:13:38,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:40,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:43,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:45,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:48,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:48,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:13:48,901 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:13:48,902 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:13:48,988 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:13:49,632 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:13:50,632 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:13:50,989 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:53,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:13:55,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:13:58,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:00,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:05,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:07,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:07,328 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:14:07,331 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:14:07,333 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:14:07,416 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:14:07,524 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:14:07,643 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:14:08,644 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:14:11,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:12,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:16,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:17,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:21,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:24,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:24,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:14:24,233 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:14:24,234 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:14:24,315 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:14:24,654 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:14:26,655 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:14:27,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:29,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:32,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:34,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:37,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:37,530 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:14:42,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:42,030 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:14:42,035 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:14:42,036 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:14:42,117 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:14:42,665 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:14:43,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:44,666 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:14:47,026 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:48,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:52,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:53,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:58,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:14:58,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:58,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:14:58,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:14:58,663 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:14:58,664 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:14:58,737 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:14:59,736 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:15:00,737 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:15:03,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:03,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:07,535 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:15:08,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:09,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:14,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:16,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:16,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:15:16,052 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:15:16,052 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:15:16,135 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:15:16,742 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:15:18,743 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:15:20,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:21,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:25,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:26,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:30,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:33,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:33,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:33,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:15:33,570 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:15:33,571 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:15:33,652 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:15:33,749 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:15:34,750 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:15:35,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:37,541 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:15:38,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:41,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:43,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:46,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:50,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:50,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:50,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:15:50,561 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:15:50,561 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:15:50,646 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:15:50,756 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:15:51,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:15:52,757 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:15:55,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:15:56,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:00,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:01,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:06,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:07,547 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:16:09,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:09,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:16:09,061 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:16:09,062 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:16:09,146 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:16:09,764 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:16:10,764 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:16:12,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:14,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:17,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:19,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:22,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:25,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:25,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:16:25,616 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:16:25,617 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:16:25,699 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:16:25,770 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:16:26,771 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:16:27,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:30,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:32,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:35,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:37,553 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:16:38,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:43,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:43,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:43,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:16:43,883 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:16:43,884 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:16:43,965 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:16:44,778 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:16:44,779 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:16:48,874 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:48,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:53,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:16:53,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:16:58,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:00,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:00,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:17:00,349 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:17:00,349 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:17:00,432 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:17:00,793 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:17:02,794 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:17:04,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:05,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:07,559 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:17:09,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:10,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:14,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:17,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:17,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:17,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:17:17,803 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:17:17,804 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:17:17,884 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:17:18,800 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:17:18,801 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:17:19,886 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:22,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:24,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:27,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:29,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:34,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:35,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:35,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:35,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:17:35,242 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:17:35,242 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:17:35,324 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:17:35,806 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:17:36,807 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:17:37,564 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:17:40,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:40,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:45,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:45,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:50,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:52,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:52,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:17:52,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:17:52,109 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:17:52,110 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:17:52,193 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:17:52,813 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:17:52,814 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:17:56,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:17:57,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:01,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:02,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:06,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:07,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:07,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:07,567 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:18:10,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:18:10,972 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:18:10,973 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:18:11,058 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:18:11,822 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:18:12,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:12,822 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:18:12,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:17,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:17,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:22,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:24,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:24,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:27,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:27,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:18:27,447 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:18:27,449 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:18:27,529 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:18:27,829 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:18:28,829 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:18:29,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:32,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:34,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:37,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:37,572 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:18:39,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:42,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:45,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:45,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:18:45,950 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:18:45,950 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:18:46,032 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:18:46,836 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:18:46,837 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:18:48,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:50,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:53,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:18:55,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:18:58,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:02,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:02,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:19:02,234 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:19:02,234 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:19:02,318 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:19:02,843 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:19:02,843 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:19:03,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:07,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:07,579 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:19:08,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:12,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:13,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:18,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:19,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:19,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:19:19,323 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:19:19,323 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:19:19,406 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:19:19,853 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:19:20,854 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:19:24,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:24,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:29,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:29,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:34,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:36,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:36,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:19:36,860 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:19:36,860 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:19:36,942 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:19:36,943 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:19:37,584 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:19:37,943 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:19:38,943 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:19:39,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:41,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:44,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:46,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:49,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:53,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:53,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:53,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:19:53,755 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:19:53,756 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:19:53,836 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:19:53,953 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:19:54,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:19:54,954 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:19:58,749 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:19:59,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:03,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:04,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:07,590 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:20:09,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:09,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:10,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:12,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:20:12,975 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:20:12,975 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:20:13,058 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:20:13,966 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:20:14,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:14,967 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:20:16,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:19,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:21,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:26,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:26,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:29,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:20:29,505 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:20:29,505 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:20:29,589 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:20:29,977 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:20:30,977 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:20:31,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:31,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:36,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:36,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:37,595 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:20:41,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:41,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:46,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:48,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:48,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:20:48,343 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:20:48,343 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:20:48,424 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:20:48,989 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:20:48,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:20:52,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:53,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:20:57,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:20:58,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:02,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:04,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:04,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:21:04,656 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:21:04,657 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:21:04,741 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:21:04,999 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:21:07,001 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:21:07,601 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:21:07,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:09,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:12,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:14,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:17,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:21,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:21,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:21:21,344 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:21:21,345 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:21:21,426 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:21:22,011 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:21:23,011 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:21:23,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:26,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:28,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:31,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:33,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:37,607 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:21:38,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:38,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:38,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:21:38,927 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:21:38,928 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:21:39,009 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:21:39,022 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:21:41,023 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:21:43,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:44,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:48,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:49,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:54,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:21:55,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:55,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:21:55,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:21:55,742 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:21:55,742 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:21:55,825 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:21:56,033 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:21:57,033 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:21:59,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:00,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:04,832 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:05,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:07,613 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:22:10,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:10,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:10,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:14,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:22:14,838 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:22:14,839 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:22:14,923 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:22:15,045 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:22:15,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:16,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:17,046 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:22:20,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:21,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:25,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:28,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:28,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:30,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:31,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:22:31,526 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:22:31,526 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:22:31,608 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:22:32,056 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:22:33,056 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:22:33,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:36,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:37,619 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:22:38,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:41,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:43,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:46,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:49,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:50,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:22:50,005 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:22:50,005 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:22:50,088 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:22:51,087 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:22:51,088 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:22:52,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:54,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:22:57,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:22:59,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:02,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:06,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:06,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:23:06,613 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:23:06,613 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:23:06,695 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:23:07,098 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:23:07,098 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:23:07,624 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:23:07,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:11,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:12,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:16,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:17,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:22,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:23,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:23,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:23:23,417 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:23:23,418 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:23:23,500 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:23:24,108 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:23:25,109 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:23:28,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:28,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:33,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:33,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:37,631 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:23:38,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:41,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:41,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:41,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:23:41,219 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:23:41,219 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:23:41,299 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:23:42,120 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:23:43,120 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:23:44,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:46,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:49,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:51,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:54,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:23:58,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:58,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:23:58,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:23:58,154 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:23:58,155 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:23:58,236 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:23:59,130 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:23:59,131 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:24:00,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:03,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:05,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:07,636 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:24:08,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:10,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:13,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:14,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:15,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:15,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:15,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:15,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:15,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:18,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:19,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:20,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:21,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:21,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:21,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:21,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:21,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:21,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:22,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:23,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:25,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:25,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:27,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:27,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:27,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:27,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:27,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:27,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:28,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:28,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:30,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:31,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:33,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:33,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:33,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:33,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:33,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:34,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:34,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:34,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:35,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:37,642 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:24:38,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:38,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:39,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:39,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:39,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:40,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:40,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:40,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:40,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:40,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:43,157 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:44,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:45,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:45,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:45,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:45,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:46,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:46,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:46,643 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:47,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:48,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:50,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:50,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:52,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:52,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:52,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:52,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:52,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:52,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:53,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:53,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:55,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:24:56,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:58,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:24:58,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:58,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:58,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:58,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:59,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:59,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:24:59,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:00,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:03,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:03,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:04,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:04,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:04,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:05,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:05,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:05,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:05,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:05,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:07,648 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:25:08,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:09,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:10,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:10,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:10,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:10,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:11,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:11,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:11,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:12,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:13,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:15,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:15,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:17,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:17,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:17,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:17,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:17,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:17,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:18,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:18,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:20,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:21,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:23,167 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:23,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:23,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:23,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:23,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:24,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:24,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:24,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:25,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:28,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:28,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:29,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:29,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:29,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:30,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:30,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:30,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:30,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:30,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:33,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:34,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:35,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:35,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:35,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:35,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:36,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:36,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:36,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:37,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:37,654 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:25:38,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:40,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:40,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:42,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:42,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:42,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:42,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:42,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:42,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:43,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:43,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:45,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:46,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:48,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:48,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:48,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:48,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:48,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:49,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:49,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:49,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:50,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:52,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:53,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:54,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:54,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:54,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:55,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:55,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:55,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:55,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:25:55,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:25:58,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:25:59,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:00,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:00,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:00,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:00,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:01,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:01,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:01,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:02,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:03,177 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:05,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:05,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:07,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:07,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:07,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:07,660 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:26:07,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:07,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:07,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:08,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:08,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:10,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:11,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:13,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:13,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:13,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:13,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:13,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:14,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:14,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:14,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:15,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:18,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:18,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:19,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:19,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:19,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:20,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:20,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:20,379 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:20,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:20,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:23,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:24,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:25,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:25,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:25,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:25,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:26,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:26,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:26,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:27,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:28,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:30,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:30,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:32,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:32,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:32,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:32,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:32,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:32,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:33,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:33,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:35,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:36,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:37,666 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:26:38,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:38,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:38,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:38,467 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:38,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:39,026 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:39,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:39,560 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:40,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:43,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:43,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:44,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:44,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:44,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:45,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:45,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:45,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:45,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:45,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:48,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:49,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:50,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:50,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:50,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:50,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:51,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:51,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:51,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:52,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:53,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:55,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:55,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:26:57,257 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:57,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:57,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:57,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:57,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:57,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:26:58,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:26:58,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:00,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:01,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:03,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:03,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:03,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:03,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:03,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:04,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:04,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:04,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:05,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:07,672 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:27:08,081 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:08,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:09,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:09,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:09,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:10,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:10,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:10,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:10,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:10,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:13,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:14,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:15,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:16,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:16,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:16,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:16,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:16,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:16,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:17,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:18,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:20,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:20,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:22,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:22,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:22,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:22,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:22,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:22,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:23,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:23,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:25,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:26,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:28,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:28,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:28,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:28,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:29,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:29,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:29,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:29,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:30,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:33,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:33,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:34,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:34,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:34,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:34,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:35,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:37,677 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:27:38,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:40,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:43,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:45,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:48,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:50,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:27:50,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:53,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:27:55,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:27:58,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:00,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:03,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:05,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:05,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:05,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:07,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:07,683 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:28:08,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:10,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:10,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:10,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:11,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:11,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:11,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:12,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:12,049 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:13,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:13,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:15,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:17,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:17,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:17,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:17,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:17,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:18,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:18,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:18,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:19,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:20,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:23,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:23,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:23,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:23,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:23,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:24,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:24,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:24,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:25,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:26,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:28,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:29,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:29,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:30,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:30,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:30,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:30,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:30,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:30,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:32,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:33,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:35,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:35,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:35,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:36,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:36,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:36,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:37,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:37,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:37,689 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:28:38,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:38,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:40,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:42,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:42,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:42,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:42,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:42,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:43,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:43,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:43,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:44,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:45,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:48,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:48,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:48,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:48,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:48,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:49,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:49,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:49,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:50,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:51,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:53,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:28:54,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:54,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:55,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:55,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:55,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:55,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:55,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:55,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:28:57,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:28:58,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:00,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:00,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:00,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:01,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:01,358 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:01,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:01,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:01,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:03,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:03,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:05,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:07,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:07,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:07,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:07,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:07,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:07,695 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:29:08,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:08,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:08,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:09,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:10,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:13,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:13,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:13,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:13,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:13,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:13,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:14,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:14,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:15,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:16,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:18,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:19,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:19,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:19,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:20,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:20,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:20,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:20,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:20,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:22,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:23,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:25,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:25,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:25,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:26,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:26,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:26,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:26,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:26,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:28,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:28,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:30,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:31,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:32,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:32,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:32,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:32,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:33,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:33,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:33,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:34,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:35,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:37,701 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:29:38,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:38,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:38,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:38,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:38,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:38,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:39,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:39,352 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:40,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:41,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:43,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:44,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:44,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:44,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:45,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:45,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:45,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:45,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:45,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:47,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:48,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:50,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:50,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:50,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:51,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:51,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:51,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:51,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:51,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:53,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:53,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:55,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:29:56,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:56,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:57,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:57,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:57,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:58,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:58,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:29:58,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:29:59,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:00,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:03,223 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:03,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:03,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:03,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:03,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:03,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:04,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:04,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:05,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:06,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:07,706 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:30:08,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:09,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:09,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:09,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:10,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:10,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:10,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:10,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:10,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:12,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:13,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:15,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:15,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:15,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:16,192 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:16,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:16,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:16,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:16,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:18,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:18,599 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:20,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:22,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:22,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:22,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:22,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:22,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:23,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:23,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:23,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:24,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:25,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:28,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:28,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:28,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:28,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:28,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:28,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:29,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:29,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:30,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:31,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:33,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:34,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:34,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:35,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:35,204 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:35,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:35,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:35,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:35,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:37,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:37,712 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:30:38,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:40,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:40,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:40,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:41,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:41,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:41,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:41,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:41,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:43,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:43,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:45,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:47,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:47,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:47,620 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:47,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:47,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:48,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:48,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:48,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:50,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:50,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:53,249 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:53,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:53,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:53,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:53,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:54,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:54,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:54,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:55,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:30:56,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:58,250 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:30:59,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:30:59,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:00,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:00,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:00,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:00,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:00,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:00,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:02,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:03,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:05,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:05,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:05,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:06,425 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:06,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:06,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:07,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:07,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:07,718 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:31:08,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:08,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:10,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:12,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:12,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:12,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:12,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:12,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:13,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:13,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:13,375 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:15,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:15,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:18,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:18,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:18,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:18,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:19,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:19,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:19,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:19,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:20,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:21,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:23,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:24,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:24,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:25,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:25,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:25,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:25,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:25,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:25,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:28,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:30,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:33,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:35,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:37,724 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:31:38,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:40,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:41,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:44,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:44,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:45,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:48,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:31:48,715 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:31:48,817 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:31:48,869 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:31:49,478 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:31:50,478 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:31:50,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:50,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:31:55,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:31:55,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:00,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:00,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:05,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:32:05,488 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:32:05,488 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:32:05,570 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:32:06,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:06,570 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:32:06,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:07,728 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:32:08,572 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:32:11,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:11,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:16,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:16,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:21,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:22,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:22,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:32:22,342 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:32:22,342 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:32:22,425 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:32:22,580 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:32:24,582 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:32:27,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:27,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:32,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:32,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:37,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:37,734 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:32:38,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:38,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:32:38,953 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:32:38,953 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:32:39,037 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:32:39,591 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:32:40,592 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:32:43,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:43,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:48,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:48,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:53,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:56,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:32:56,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:32:56,260 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:32:56,260 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:32:56,341 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:32:56,602 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:32:58,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:32:58,603 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:33:01,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:03,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:06,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:07,740 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:33:08,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:13,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:14,375 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:14,381 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:33:14,384 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:33:14,385 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:33:14,468 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:33:14,614 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:33:16,615 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:33:19,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:19,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:24,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:24,471 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:29,472 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:29,749 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:32,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:33:32,125 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:33:32,126 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:33:32,208 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:33:32,625 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:33:34,627 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:33:35,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:35,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:37,746 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:33:40,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:40,749 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:45,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:45,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:49,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:33:49,892 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:33:49,892 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:33:49,974 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:33:50,637 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:33:50,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:51,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:33:52,638 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:33:55,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:33:56,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:00,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:01,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:05,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:07,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:07,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:34:07,401 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:34:07,402 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:34:07,488 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:34:07,648 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:34:07,752 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:34:08,648 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:34:11,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:12,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:16,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:17,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:21,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:24,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:24,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:34:24,468 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:34:24,469 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:34:24,552 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:34:24,659 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:34:26,660 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:34:27,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:29,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:32,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:34,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:37,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:37,757 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:34:41,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:41,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:34:41,491 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:34:41,492 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:34:41,575 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:34:41,670 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:34:42,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:42,671 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:34:46,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:47,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:51,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:52,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:57,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:34:57,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:34:57,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:34:57,969 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:34:57,970 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:34:58,051 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:34:58,681 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:35:00,682 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:35:02,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:03,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:07,764 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:35:07,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:08,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:13,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:15,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:15,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:15,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:35:15,557 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:35:15,557 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:35:15,639 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:35:15,692 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:35:16,693 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:35:18,694 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:35:19,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:20,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:24,641 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:25,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:29,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:31,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:33,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:35:33,573 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:35:33,573 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:35:33,656 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:35:33,704 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:35:34,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:34,705 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:35:36,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:37,769 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:35:39,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:41,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:44,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:47,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:47,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:49,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:51,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:35:51,169 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:35:51,169 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:35:51,251 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:35:51,748 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:35:52,749 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:35:53,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:35:55,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:35:58,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:00,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:03,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:05,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:07,775 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:36:08,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:08,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:36:08,823 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:36:08,823 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:36:08,904 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:36:09,764 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:36:10,764 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:36:10,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:13,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:15,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:18,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:20,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:25,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:25,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:36:25,899 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:36:25,900 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:36:25,980 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:36:25,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:26,775 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:36:28,776 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:36:30,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:30,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:35,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:35,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:37,781 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:36:41,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:43,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:43,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:36:43,363 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:36:43,364 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:36:43,446 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:36:43,799 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:36:44,800 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:36:47,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:48,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:52,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:53,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:57,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:36:59,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:36:59,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:36:59,825 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:36:59,825 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:36:59,907 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:37:00,811 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:37:00,811 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:37:02,812 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:37:02,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:04,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:07,787 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:37:08,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:09,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:13,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:17,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:17,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:17,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:37:17,503 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:37:17,503 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:37:17,585 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:37:17,822 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:37:18,822 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:37:19,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:22,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:24,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:27,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:29,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:34,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:35,182 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:35,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:35,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:37:35,212 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:37:35,213 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:37:35,293 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:37:35,833 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:37:36,834 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:37:37,792 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:37:39,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:40,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:44,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:45,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:49,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:52,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:52,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:37:52,884 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:37:52,885 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:37:52,965 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:37:53,845 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:37:54,846 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:37:54,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:37:57,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:37:59,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:02,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:04,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:07,798 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:38:10,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:10,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:10,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:38:10,829 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:38:10,830 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:38:10,910 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:38:10,911 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:38:11,911 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:38:12,911 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:38:15,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:15,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:20,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:20,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:25,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:27,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:27,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:38:27,975 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:38:27,976 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:38:28,057 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:38:28,923 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:38:28,923 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:38:30,924 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:38:31,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:32,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:36,059 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:37,804 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:38:37,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:41,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:45,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:45,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:38:45,514 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:38:45,514 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:38:45,595 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:38:45,934 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:38:46,935 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:38:47,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:50,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:52,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:38:55,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:38:57,598 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:02,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:02,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:39:02,141 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:39:02,141 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:39:02,225 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:39:02,945 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:39:02,946 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:39:03,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:04,946 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:39:07,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:07,810 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:39:08,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:12,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:13,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:18,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:19,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:19,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:19,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:39:19,325 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:39:19,326 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:39:19,407 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:39:19,956 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:39:20,957 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:39:24,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:24,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:29,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:29,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:34,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:37,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:37,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:37,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:39:37,330 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:39:37,330 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:39:37,412 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:39:37,816 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:39:37,967 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:39:38,968 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:39:39,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:42,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:44,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:47,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:49,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:54,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:54,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:39:54,781 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:39:54,782 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:39:54,863 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:39:54,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:39:54,978 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:39:56,979 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:39:59,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:39:59,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:04,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:40:04,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:07,822 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:40:10,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:12,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:40:12,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:40:12,859 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:40:12,860 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:40:12,940 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:40:12,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:40:14,991 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:40:15,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:17,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:40:20,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:22,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:40:25,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:29,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:40:29,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:40:29,941 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:40:29,942 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:40:30,022 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:40:31,021 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:40:31,021 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:40:31,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:34,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:40:36,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:37,827 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:40:39,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:40:41,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:46,832 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:47,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:40:47,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:40:47,813 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:40:47,813 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:40:47,894 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:40:48,032 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:40:49,033 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:40:51,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:52,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:40:56,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:40:57,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:01,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:04,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:04,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:04,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:41:04,719 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:41:04,720 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:41:04,801 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:41:05,043 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:41:07,044 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:41:07,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:07,833 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:41:09,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:12,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:14,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:17,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:21,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:21,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:21,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:41:21,867 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:41:21,867 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:41:21,948 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:41:22,054 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:41:22,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:23,055 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:41:26,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:27,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:31,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:32,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:37,840 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:41:38,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:39,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:39,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:39,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:41:39,111 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:41:39,112 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:41:39,192 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:41:40,066 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:41:41,066 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:41:44,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:44,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:49,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:49,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:54,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:41:55,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:55,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:41:55,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:41:55,848 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:41:55,849 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:41:55,932 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:41:56,076 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:41:57,077 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:41:59,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:00,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:04,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:05,844 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:07,845 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:42:10,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:14,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:14,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:14,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:42:14,325 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:42:14,325 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:42:14,407 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:42:15,088 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:42:15,089 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:42:16,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:19,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:21,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:24,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:26,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:31,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:31,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:42:31,324 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:42:31,325 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:42:31,406 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:42:32,099 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:42:32,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:33,100 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:42:36,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:37,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:37,851 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:42:41,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:42,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:46,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:47,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:49,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:42:49,529 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:42:49,530 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:42:49,611 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:42:50,111 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:42:51,111 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:42:51,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:53,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:42:56,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:42:58,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:01,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:03,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:06,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:43:06,512 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:43:06,513 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:43:06,595 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:43:07,122 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:43:07,122 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:43:07,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:07,857 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:43:08,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:09,123 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:43:12,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:13,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:17,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:18,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:23,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:23,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:43:23,824 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:43:23,824 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:43:23,904 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:43:23,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:24,132 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:43:25,133 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:43:28,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:28,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:33,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:33,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:37,863 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:43:39,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:40,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:40,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:40,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:43:40,910 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:43:40,910 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:43:40,995 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:43:41,146 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:43:43,148 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:43:44,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:45,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:49,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:50,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:54,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:43:57,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:57,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:43:57,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:43:57,323 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:43:57,324 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:43:57,405 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:43:58,158 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:43:59,158 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:44:00,406 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:02,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:05,407 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:07,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:07,869 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:44:10,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:15,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:15,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:15,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:44:15,658 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:44:15,658 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:44:15,741 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:44:16,169 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:44:16,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:17,169 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:44:20,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:21,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:25,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:26,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:31,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:32,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:32,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:44:32,340 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:44:32,340 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:44:32,422 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:44:33,180 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:44:33,180 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:44:37,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:37,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:37,874 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:44:42,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:42,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:47,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:50,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:50,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:44:50,518 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:44:50,518 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:44:50,599 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:44:51,191 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:44:51,191 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:44:53,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:44:55,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:44:58,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:00,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:03,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:07,880 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:45:08,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:08,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:45:08,149 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:45:08,149 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:45:08,232 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:45:09,230 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:45:09,231 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:45:09,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:13,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:14,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:18,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:19,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:24,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:25,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:25,550 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:45:25,553 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:45:25,553 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:45:25,634 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:45:26,253 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:45:27,257 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:45:29,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:30,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:34,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:35,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:37,886 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:45:39,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:43,268 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:43,269 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:43,274 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:45:43,277 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:45:43,278 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:45:43,364 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:45:44,265 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:45:45,265 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:45:45,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:48,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:50,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:53,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:55,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:45:59,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:45:59,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:45:59,837 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:45:59,838 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:45:59,922 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:46:00,275 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:46:00,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:01,276 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:46:04,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:05,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:07,892 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:46:09,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:11,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:15,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:16,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:18,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:46:18,135 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:46:18,136 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:46:18,217 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:46:18,286 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:46:19,287 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:46:20,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:22,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:25,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:27,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:31,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:31,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:32,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:34,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:46:34,397 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:46:34,397 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:46:34,481 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:46:35,297 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:46:35,298 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:46:36,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:37,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:37,898 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:46:41,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:42,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:46,391 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:47,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:52,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:52,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:46:52,379 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:46:52,379 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:46:52,460 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:46:53,308 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:46:53,309 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:46:53,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:46:57,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:46:58,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:02,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:03,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:07,903 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:47:08,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:09,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:09,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:47:09,619 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:47:09,620 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:47:09,701 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:47:10,319 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:47:11,320 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:47:14,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:14,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:19,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:19,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:24,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:26,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:26,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:47:26,864 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:47:26,864 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:47:26,944 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:47:27,330 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:47:27,330 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:47:29,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:31,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:34,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:36,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:37,909 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:47:40,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:45,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:45,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:47:45,037 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:47:45,037 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:47:45,121 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:47:45,341 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:47:45,342 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:47:46,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:47,342 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:47:50,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:51,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:47:55,029 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:47:56,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:01,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:01,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:01,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:01,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:48:01,976 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:48:01,976 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:48:02,057 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:48:02,352 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:48:03,353 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:48:06,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:07,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:07,915 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:48:11,970 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:12,918 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:17,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:20,026 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:20,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:48:20,035 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:48:20,036 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:48:20,118 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:48:20,363 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:48:21,364 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:48:23,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:25,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:28,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:30,028 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:33,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:36,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:36,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:36,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:48:36,303 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:48:36,303 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:48:36,383 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:48:36,384 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:48:37,384 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:48:37,921 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:48:38,923 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:41,296 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:43,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:46,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:48,925 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:53,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:48:54,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:54,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:48:54,119 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:48:54,120 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:48:54,201 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:48:54,395 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:48:55,395 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:48:59,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:48:59,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:04,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:04,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:07,927 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:49:09,930 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:11,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:11,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:49:11,160 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:49:11,160 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:49:11,242 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:49:11,406 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:49:12,406 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:49:14,408 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:49:15,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:16,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:20,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:21,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:25,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:26,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:28,737 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:49:28,740 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:49:28,740 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:49:28,823 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:49:29,417 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:49:30,418 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:49:30,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:31,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:35,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:36,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:37,933 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:49:40,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:41,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:45,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:46,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:46,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:49:46,778 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:49:46,778 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:49:46,863 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:49:47,429 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:49:48,429 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:49:51,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:51,864 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:49:56,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:49:56,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:01,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:03,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:03,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:03,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:50:03,918 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:50:03,919 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:50:04,001 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:50:04,439 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:50:06,441 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:50:07,002 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:07,938 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:50:08,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:12,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:13,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:17,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:21,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:21,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:21,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:50:21,816 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:50:21,817 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:50:21,899 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:50:22,451 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:50:23,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:24,452 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:50:26,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:28,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:31,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:33,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:37,942 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:50:38,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:38,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:38,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:50:38,344 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:50:38,346 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:50:38,427 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:50:38,461 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:50:39,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:40,462 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:50:43,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:44,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:48,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:49,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:54,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:50:55,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:50:55,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:50:55,773 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:50:55,773 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:50:55,855 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:50:56,472 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:50:58,473 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:50:59,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:00,766 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:04,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:05,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:07,948 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:51:09,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:12,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:12,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:51:13,000 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:51:13,000 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:51:13,081 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:51:13,485 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:51:14,485 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:51:15,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:17,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:20,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:22,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:25,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:28,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:30,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:30,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:51:30,872 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:51:30,873 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:51:30,954 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:51:31,496 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:51:32,497 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:51:33,865 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:35,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:37,954 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:51:38,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:40,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:44,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:45,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:48,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:51:48,645 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:51:48,646 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:51:48,727 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:51:49,508 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:51:50,509 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:51:50,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:51,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:51:55,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:51:56,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:00,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:01,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:06,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:06,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:06,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:52:06,354 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:52:06,355 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:52:06,436 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:52:06,519 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:52:07,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:07,959 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:52:08,520 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:52:11,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:12,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:16,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:17,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:22,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:24,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:24,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:24,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:52:24,009 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:52:24,009 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:52:24,094 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:52:24,531 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:52:26,532 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:52:28,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:29,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:33,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:34,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:37,965 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:52:38,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:41,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:41,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:41,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:52:41,157 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:52:41,158 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:52:41,239 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:52:41,542 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:52:42,542 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:52:44,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:46,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:49,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:51,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:54,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:52:58,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:52:58,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:52:58,537 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:52:58,538 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:52:58,620 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:52:59,620 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:52:59,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:00,621 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:53:03,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:04,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:07,971 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:53:08,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:09,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:14,975 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:15,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:15,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:53:15,849 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:53:15,850 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:53:15,931 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:53:16,632 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:53:18,633 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:53:20,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:20,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:25,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:25,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:30,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:31,031 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:33,258 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:53:33,261 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:53:33,262 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:53:33,343 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:53:33,643 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:53:34,643 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:53:36,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:36,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:37,977 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:53:41,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:41,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:46,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:46,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:50,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:53:50,523 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:53:50,523 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:53:50,607 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:53:50,654 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:53:52,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:52,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:53:52,655 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:53:57,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:53:57,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:02,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:02,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:07,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:07,983 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:54:08,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:08,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:08,245 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:54:08,248 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:54:08,249 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:54:08,330 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:54:08,669 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:54:10,670 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:54:13,241 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:13,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:18,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:18,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:23,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:25,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:25,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:25,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:54:25,503 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:54:25,503 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:54:25,584 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:54:25,694 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:54:26,699 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:54:28,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:30,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:33,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:35,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:37,988 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:54:38,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:43,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:43,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:54:43,391 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:54:43,392 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:54:43,473 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:54:43,725 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:54:44,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:44,729 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:54:48,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:49,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:53,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:54:54,475 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:54:59,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:00,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:00,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:00,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:55:00,484 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:55:00,485 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:55:00,567 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:55:00,747 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:55:02,749 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:55:04,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:05,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:07,995 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:55:09,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:10,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:14,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:17,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:18,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:55:18,004 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:55:18,004 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:55:18,086 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:55:18,759 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:55:20,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:20,760 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:55:22,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:25,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:27,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:30,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:35,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:35,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:35,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:55:35,073 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:55:35,073 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:55:35,156 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:55:35,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:35,770 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:55:36,771 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:55:38,000 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:55:40,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:41,003 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:45,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:46,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:51,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:51,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:51,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:55:51,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:55:51,936 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:55:51,937 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:55:52,020 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:55:52,781 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:55:54,782 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:55:56,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:55:56,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:01,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:01,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:06,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:08,006 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:56:09,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:09,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:09,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:56:09,718 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:56:09,719 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:56:09,800 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:56:10,800 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:56:10,800 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:56:11,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:14,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:16,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:19,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:21,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:26,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:26,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:26,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:56:26,490 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:56:26,490 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:56:26,572 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:56:26,810 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:56:27,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:28,812 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:56:31,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:32,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:36,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:37,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:38,011 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:56:43,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:44,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:44,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:56:44,966 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:56:44,967 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:56:45,048 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:56:45,822 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:56:46,823 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:56:48,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:49,958 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:53,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:56:54,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:56:58,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:02,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:02,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:57:02,071 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:57:02,071 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:57:02,154 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:57:02,833 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:57:03,155 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:04,835 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:57:07,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:08,017 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:57:09,020 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:12,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:14,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:19,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:19,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:19,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:57:19,993 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:57:19,993 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:57:20,077 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:57:20,845 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:57:20,846 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:57:22,847 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:57:24,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:24,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:29,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:29,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:34,081 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:37,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:37,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:57:37,170 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:57:37,170 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:57:37,252 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:57:37,856 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:57:38,023 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:57:38,857 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:57:40,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:42,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:45,026 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:47,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:50,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:54,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:54,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:57:54,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:57:54,073 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:57:54,074 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:57:54,158 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:57:54,868 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:57:54,868 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:57:55,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:57:56,869 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:57:59,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:00,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:04,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:05,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:08,029 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:58:11,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:11,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:11,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:11,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:58:11,619 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:58:11,620 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:58:11,702 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:58:11,879 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:58:12,880 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:58:16,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:16,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:21,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:21,705 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:26,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:28,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:28,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:28,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:58:28,058 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:58:28,058 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:58:28,140 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:58:28,890 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:58:28,891 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:58:32,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:33,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:37,142 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:38,035 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:58:38,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:43,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:46,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:46,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:58:46,508 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:58:46,509 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:58:46,590 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:58:46,902 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:58:48,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:48,903 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:58:51,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:53,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:58:56,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:58:58,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:03,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:03,309 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:59:03,312 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:59:03,313 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:59:03,393 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:59:03,913 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:59:04,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:04,914 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:59:08,040 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:59:08,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:10,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:13,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:15,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:20,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:21,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:21,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:59:21,424 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:59:21,425 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:59:21,505 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:59:21,925 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:59:22,925 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:59:25,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:26,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:30,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:31,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:35,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:38,046 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 19:59:38,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:38,990 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:59:38,993 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:59:38,993 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:59:39,075 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:59:39,936 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:59:40,937 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:59:41,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:43,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:46,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:48,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:51,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:56,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 19:59:56,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 19:59:56,267 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 19:59:56,270 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 19:59:56,270 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 19:59:56,354 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 19:59:56,947 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 19:59:56,948 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 19:59:58,949 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:00:01,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:01,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:06,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:06,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:08,052 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:00:12,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:13,393 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:13,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:13,402 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:00:13,403 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:00:13,485 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:00:13,958 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:00:14,959 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:00:17,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:18,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:22,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:23,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:27,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:29,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:29,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:29,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:29,781 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:00:29,782 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:00:29,864 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:00:29,969 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:00:30,970 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:00:32,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:34,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:37,867 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:38,058 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:00:39,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:43,061 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:44,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:46,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:46,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:47,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:47,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:47,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:48,062 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:49,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:50,993 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:52,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:52,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:52,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:53,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:53,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:53,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:53,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:53,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:54,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:00:57,230 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:58,064 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:00:58,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:58,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:58,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:59,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:59,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:59,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:00:59,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:00,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:03,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:03,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:04,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:05,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:05,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:05,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:05,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:05,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:05,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:06,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:08,064 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:01:08,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:09,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:09,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:11,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:11,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:11,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:11,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:11,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:12,094 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:12,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:13,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:14,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:15,929 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:17,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:17,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:17,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:18,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:18,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:18,201 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:18,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:18,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:19,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:22,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:23,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:23,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:24,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:24,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:24,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:24,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:24,651 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:24,786 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:25,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:28,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:28,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:29,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:30,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:30,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:30,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:30,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:30,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:30,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:31,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:33,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:34,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:34,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:36,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:36,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:36,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:36,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:37,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:37,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:37,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:38,069 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:01:38,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:39,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:40,938 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:42,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:42,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:42,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:43,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:43,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:43,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:43,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:43,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:44,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:47,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:48,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:48,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:48,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:48,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:49,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:49,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:49,606 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:49,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:50,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:53,075 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:53,442 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:54,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:01:55,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:55,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:55,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:55,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:55,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:55,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:56,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:58,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:01:59,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:01:59,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:01,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:01,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:01,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:01,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:02,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:02,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:02,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:03,077 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:04,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:05,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:07,701 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:07,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:07,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:08,076 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:02:08,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:08,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:08,275 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:08,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:08,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:09,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:12,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:13,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:14,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:14,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:14,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:14,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:14,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:14,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:14,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:15,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:18,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:18,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:19,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:20,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:20,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:20,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:20,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:20,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:20,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:21,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:23,081 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:24,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:24,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:26,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:26,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:26,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:27,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:27,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:27,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:27,637 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:28,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:29,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:31,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:32,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:32,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:32,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:33,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:33,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:33,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:33,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:33,869 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:34,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:37,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:38,081 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:02:38,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:39,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:39,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:39,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:39,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:39,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:39,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:39,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:40,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:43,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:43,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:44,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:45,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:45,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:45,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:45,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:45,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:45,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:46,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:48,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:49,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:49,979 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:51,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:51,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:51,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:52,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:52,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:52,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:52,801 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:53,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:54,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:02:56,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:57,928 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:57,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:57,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:58,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:02:58,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:58,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:58,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:59,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:02:59,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:02,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:03,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:04,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:04,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:04,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:04,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:04,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:04,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:04,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:05,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:08,087 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:03:08,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:08,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:09,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:10,468 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:10,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:10,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:10,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:11,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:11,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:11,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:13,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:14,814 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:15,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:16,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:16,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:16,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:17,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:17,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:17,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:17,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:18,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:19,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:21,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:22,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:23,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:23,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:23,092 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:23,461 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:23,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:23,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:24,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:24,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:27,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:28,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:29,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:29,314 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:29,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:29,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:29,817 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:29,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:29,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:30,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:33,094 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:33,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:34,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:35,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:35,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:35,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:35,992 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:36,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:36,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:36,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:38,093 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:03:38,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:39,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:40,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:41,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:41,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:41,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:42,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:42,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:42,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:42,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:43,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:44,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:46,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:48,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:48,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:48,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:48,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:48,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:48,630 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:48,728 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:49,167 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:49,822 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:52,631 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:53,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:54,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:54,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:54,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:54,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:54,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:03:54,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:55,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:55,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:58,099 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:03:58,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:03:59,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:00,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:00,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:00,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:01,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:01,162 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:01,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:01,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:03,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:04,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:05,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:06,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:06,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:06,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:06,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:08,099 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:04:08,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:09,828 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:13,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:14,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:18,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:19,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:22,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:23,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:24,831 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:28,105 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:29,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:33,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:34,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:37,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:37,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:38,104 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:04:38,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:39,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:39,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:42,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:42,941 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:43,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:43,408 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:43,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:43,626 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:44,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:44,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:44,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:45,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:48,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:49,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:49,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:49,653 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:49,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:49,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:49,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:50,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:50,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:52,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:53,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:54,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:04:55,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:55,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:56,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:56,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:56,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:56,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:56,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:58,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:04:58,469 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:04:59,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:02,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:02,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:02,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:02,600 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:02,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:03,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:03,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:03,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:04,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:04,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:08,110 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:05:08,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:08,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:08,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:08,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:08,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:08,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:09,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:09,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:09,842 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:11,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:13,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:14,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:14,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:14,844 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:14,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:15,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:15,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:15,642 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:15,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:17,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:18,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:19,844 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:20,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:20,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:21,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:21,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:21,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:21,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:21,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:23,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:23,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:24,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:27,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:27,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:27,467 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:27,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:27,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:28,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:28,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:28,119 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:29,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:29,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:33,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:33,266 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:33,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:33,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:33,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:33,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:34,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:34,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:34,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:36,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:38,117 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:05:38,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:39,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:39,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:39,849 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:39,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:40,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:40,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:40,665 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:40,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:42,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:43,120 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:44,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:45,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:45,836 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:46,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:46,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:46,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:46,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:46,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:48,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:48,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:49,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:52,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:52,072 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:52,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:52,624 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:52,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:53,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:53,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:53,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:54,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:05:54,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:58,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:05:58,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:58,345 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:58,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:58,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:59,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:59,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:59,470 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:05:59,854 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:01,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:03,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:04,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:04,592 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:04,855 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:05,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:05,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:05,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:05,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:05,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:07,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:08,122 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:06:08,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:09,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:10,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:10,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:11,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:11,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:11,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:12,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:12,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:13,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:13,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:14,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:17,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:17,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:17,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:17,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:17,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:18,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:18,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:18,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:19,874 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:20,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:23,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:23,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:23,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:24,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:24,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:24,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:24,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:24,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:24,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:26,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:28,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:29,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:29,829 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:29,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:30,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:30,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:30,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:30,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:30,940 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:32,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:33,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:34,877 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:36,063 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:36,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:36,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:36,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:36,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:37,200 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:37,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:38,128 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:06:38,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:38,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:39,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:42,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:42,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:42,874 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:42,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:43,086 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:43,132 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:43,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:43,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:44,879 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:45,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:48,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:48,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:48,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:49,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:49,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:49,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:49,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:49,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:49,880 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:51,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:53,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:54,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:06:54,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:54,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:55,443 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:55,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:55,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:56,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:56,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:57,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:06:58,134 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:06:59,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:01,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:01,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:01,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:01,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:01,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:02,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:02,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:03,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:04,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:04,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:07,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:07,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:08,134 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:07:08,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:08,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:08,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:08,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:08,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:08,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:09,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:10,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:13,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:14,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:14,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:14,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:14,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:14,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:14,887 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:15,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:15,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:16,917 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:18,138 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:19,888 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:20,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:20,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:20,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:20,939 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:21,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:21,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:21,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:23,139 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:23,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:24,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:26,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:26,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:27,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:27,183 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:27,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:27,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:27,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:28,140 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:29,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:29,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:32,935 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:32,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:33,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:33,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:33,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:33,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:34,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:34,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:34,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:35,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:38,139 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:07:38,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:39,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:39,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:39,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:39,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:39,878 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:39,893 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:40,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:40,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:42,117 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:43,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:44,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:45,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:45,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:46,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:46,148 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:46,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:46,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:46,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:48,144 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:48,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:49,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:51,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:51,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:52,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:52,417 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:52,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:52,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:52,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:53,145 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:54,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:54,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:07:58,146 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:07:58,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:58,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:58,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:58,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:58,868 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:59,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:59,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:07:59,897 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:03,147 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:04,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:08,145 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:08:08,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:09,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:13,149 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:14,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:08:14,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:18,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:21,833 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:21,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:08:22,059 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:08:22,159 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:08:22,211 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:08:22,283 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:08:23,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:23,284 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:08:26,834 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:28,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:31,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:33,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:38,152 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:08:38,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:38,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:38,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:08:38,350 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:08:38,350 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:08:38,432 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:08:39,294 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:08:39,295 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:08:43,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:43,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:48,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:48,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:53,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:08:55,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:08:55,365 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:08:55,368 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:08:55,369 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:08:55,449 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:08:56,305 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:08:57,306 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:08:58,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:00,362 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:03,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:05,364 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:08,157 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:09:09,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:12,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:12,608 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:09:12,612 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:09:12,612 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:09:12,694 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:09:13,315 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:09:13,316 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:09:14,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:17,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:19,696 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:22,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:24,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:28,946 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:28,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:09:28,953 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:09:28,954 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:09:29,038 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:09:29,325 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:09:29,326 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:09:30,039 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:31,327 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:09:33,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:35,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:38,163 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:09:38,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:40,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:44,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:45,167 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:48,590 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:09:48,594 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:09:48,594 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:09:48,676 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:09:49,338 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:09:49,338 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:09:50,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:50,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:09:55,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:09:55,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:00,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:01,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:01,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:04,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:10:04,989 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:10:04,989 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:10:05,072 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:10:05,347 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:10:05,348 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:10:06,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:06,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:08,169 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:10:11,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:11,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:16,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:16,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:21,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:24,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:24,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:10:24,025 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:10:24,026 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:10:24,108 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:10:24,360 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:10:25,360 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:10:26,361 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:10:27,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:29,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:32,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:34,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:37,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:38,175 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:10:40,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:40,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:10:40,344 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:10:40,345 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:10:40,427 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:10:41,426 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:10:41,427 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:10:42,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:45,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:47,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:50,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:52,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:57,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:10:57,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:10:57,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:10:57,529 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:10:57,529 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:10:57,612 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:10:58,437 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:10:59,438 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:11:02,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:02,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:07,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:07,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:08,181 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:11:13,184 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:14,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:14,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:11:14,654 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:11:14,654 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:11:14,738 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:11:15,448 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:11:15,449 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:11:18,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:19,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:23,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:24,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:28,741 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:30,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:30,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:30,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:11:30,815 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:11:30,816 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:11:30,897 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:11:31,459 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:11:31,460 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:11:32,460 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:11:33,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:35,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:38,186 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:11:39,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:40,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:44,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:46,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:46,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:49,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:50,284 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:11:50,287 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:11:50,288 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:11:50,369 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:11:50,471 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:11:51,472 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:11:52,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:52,473 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:11:54,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:11:57,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:11:59,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:03,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:03,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:04,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:06,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:12:06,715 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:12:06,715 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:12:06,796 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:12:07,482 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:12:07,483 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:12:08,192 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:12:08,483 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:12:08,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:10,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:13,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:15,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:18,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:20,197 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:25,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:26,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:26,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:12:26,351 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:12:26,352 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:12:26,433 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:12:26,494 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:12:27,498 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:12:28,498 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:12:30,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:31,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:35,436 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:36,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:38,198 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:12:41,202 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:42,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:42,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:12:42,785 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:12:42,785 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:12:42,869 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:12:43,508 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:12:44,509 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:12:46,871 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:47,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:51,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:12:52,779 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:12:56,872 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:00,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:00,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:13:00,662 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:13:00,663 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:13:00,744 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:13:01,520 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:13:02,521 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:13:02,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:05,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:07,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:08,204 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:13:10,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:13,208 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:17,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:17,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:17,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:13:17,527 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:13:17,528 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:13:17,609 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:13:17,611 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:13:18,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:19,613 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:13:20,614 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:13:22,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:23,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:27,520 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:28,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:33,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:34,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:34,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:34,019 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:13:34,022 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:13:34,023 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:13:34,104 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:13:34,623 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:13:36,624 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:13:38,209 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:13:39,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:39,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:44,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:44,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:49,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:52,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:52,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:13:52,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:13:52,025 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:13:52,025 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:13:52,107 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:13:52,635 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:13:54,636 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:13:55,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:13:57,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:00,110 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:02,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:05,111 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:08,215 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:14:08,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:08,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:08,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:14:08,349 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:14:08,350 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:14:08,437 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:14:08,645 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:14:10,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:10,647 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:14:13,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:15,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:18,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:20,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:24,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:24,035 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:25,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:27,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:14:27,374 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:14:27,374 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:14:27,456 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:14:27,657 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:14:28,658 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:14:29,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:30,457 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:34,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:35,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:38,222 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:14:40,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:41,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:43,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:14:43,951 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:14:43,952 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:14:44,034 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:14:44,668 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:14:45,943 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:46,669 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:14:47,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:50,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:52,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:14:55,945 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:14:57,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:02,038 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:02,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:02,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:15:02,465 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:15:02,466 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:15:02,548 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:15:02,679 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:15:04,681 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:15:07,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:07,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:08,228 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:15:12,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:13,232 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:18,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:19,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:19,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:15:19,353 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:15:19,353 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:15:19,435 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:15:19,690 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:15:20,691 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:15:23,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:24,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:28,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:29,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:33,441 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:34,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:36,666 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:15:36,669 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:15:36,670 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:15:36,751 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:15:36,752 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:15:38,233 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:15:38,752 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:15:39,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:39,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:39,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:44,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:44,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:49,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:50,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:54,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:15:54,199 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:15:54,200 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:15:54,281 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:15:54,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:15:54,763 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:15:56,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:56,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:15:56,764 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:15:59,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:01,191 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:04,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:07,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:08,239 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:16:10,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:10,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:16:10,588 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:16:10,589 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:16:10,674 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:16:10,773 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:16:12,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:12,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:12,775 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:16:15,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:17,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:20,677 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:22,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:25,678 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:28,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:28,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:28,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:16:28,812 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:16:28,813 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:16:28,896 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:16:29,786 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:16:30,787 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:16:30,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:33,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:35,898 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:38,245 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:16:38,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:41,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:45,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:45,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:45,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:16:45,550 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:16:45,550 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:16:45,632 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:16:45,796 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:16:46,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:46,797 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:16:50,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:51,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:16:55,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:16:56,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:01,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:01,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:01,636 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:04,466 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:17:04,469 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:17:04,470 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:17:04,551 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:17:04,809 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:17:06,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:06,810 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:17:07,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:08,251 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:17:11,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:13,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:17,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:18,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:21,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:17:21,396 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:17:21,396 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:17:21,477 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:17:21,820 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:17:22,821 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:17:23,387 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:23,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:28,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:28,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:33,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:33,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:38,257 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:17:39,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:39,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:39,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:17:39,402 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:17:39,402 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:17:39,484 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:17:39,831 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:17:40,832 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:17:44,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:44,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:49,396 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:49,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:54,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:17:56,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:17:56,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:17:56,583 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:17:56,584 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:17:56,668 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:17:56,842 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:17:58,843 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:17:59,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:01,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:04,671 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:06,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:08,262 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:18:10,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:13,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:13,490 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:13,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:18:13,498 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:18:13,498 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:18:13,580 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:18:13,853 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:18:14,854 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:18:15,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:18,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:20,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:23,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:25,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:30,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:30,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:30,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:30,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:18:30,959 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:18:30,959 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:18:31,041 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:18:31,865 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:18:32,866 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:18:35,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:36,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:38,268 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:18:40,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:41,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:46,272 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:47,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:47,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:47,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:18:47,690 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:18:47,691 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:18:47,772 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:18:47,876 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:18:48,876 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:18:51,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:52,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:18:56,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:18:57,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:01,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:02,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:02,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:05,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:19:05,955 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:19:05,956 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:19:06,037 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:19:06,888 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:19:07,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:08,274 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:19:08,889 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:19:08,947 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:12,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:13,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:17,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:19,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:22,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:22,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:19:22,912 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:19:22,913 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:19:22,997 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:19:23,899 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:19:24,900 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:19:24,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:27,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:29,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:32,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:34,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:38,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:38,280 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:19:41,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:41,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:19:41,241 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:19:41,242 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:19:41,324 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:19:41,911 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:19:42,911 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:19:43,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:46,235 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:48,326 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:51,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:53,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:58,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:19:58,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:19:58,307 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:19:58,308 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:19:58,389 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:19:58,389 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:19:58,922 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:20:00,923 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:20:03,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:03,390 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:08,285 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:20:08,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:09,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:14,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:15,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:15,889 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:15,894 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:20:15,897 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:20:15,898 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:20:15,979 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:20:15,980 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:20:16,980 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:20:18,981 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:20:19,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:20,890 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:24,983 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:25,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:29,984 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:32,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:32,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:32,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:20:32,597 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:20:32,597 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:20:32,679 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:20:32,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:20:34,991 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:20:35,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:37,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:38,291 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:20:41,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:42,593 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:46,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:49,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:49,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:20:49,575 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:20:49,576 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:20:49,659 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:20:50,001 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:20:51,002 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:20:51,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:54,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:20:56,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:20:59,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:01,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:05,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:06,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:07,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:21:07,103 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:21:07,104 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:21:07,186 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:21:08,013 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:21:08,297 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:21:09,014 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:21:11,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:12,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:16,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:17,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:22,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:24,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:24,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:24,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:21:24,550 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:21:24,551 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:21:24,632 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:21:25,024 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:21:27,025 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:21:27,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:29,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:32,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:34,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:37,635 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:38,303 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:21:39,847 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:39,848 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:43,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:43,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:21:43,167 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:21:43,168 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:21:43,251 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:21:44,036 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:21:45,037 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:21:45,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:48,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:50,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:53,254 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:21:55,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:21:58,255 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:00,410 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:00,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:22:00,419 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:22:00,419 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:22:00,501 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:22:01,046 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:22:03,048 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:22:03,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:05,411 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:08,309 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:22:09,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:10,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:14,313 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:18,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:18,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:22:18,490 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:22:18,491 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:22:18,573 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:22:19,070 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:22:19,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:21,072 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:22:23,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:24,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:28,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:29,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:34,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:35,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:35,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:22:35,057 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:22:35,058 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:22:35,143 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:22:35,143 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:22:37,146 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:22:38,314 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:22:40,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:40,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:45,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:45,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:50,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:52,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:52,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:22:52,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:22:52,625 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:22:52,626 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:22:52,709 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:22:53,172 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:22:55,174 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:22:55,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:22:57,617 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:00,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:02,618 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:05,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:08,321 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:23:09,331 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:09,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:09,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:23:09,339 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:23:09,340 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:23:09,425 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:23:10,199 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:23:11,200 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:23:11,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:14,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:16,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:19,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:21,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:26,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:26,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:26,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:26,588 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:23:26,591 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:23:26,592 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:23:26,673 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:23:27,225 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:23:29,227 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:23:31,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:31,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:36,584 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:36,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:38,326 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:23:42,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:44,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:44,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:23:44,667 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:23:44,667 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:23:44,749 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:23:45,256 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:23:47,258 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:23:47,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:49,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:52,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:23:54,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:23:57,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:01,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:01,695 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:24:01,698 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:24:01,698 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:24:01,780 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:24:02,282 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:24:02,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:03,283 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:24:06,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:07,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:08,332 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:24:11,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:13,336 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:16,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:18,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:20,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:24:20,690 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:24:20,691 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:24:20,772 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:24:21,314 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:24:22,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:23,316 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:24:23,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:27,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:28,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:33,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:34,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:37,431 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:24:37,434 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:24:37,435 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:24:37,517 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:24:38,338 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:24:38,353 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:24:39,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:39,353 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:24:39,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:44,341 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:44,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:49,342 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:49,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:54,343 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:24:55,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:24:55,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:24:55,663 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:24:55,664 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:24:55,745 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:24:56,368 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:24:57,370 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:24:59,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:00,655 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:04,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:05,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:08,344 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:25:10,347 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:12,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:12,152 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:12,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:25:12,159 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:25:12,160 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:25:12,241 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:25:12,395 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:25:13,399 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:25:16,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:17,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:21,243 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:22,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:26,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:29,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:29,127 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:25:29,131 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:25:29,131 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:25:29,212 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:25:29,425 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:25:31,427 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:25:32,214 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:34,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:37,215 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:38,349 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:25:39,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:42,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:46,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:46,400 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:46,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:25:46,407 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:25:46,408 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:25:46,490 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:25:46,492 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:25:47,491 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:25:47,491 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:25:47,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:49,492 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:25:51,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:52,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:25:56,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:25:57,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:02,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:03,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:03,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:03,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:26:03,266 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:26:03,267 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:26:03,349 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:26:03,502 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:26:05,503 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:26:08,260 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:08,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:08,355 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:26:13,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:13,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:18,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:18,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:18,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:22,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:26:22,443 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:26:22,443 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:26:22,524 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:26:22,526 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:26:23,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:23,525 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:26:23,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:28,434 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:28,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:33,435 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:33,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:38,361 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:26:39,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:39,207 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:39,209 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:26:39,212 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:26:39,212 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:26:39,295 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:26:39,537 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:26:41,539 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:26:44,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:44,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:49,206 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:49,298 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:54,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:26:54,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:26:58,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:26:58,144 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:26:58,145 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:26:58,227 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:26:58,564 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:26:59,566 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:27:00,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:00,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:05,136 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:05,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:08,367 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:27:10,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:10,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:14,325 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:27:14,328 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:27:14,329 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:27:14,411 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:27:14,589 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:27:15,412 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:15,594 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:27:16,320 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:20,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:21,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:25,414 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:26,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:30,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:31,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:31,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:27:31,976 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:27:31,976 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:27:32,059 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:27:32,617 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:27:33,618 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:27:36,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:36,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:38,373 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:27:41,376 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:41,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:46,377 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:48,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:48,995 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:49,001 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:27:49,004 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:27:49,004 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:27:49,086 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:27:49,628 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:27:51,630 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:27:52,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:53,996 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:27:57,088 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:27:58,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:02,089 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:05,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:05,967 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:05,972 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:28:05,975 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:28:05,975 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:28:06,058 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:28:06,639 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:28:07,640 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:28:08,060 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:08,379 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:28:10,968 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:13,382 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:15,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:18,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:23,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:24,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:24,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:24,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:28:24,168 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:28:24,168 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:28:24,251 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:28:24,651 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:28:25,651 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:28:29,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:29,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:34,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:34,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:38,383 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:28:39,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:40,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:40,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:28:40,820 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:28:40,822 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:28:40,903 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:28:41,661 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:28:41,662 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:28:43,662 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:28:44,904 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:45,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:49,905 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:50,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:54,906 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:28:56,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:56,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:28:59,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:28:59,730 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:28:59,731 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:28:59,815 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:29:00,673 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:29:00,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:01,674 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:29:01,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:05,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:06,724 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:08,388 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:29:11,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:12,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:12,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:15,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:29:15,958 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:29:15,959 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:29:16,039 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:29:16,684 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:29:17,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:17,684 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:29:17,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:22,041 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:22,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:27,042 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:27,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:32,043 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:34,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:34,286 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:29:34,289 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:29:34,289 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:29:34,370 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:29:34,695 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:29:35,696 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:29:37,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:38,394 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:29:39,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:42,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:44,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:47,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:51,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:51,216 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:29:51,219 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:29:51,219 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:29:51,301 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:29:51,707 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:29:53,302 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:29:53,709 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:29:56,211 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:29:58,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:01,212 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:03,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:06,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:08,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:08,400 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:30:08,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:30:08,724 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:30:08,725 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:30:08,807 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:30:09,719 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:30:09,720 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:30:11,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:11,717 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:13,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:16,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:18,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:22,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:23,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:26,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:30:26,286 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:30:26,287 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:30:26,370 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:30:26,730 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:30:27,730 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:30:28,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:29,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:33,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:34,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:38,406 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:30:39,409 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:39,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:43,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:30:43,018 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:30:43,019 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:30:43,100 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:30:43,741 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:30:45,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:45,011 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:45,101 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:45,742 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:30:50,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:50,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:30:55,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:30:55,103 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:00,104 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:01,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:01,225 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:01,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:31:01,232 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:31:01,233 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:31:01,313 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:31:01,752 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:31:03,754 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:31:05,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:06,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:08,412 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:31:10,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:11,228 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:15,416 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:17,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:17,911 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:17,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:31:17,919 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:31:17,920 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:31:18,003 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:31:18,763 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:31:19,764 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:31:21,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:22,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:26,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:27,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:31,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:33,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:33,673 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:36,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:36,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:31:36,678 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:31:36,678 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:31:36,760 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:31:36,775 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:31:37,776 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:31:38,418 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:31:39,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:41,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:44,670 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:46,421 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:51,423 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:53,418 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:31:53,424 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:31:53,427 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:31:53,428 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:31:53,510 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:31:53,786 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:31:55,787 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:31:56,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:31:58,419 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:01,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:03,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:06,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:08,424 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:32:11,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:11,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:32:11,286 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:32:11,286 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:32:11,369 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:32:11,798 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:32:12,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:13,799 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:32:16,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:17,372 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:21,280 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:22,373 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:27,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:28,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:28,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:32:28,233 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:32:28,233 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:32:28,314 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:32:28,809 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:32:29,810 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:32:33,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:33,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:38,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:38,316 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:38,429 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:32:43,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:45,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:45,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:45,244 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:32:45,247 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:32:45,247 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:32:45,328 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:32:45,820 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:32:47,821 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:32:49,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:50,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:54,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:32:55,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:32:59,332 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:02,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:02,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:02,402 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:33:02,405 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:33:02,406 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:33:02,487 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:33:02,831 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:33:03,832 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:33:04,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:07,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:08,436 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:33:10,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:12,401 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:15,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:19,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:19,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:19,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:33:19,455 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:33:19,455 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:33:19,537 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:33:19,842 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:33:20,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:21,844 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:33:24,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:25,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:29,448 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:30,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:35,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:35,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:35,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:38,199 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:33:38,203 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:33:38,203 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:33:38,287 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:33:38,441 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:33:38,855 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:33:39,855 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:33:40,195 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:41,444 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:45,196 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:46,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:51,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:51,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:33:55,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:33:55,076 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:33:55,077 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:33:55,157 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:33:55,866 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:33:55,867 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:33:57,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:33:57,158 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:02,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:02,159 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:07,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:07,160 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:08,447 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:34:12,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:13,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:13,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:34:13,686 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:34:13,687 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:34:13,771 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:34:13,878 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:34:15,879 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:34:17,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:18,680 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:22,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:23,681 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:27,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:30,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:30,118 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:34:30,120 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:34:30,121 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:34:30,203 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:34:30,889 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:34:31,890 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:34:33,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:35,115 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:38,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:38,453 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:34:40,116 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:43,456 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:45,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:48,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:34:48,078 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:34:48,078 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:34:48,161 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:34:48,901 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:34:49,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:49,901 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:34:51,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:51,070 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:54,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:34:56,071 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:34:59,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:01,969 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:04,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:05,094 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:35:05,097 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:35:05,098 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:35:05,180 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:35:05,911 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:35:05,912 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:35:07,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:07,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:08,459 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:35:09,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:12,091 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:14,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:17,093 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:19,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:22,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:22,333 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:22,339 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:35:22,342 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:35:22,343 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:35:22,426 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:35:22,922 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:35:23,923 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:35:25,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:27,334 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:30,428 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:32,335 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:35,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:37,873 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:38,465 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:35:40,021 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:35:40,024 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:35:40,025 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:35:40,105 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:35:40,934 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:35:41,106 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:41,935 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:35:43,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:46,107 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:48,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:51,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:53,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:53,276 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:35:56,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:35:56,986 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:35:56,990 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:35:56,990 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:35:57,071 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:35:57,945 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:35:57,946 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:35:58,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:02,073 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:03,980 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:07,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:08,470 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:36:08,981 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:12,474 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:15,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:15,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:36:15,362 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:36:15,363 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:36:15,448 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:36:15,957 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:36:15,958 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:36:18,449 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:20,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:23,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:25,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:28,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:31,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:31,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:36:31,536 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:36:31,537 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:36:31,617 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:36:31,967 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:36:33,619 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:33,969 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:36:34,969 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:36:36,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:38,476 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:36:39,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:41,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:44,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:49,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:49,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:49,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:36:49,826 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:36:49,827 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:36:49,909 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:36:49,979 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:36:51,980 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:36:52,981 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:36:54,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:54,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:36:59,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:36:59,913 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:04,914 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:06,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:06,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:06,709 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:37:06,710 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:37:06,791 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:37:06,990 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:37:07,991 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:37:08,482 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:37:08,992 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:37:10,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:11,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:15,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:16,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:20,486 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:21,706 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:23,473 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:24,065 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:24,171 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:24,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:24,704 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:25,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:26,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:28,122 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:29,843 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:29,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:29,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:30,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:30,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:30,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:30,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:30,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:31,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:34,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:35,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:36,094 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:36,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:36,109 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:36,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:36,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:36,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:36,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:37,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:38,488 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:37:40,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:40,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:41,710 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:42,346 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:42,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:42,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:42,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:42,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:42,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:43,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:45,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:46,711 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:46,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:48,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:48,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:48,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:49,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:49,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:49,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:49,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:50,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:51,712 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:53,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:54,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:54,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:54,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:55,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:55,399 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:55,494 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:37:55,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:55,936 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:37:56,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:37:59,349 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:00,495 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:01,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:01,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:01,098 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:01,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:01,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:01,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:01,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:02,180 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:05,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:05,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:06,716 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:07,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:07,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:07,386 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:07,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:07,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:08,047 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:08,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:08,494 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:38:10,497 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:11,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:11,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:13,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:13,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:13,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:14,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:14,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:14,303 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:14,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:15,498 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:16,719 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:18,181 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:19,959 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:19,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:19,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:20,413 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:20,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:20,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:20,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:21,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:21,720 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:24,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:25,500 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:26,198 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:26,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:26,213 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:26,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:26,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:26,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:26,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:27,315 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:30,501 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:30,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:31,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:32,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:32,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:32,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:32,966 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:33,067 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:33,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:33,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:35,502 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:36,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:37,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:38,500 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:38:38,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:38,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:38,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:39,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:39,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:39,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:39,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:40,503 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:41,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:43,329 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:45,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:45,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:45,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:45,504 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:45,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:45,625 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:45,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:46,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:46,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:49,583 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:50,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:51,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:51,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:51,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:51,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:51,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:51,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:52,018 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:52,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:55,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:38:55,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:56,730 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:38:57,674 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:57,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:57,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:58,130 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:58,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:58,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:38:58,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:00,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:01,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:02,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:03,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:03,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:03,932 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:04,388 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:04,493 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:04,591 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:05,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:05,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:06,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:08,505 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:08,506 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:39:10,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:10,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:10,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:10,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:10,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:10,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:10,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:11,340 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:11,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:14,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:15,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:16,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:16,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:16,506 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:16,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:16,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:17,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:17,166 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:17,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:20,510 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:21,078 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:21,736 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:22,824 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:22,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:22,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:23,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:23,380 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:23,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:23,919 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:25,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:26,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:27,327 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:29,068 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:29,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:29,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:29,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:29,647 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:29,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:30,185 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:30,512 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:31,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:33,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:35,447 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:35,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:35,462 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:35,513 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:35,903 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:36,004 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:36,100 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:36,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:36,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:38,512 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:39:39,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:40,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:41,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:41,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:41,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:41,742 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:42,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:42,270 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:42,367 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:42,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:45,516 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:46,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:46,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:48,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:48,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:48,048 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:48,492 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:48,594 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:48,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:49,135 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:50,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:51,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:52,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:54,285 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:54,300 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:54,301 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:54,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:54,863 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:54,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:55,403 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:39:55,517 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:39:56,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:39:58,991 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:00,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:00,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:00,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:00,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:01,187 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:01,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:01,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:01,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:01,832 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:05,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:05,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:06,749 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:06,985 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:07,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:07,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:07,458 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:07,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:07,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:08,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:08,517 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:40:10,521 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:11,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:11,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:13,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:13,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:13,360 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:13,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:13,902 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:13,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:14,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:15,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:16,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:17,856 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:19,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:19,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:19,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:20,069 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:20,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:20,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:20,522 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:20,714 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:21,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:24,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:25,523 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:25,942 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:25,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:25,957 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:26,398 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:26,499 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:26,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:26,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:27,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:30,452 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:30,524 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:31,756 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:32,189 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:32,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:32,205 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:32,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:32,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:32,866 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:33,308 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:35,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:36,757 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:36,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:38,524 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:40:38,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:38,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:38,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:38,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:39,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:39,193 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:39,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:40,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:41,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:43,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:44,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:44,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:44,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:44,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:40:45,528 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:46,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:50,529 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:51,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:40:55,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:40:56,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:00,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:00,530 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:01,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:05,531 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:06,765 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:08,529 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:41:10,532 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:11,767 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:15,437 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:15,515 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:15,533 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:16,768 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:17,299 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:20,534 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:20,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:20,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:21,190 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:21,306 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:21,405 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:21,769 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:21,853 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:21,858 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:23,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:25,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:26,770 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:26,954 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:26,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:27,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:27,525 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:27,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:28,080 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:28,084 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:29,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:30,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:31,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:33,330 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:33,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:33,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:33,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:34,032 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:34,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:34,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:35,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:36,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:36,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:38,535 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:41:39,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:39,623 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:40,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:40,194 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:40,292 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:40,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:40,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:40,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:41,774 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:42,514 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:45,539 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:46,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:46,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:46,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:46,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:46,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:46,775 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:47,156 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:47,161 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:48,931 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:50,540 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:51,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:52,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:52,307 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:52,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:52,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:52,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:53,415 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:53,420 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:55,188 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:55,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:41:56,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:41:58,676 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:58,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:59,167 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:59,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:59,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:59,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:41:59,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:00,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:01,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:01,780 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:04,955 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:04,974 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:05,426 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:05,535 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:05,543 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:05,633 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:06,074 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:06,079 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:06,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:07,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:08,541 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:42:10,544 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:11,337 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:11,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:11,782 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:11,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:11,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:12,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:12,491 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:12,496 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:14,253 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:15,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:16,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:17,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:17,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:18,076 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:18,186 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:18,283 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:18,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:18,729 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:20,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:20,546 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:21,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:23,988 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:24,016 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:24,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:24,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:24,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:25,137 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:25,141 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:25,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:26,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:26,891 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:30,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:30,271 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:30,547 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:30,713 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:30,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:30,922 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:31,370 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:31,374 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:31,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:33,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:35,548 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:36,632 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:36,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:36,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:37,114 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:37,226 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:37,324 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:37,771 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:37,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:38,546 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:42:39,541 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:40,550 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:41,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:42,896 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:42,916 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:43,355 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:43,465 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:43,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:44,017 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:44,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:45,550 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:45,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:46,792 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:49,297 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:49,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:49,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:49,900 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:50,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:50,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:50,464 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:50,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:51,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:52,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:55,552 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:42:55,587 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:55,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:56,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:56,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:56,261 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:56,718 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:56,723 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:42:56,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:42:58,488 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:00,553 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:01,796 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:01,973 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:02,000 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:02,459 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:02,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:02,675 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:03,121 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:03,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:04,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:05,554 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:06,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:08,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:08,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:08,553 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:43:08,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:08,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:08,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:09,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:09,357 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:10,556 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:11,125 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:11,799 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:14,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:14,652 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:15,112 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:15,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:15,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:15,557 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:15,773 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:15,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:16,800 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:17,545 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:20,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:20,907 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:20,926 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:21,371 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:21,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:21,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:21,802 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:22,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:22,027 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:23,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:25,559 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:26,803 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:27,295 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:27,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:27,787 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:27,899 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:27,997 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:28,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:28,451 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:30,227 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:30,560 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:31,804 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:33,582 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:33,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:34,044 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:34,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:34,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:34,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:34,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:35,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:36,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:36,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:38,558 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:43:39,963 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:39,987 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:40,446 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:40,558 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:40,561 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:40,657 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:41,108 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:41,113 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:41,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:42,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:45,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:46,237 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:46,256 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:46,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:46,808 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:46,810 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:46,915 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:47,356 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:47,361 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:49,133 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:50,563 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:51,809 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:52,607 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:52,634 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:53,090 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:53,203 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:53,304 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:53,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:53,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:55,527 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:55,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:43:56,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:43:58,881 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:58,901 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:59,344 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:59,453 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:59,551 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:59,994 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:43:59,999 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:00,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:01,772 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:01,812 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:05,281 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:05,305 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:05,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:05,781 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:05,895 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:05,998 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:06,445 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:06,450 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:06,813 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:08,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:08,564 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:44:10,567 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:11,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:11,589 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:11,815 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:12,033 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:12,143 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:12,242 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:12,694 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:12,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:14,467 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:15,568 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:16,816 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:17,944 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:17,971 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:18,433 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:18,549 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:18,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:19,097 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:19,102 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:20,569 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:20,870 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:21,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:24,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:24,251 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:24,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:24,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:24,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:25,354 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:25,359 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:25,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:26,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:27,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:30,571 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:30,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:30,640 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:31,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:31,210 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:31,312 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:31,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:31,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:31,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:33,526 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:35,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:36,821 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:36,892 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:36,912 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:37,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:37,463 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:37,562 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:38,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:38,012 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:38,570 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:44:40,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:41,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:45,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:46,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:50,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:44:51,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:44:53,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:44:55,576 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:00,317 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:00,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:45:00,551 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:45:00,652 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:45:00,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:00,729 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:45:01,295 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:45:01,296 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:45:03,297 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:45:05,318 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:05,731 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:08,575 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:45:10,319 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:11,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:16,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:17,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:17,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:45:17,767 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:45:17,768 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:45:17,849 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:45:18,307 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:45:19,308 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:45:21,850 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:22,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:26,851 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:27,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:31,852 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:35,163 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:35,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:45:35,172 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:45:35,173 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:45:35,257 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:45:35,318 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:45:37,259 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:37,319 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:45:38,581 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:45:40,164 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:42,585 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:45,165 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:47,586 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:51,793 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:51,797 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:45:51,800 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:45:51,801 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:45:51,881 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:45:52,329 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:45:52,882 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:45:53,329 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:45:56,794 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:45:57,883 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:01,795 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:02,884 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:07,885 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:08,587 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:46:09,168 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:09,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:46:09,178 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:46:09,179 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:46:09,261 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:46:09,340 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:46:11,341 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:46:13,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:14,169 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:18,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:19,170 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:23,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:26,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:26,220 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:26,224 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:46:26,227 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:46:26,228 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:46:26,309 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:46:26,350 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:46:27,351 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:46:28,310 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:31,221 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:33,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:36,222 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:38,311 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:38,593 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:46:43,597 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:44,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:44,392 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:44,397 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:46:44,400 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:46:44,400 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:46:44,482 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:46:45,375 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:46:45,375 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:46:47,377 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:46:49,394 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:49,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:54,395 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:46:54,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:46:59,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:01,648 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:01,654 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:47:01,657 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:47:01,658 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:47:01,739 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:47:02,401 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:47:03,403 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:47:04,740 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:06,649 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:08,598 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:47:10,601 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:11,650 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:15,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:19,082 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:19,087 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:47:19,090 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:47:19,091 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:47:19,171 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:47:19,432 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:47:21,172 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:21,434 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:47:24,083 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:26,173 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:29,085 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:31,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:36,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:36,454 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:36,460 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:47:36,463 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:47:36,463 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:47:36,544 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:47:37,451 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:47:37,452 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:47:38,600 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:47:39,452 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:47:41,455 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:41,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:46,456 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:46,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:51,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:53,481 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:47:53,487 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:47:53,490 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:47:53,490 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:47:53,572 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:47:54,462 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:47:55,463 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:47:57,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:47:58,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:02,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:03,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:07,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:08,606 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:48:10,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:10,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:48:10,762 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:48:10,763 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:48:10,844 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:48:11,473 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:48:12,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:13,474 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:48:15,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:17,845 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:20,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:22,846 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:27,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:27,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:27,693 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:48:27,695 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:48:27,696 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:48:27,776 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:48:28,484 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:48:28,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:29,485 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:48:32,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:33,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:37,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:38,612 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:48:39,615 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:44,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:45,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:45,348 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:45,353 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:48:45,356 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:48:45,357 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:48:45,437 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:48:45,495 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:48:47,496 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:48:50,350 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:50,438 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:48:55,351 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:48:55,439 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:00,440 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:03,246 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:03,252 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:49:03,255 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:49:03,256 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:49:03,336 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:49:03,507 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:49:05,508 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:49:06,338 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:08,247 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:08,618 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:49:11,621 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:13,248 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:16,622 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:18,798 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:20,937 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:49:20,940 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:49:20,940 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:49:21,021 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:49:21,519 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:49:21,519 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:49:22,022 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:23,520 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:49:23,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:27,023 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:28,934 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:32,024 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:34,672 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:37,025 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:38,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:49:38,485 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:49:38,486 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:49:38,568 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:49:38,623 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:49:39,567 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:49:39,568 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:49:40,476 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:42,626 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:45,477 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:47,627 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:50,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:52,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:49:55,818 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:49:55,823 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:49:55,826 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:49:55,826 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:49:55,907 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:49:56,578 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:49:57,578 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:49:57,908 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:00,819 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:02,909 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:05,820 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:07,910 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:08,629 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:50:12,749 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:12,749 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:12,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:50:12,756 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:50:12,757 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:50:12,837 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:50:13,588 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:50:13,589 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:50:13,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:17,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:18,839 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:22,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:23,840 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:28,841 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:29,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:29,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:29,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:50:29,695 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:50:29,695 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:50:29,776 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:50:30,599 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:50:31,600 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:50:34,687 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:34,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:38,635 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:50:39,688 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:40,638 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:45,639 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:46,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:46,574 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:46,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:50:46,580 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:50:46,581 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:50:46,659 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:50:46,660 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:50:47,659 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:50:50,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:51,575 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:50:55,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:50:56,577 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:00,664 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:04,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:04,564 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:04,570 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:51:04,572 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:51:04,573 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:51:04,608 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:51:04,670 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:51:05,671 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:51:06,610 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:08,641 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:51:09,565 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:11,644 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:14,566 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:16,645 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:20,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:20,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:21,646 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:22,294 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:51:22,298 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:51:22,298 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:51:22,381 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:51:22,681 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:51:23,682 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:51:25,290 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:27,383 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:30,291 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:32,384 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:36,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:36,229 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:37,385 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:38,646 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:51:39,956 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:51:39,959 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:51:39,960 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:51:40,044 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:51:40,693 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:51:41,693 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:51:41,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:43,045 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:46,951 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:48,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:51,952 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:53,046 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:57,507 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:51:57,511 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:51:57,514 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:51:57,515 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:51:57,594 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:51:57,703 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:51:58,595 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:51:59,705 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:52:02,508 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:03,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:07,509 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:08,596 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:08,652 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:52:13,656 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:14,231 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:14,236 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:52:14,239 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:52:14,240 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:52:14,319 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:52:14,714 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:52:15,715 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:52:19,233 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:19,321 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:24,234 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:24,322 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:29,323 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:31,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:31,746 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:31,752 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:52:31,755 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:52:31,755 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:52:31,834 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:52:32,726 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:52:33,726 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:52:34,835 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:36,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:38,658 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:52:40,661 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:41,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:45,662 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:48,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:48,698 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:48,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:52:48,705 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:52:48,706 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:52:48,787 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:52:48,788 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:52:49,787 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:52:50,789 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:53,699 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:52:55,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:52:58,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:00,790 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:05,791 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:06,602 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:06,603 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:06,609 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:53:06,611 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:53:06,613 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:53:06,693 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:53:06,798 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:53:07,799 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:53:08,664 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:53:11,604 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:11,667 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:16,605 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:16,668 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:21,669 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:24,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:24,008 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:24,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:53:24,016 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:53:24,016 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:53:24,053 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:53:24,809 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:53:25,810 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:53:27,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:29,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:32,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:34,010 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:37,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:38,671 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:53:41,733 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:41,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:53:41,741 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:53:41,742 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:53:41,824 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:53:42,823 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:53:42,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:43,824 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:53:46,734 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:47,825 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:51,735 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:52,826 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:57,827 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:53:59,277 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:53:59,282 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:53:59,285 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:53:59,286 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:53:59,367 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:53:59,834 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:53:59,835 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:54:01,836 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:54:03,368 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:04,278 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:08,369 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:08,676 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:54:09,279 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:13,679 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:15,961 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:15,965 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:54:15,969 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:54:15,969 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:54:16,049 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:54:16,845 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:54:17,845 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:54:19,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:20,962 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:24,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:25,964 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:29,052 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:33,976 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:33,982 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:54:33,985 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:54:33,986 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:54:34,065 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:54:34,066 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:34,856 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:54:35,857 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:54:38,682 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:54:38,977 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:39,685 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:43,978 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:44,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:49,686 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:50,743 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:50,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:54:50,754 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:54:50,755 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:54:50,835 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:54:50,866 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:54:51,867 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:54:54,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:54:55,744 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:54:59,837 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:00,745 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:04,838 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:08,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:08,482 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:08,489 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:55:08,492 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:55:08,493 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:55:08,574 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:55:08,687 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:55:08,878 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:55:09,878 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:55:10,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:13,483 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:15,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:18,485 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:20,692 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:25,611 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:25,612 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:25,616 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:55:25,619 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:55:25,620 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:55:25,699 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:55:25,700 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:25,889 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:55:27,890 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:55:30,613 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:30,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:35,614 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:35,703 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:38,694 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:55:41,697 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:43,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:43,427 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:43,432 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:55:43,435 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:55:43,436 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:55:43,517 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:55:43,900 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:55:45,901 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:55:47,518 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:48,429 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:52,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:55:53,430 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:55:57,519 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:01,034 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:01,040 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:56:01,043 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:56:01,044 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:56:01,122 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:56:01,912 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:56:01,912 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:56:03,123 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:06,036 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:08,124 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:08,698 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:56:11,037 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:13,702 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:18,174 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:18,178 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:56:18,182 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:56:18,182 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:56:18,262 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:56:18,922 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:56:19,262 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:19,923 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:56:23,175 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:24,263 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:28,176 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:29,264 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:34,265 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:36,287 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:36,293 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:56:36,296 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:56:36,297 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:56:36,379 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:56:36,934 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:56:37,934 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:56:38,704 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:56:39,707 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:41,288 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:44,708 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:46,289 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:49,709 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:52,783 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:52,788 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:56:52,791 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:56:52,791 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:56:52,873 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:56:52,944 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:56:53,944 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:56:54,875 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:56:57,784 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:56:59,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:02,785 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:04,876 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:08,706 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:57:10,682 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:10,689 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:10,690 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:57:10,693 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:57:10,693 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:57:10,773 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:57:10,955 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:57:11,956 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:57:15,683 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:15,776 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:20,684 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:20,777 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:25,778 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:27,658 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:27,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:27,663 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:57:27,666 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:57:27,667 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:57:27,746 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:57:27,965 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:57:29,967 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:57:31,747 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:32,659 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:36,748 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:37,660 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:38,712 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:57:42,715 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:45,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:45,150 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:45,154 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:57:45,157 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:57:45,158 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:57:45,237 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:57:45,985 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:57:45,986 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:57:48,238 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:50,151 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:53,239 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:57:55,153 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:57:58,240 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:02,478 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:02,484 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:58:02,487 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:58:02,487 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:58:02,571 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:58:02,996 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:58:03,572 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:03,996 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:58:07,479 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:08,573 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:08,717 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:58:12,480 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:13,721 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:17,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:18,722 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:19,924 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:58:19,927 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:58:19,928 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:58:20,010 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:58:21,009 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:58:22,013 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:58:22,920 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:24,013 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:27,921 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:29,014 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:33,933 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:34,015 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:37,811 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:58:37,814 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:58:37,814 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:58:37,896 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:58:38,023 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:58:38,723 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:58:39,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:39,805 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:40,025 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:58:41,025 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:58:44,725 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:44,806 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:49,726 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:49,807 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:54,727 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:58:55,005 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:58:55,009 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:58:55,012 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:58:55,013 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:58:55,093 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:58:56,093 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:58:56,093 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:59:00,006 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:00,095 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:05,007 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:05,096 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:08,729 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:59:10,732 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:12,857 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:12,862 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:59:12,865 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:59:12,866 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:59:12,947 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:59:13,104 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:59:14,105 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:59:15,948 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:17,859 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:20,949 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:22,860 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:25,950 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:29,536 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:29,542 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:59:29,545 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:59:29,545 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:59:29,627 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:59:30,117 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:59:30,118 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:59:31,628 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:34,537 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:36,629 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:38,735 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 20:59:39,538 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:41,738 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:46,739 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:47,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:47,126 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:47,131 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-15 20:59:47,134 DEBUG SenderThread:21069 [sender.py:send():378] send: history
+2024-05-15 20:59:47,135 DEBUG SenderThread:21069 [sender.py:send_request():405] send_request: summary_record
+2024-05-15 20:59:47,216 INFO SenderThread:21069 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-15 20:59:48,128 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:59:48,129 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/wandb-summary.json
+2024-05-15 20:59:49,129 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 20:59:52,128 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:52,217 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 20:59:57,129 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 20:59:57,218 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:02,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:05,273 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:00:07,219 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:08,756 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 21:00:12,691 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:00:12,759 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:17,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:19,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:00:19,758 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:00:23,578 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:25,152 INFO Thread-12 :21069 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/files/output.log
+2024-05-15 21:00:28,579 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:29,830 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:00:33,580 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:38,581 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:38,746 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 21:00:43,750 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:48,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:53,751 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:00:58,753 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:03,754 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:08,755 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:08,756 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 21:01:13,760 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:16,050 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:01:16,051 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:01:18,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:21,053 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:01:23,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:26,054 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:01:28,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:31,055 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:01:33,764 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:36,056 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:01:38,758 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 21:01:39,761 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:41,057 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:01:44,762 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:46,058 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: keepalive
+2024-05-15 21:01:49,763 DEBUG HandlerThread:21069 [handler.py:handle_request():158] handle_request: status_report
+2024-05-15 21:01:51,018 INFO memory :21069 [interfaces.py:monitor():138] Process proc.memory.rssMB has exited.
+2024-05-15 21:01:51,027 DEBUG SystemMonitor:21069 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-15 21:01:51,028 DEBUG SystemMonitor:21069 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-15 21:01:51,034 DEBUG SenderThread:21069 [sender.py:send():378] send: stats
+2024-05-15 21:01:52,246 INFO MainThread:21069 [internal.py:handle_exit():75] Internal process exited
diff --git a/wandb/run-20240515_132302-ky6h7dv9/logs/debug.log b/wandb/run-20240515_132302-ky6h7dv9/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..a5029ccd13db145abca8bcb2d9cf4b66ab3e71e7
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-15 13:23:02,546 INFO MainThread:20482 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-15 13:23:02,546 INFO MainThread:20482 [wandb_setup.py:_flush():76] Configure stats pid to 20482
+2024-05-15 13:23:02,546 INFO MainThread:20482 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-15 13:23:02,546 INFO MainThread:20482 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-15 13:23:02,546 WARNING MainThread:20482 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-15 13:23:02,546 INFO MainThread:20482 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-15 13:23:02,546 INFO MainThread:20482 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-15 13:23:02,547 WARNING MainThread:20482 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-15 13:23:02,547 INFO MainThread:20482 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-15 13:23:02,547 INFO MainThread:20482 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-15 13:23:02,547 INFO MainThread:20482 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/logs/debug.log
+2024-05-15 13:23:02,547 INFO MainThread:20482 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240515_132302-ky6h7dv9/logs/debug-internal.log
+2024-05-15 13:23:02,547 INFO MainThread:20482 [wandb_init.py:init():560] calling init triggers
+2024-05-15 13:23:02,547 INFO MainThread:20482 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-15 13:23:02,547 INFO MainThread:20482 [wandb_init.py:init():610] starting backend
+2024-05-15 13:23:02,547 INFO MainThread:20482 [wandb_init.py:init():614] setting up manager
+2024-05-15 13:23:02,549 INFO MainThread:20482 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-15 13:23:02,551 INFO MainThread:20482 [wandb_init.py:init():622] backend started and connected
+2024-05-15 13:23:02,554 INFO MainThread:20482 [wandb_init.py:init():711] updated telemetry
+2024-05-15 13:23:02,565 INFO MainThread:20482 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-15 13:23:02,947 INFO MainThread:20482 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-15 13:23:03,015 INFO MainThread:20482 [wandb_run.py:_on_init():2405] got version response
+2024-05-15 13:23:03,016 INFO MainThread:20482 [wandb_init.py:init():795] starting run threads in backend
+2024-05-15 13:23:03,892 INFO MainThread:20482 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-15 13:23:03,892 INFO MainThread:20482 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-15 13:23:03,892 INFO MainThread:20482 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-15 13:23:03,892 INFO MainThread:20482 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-15 13:23:03,893 INFO MainThread:20482 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-15 13:23:03,895 INFO MainThread:20482 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.0.dev0', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 1, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 8628, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May15_13-22-23_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero2.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_bnb_8bit', 'optim_args': 'amsgrad=True', 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.92, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 1, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-15 13:23:03,897 INFO MainThread:20482 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240515_132302-ky6h7dv9/run-ky6h7dv9.wandb b/wandb/run-20240515_132302-ky6h7dv9/run-ky6h7dv9.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..365956e27b598dc556468deb32fb2dd152f00437
--- /dev/null
+++ b/wandb/run-20240515_132302-ky6h7dv9/run-ky6h7dv9.wandb
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e8aba43adc152a281d7345b808523a576acda186a970ebfb125375ffa85d2396
+size 17682432
diff --git a/wandb/run-20240516_032739-kmszoyl0/files/conda-environment.yaml b/wandb/run-20240516_032739-kmszoyl0/files/conda-environment.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..dc18d437c5b23969e777b5b1292c6c2af51cf43c
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/files/conda-environment.yaml
@@ -0,0 +1,79 @@
+name: base
+channels:
+ - defaults
+dependencies:
+ - _libgcc_mutex=0.1=main
+ - _openmp_mutex=5.1=1_gnu
+ - anaconda-anon-usage=0.4.4=py312hfc0e8ea_100
+ - archspec=0.2.3=pyhd3eb1b0_0
+ - boltons=23.0.0=py312h06a4308_0
+ - brotli-python=1.0.9=py312h6a678d5_7
+ - bzip2=1.0.8=h5eee18b_5
+ - c-ares=1.19.1=h5eee18b_0
+ - ca-certificates=2024.3.11=h06a4308_0
+ - certifi=2024.2.2=py312h06a4308_0
+ - cffi=1.16.0=py312h5eee18b_0
+ - charset-normalizer=2.0.4=pyhd3eb1b0_0
+ - conda=24.3.0=py312h06a4308_0
+ - conda-content-trust=0.2.0=py312h06a4308_0
+ - conda-libmamba-solver=24.1.0=pyhd3eb1b0_0
+ - conda-package-handling=2.2.0=py312h06a4308_0
+ - conda-package-streaming=0.9.0=py312h06a4308_0
+ - cryptography=42.0.5=py312hdda0065_0
+ - distro=1.8.0=py312h06a4308_0
+ - expat=2.5.0=h6a678d5_0
+ - fmt=9.1.0=hdb19cb5_0
+ - icu=73.1=h6a678d5_0
+ - idna=3.4=py312h06a4308_0
+ - jsonpatch=1.33=py312h06a4308_0
+ - jsonpointer=2.1=pyhd3eb1b0_0
+ - krb5=1.20.1=h143b758_1
+ - ld_impl_linux-64=2.38=h1181459_1
+ - libarchive=3.6.2=h6ac8c49_2
+ - libcurl=8.5.0=h251f7ec_0
+ - libedit=3.1.20230828=h5eee18b_0
+ - libev=4.33=h7f8727e_1
+ - libffi=3.4.4=h6a678d5_0
+ - libgcc-ng=11.2.0=h1234567_1
+ - libgomp=11.2.0=h1234567_1
+ - libmamba=1.5.8=hfe524e5_1
+ - libmambapy=1.5.8=py312h2dafd23_1
+ - libnghttp2=1.57.0=h2d74bed_0
+ - libsolv=0.7.24=he621ea3_0
+ - libssh2=1.10.0=hdbd6064_2
+ - libstdcxx-ng=11.2.0=h1234567_1
+ - libuuid=1.41.5=h5eee18b_0
+ - libxml2=2.10.4=hf1b16e4_1
+ - lz4-c=1.9.4=h6a678d5_0
+ - menuinst=2.0.2=py312h06a4308_0
+ - ncurses=6.4=h6a678d5_0
+ - openssl=3.0.13=h7f8727e_0
+ - packaging=23.2=py312h06a4308_0
+ - pcre2=10.42=hebb0a14_0
+ - pip=23.3.1=py312h06a4308_0
+ - platformdirs=3.10.0=py312h06a4308_0
+ - pluggy=1.0.0=py312h06a4308_1
+ - pybind11-abi=4=hd3eb1b0_1
+ - pycosat=0.6.6=py312h5eee18b_0
+ - pycparser=2.21=pyhd3eb1b0_0
+ - pysocks=1.7.1=py312h06a4308_0
+ - python=3.12.2=h996f2a0_0
+ - readline=8.2=h5eee18b_0
+ - reproc=14.2.4=h295c915_1
+ - reproc-cpp=14.2.4=h295c915_1
+ - requests=2.31.0=py312h06a4308_1
+ - ruamel.yaml=0.17.21=py312h5eee18b_0
+ - setuptools=68.2.2=py312h06a4308_0
+ - sqlite=3.41.2=h5eee18b_0
+ - tk=8.6.12=h1ccaba5_0
+ - tqdm=4.65.0=py312he106c6f_0
+ - truststore=0.8.0=py312h06a4308_0
+ - tzdata=2024a=h04d1e81_0
+ - urllib3=2.1.0=py312h06a4308_1
+ - wheel=0.41.2=py312h06a4308_0
+ - xz=5.4.6=h5eee18b_0
+ - yaml-cpp=0.8.0=h6a678d5_0
+ - zlib=1.2.13=h5eee18b_0
+ - zstandard=0.19.0=py312h5eee18b_0
+ - zstd=1.5.5=hc292b87_0
+prefix: /root/miniconda3
diff --git a/wandb/run-20240516_032739-kmszoyl0/files/config.yaml b/wandb/run-20240516_032739-kmszoyl0/files/config.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..78c206fb5b79f0a6715190a68ca0558d41f51760
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/files/config.yaml
@@ -0,0 +1,757 @@
+wandb_version: 1
+
+_wandb:
+ desc: null
+ value:
+ python_version: 3.10.14
+ cli_version: 0.17.0
+ framework: huggingface
+ huggingface_version: 4.40.2
+ is_jupyter_run: false
+ is_kaggle_kernel: false
+ start_time: 1715830059
+ t:
+ 1:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 2:
+ - 1
+ - 5
+ - 11
+ - 49
+ - 51
+ - 53
+ - 55
+ - 71
+ - 84
+ - 98
+ - 99
+ - 100
+ - 105
+ 3:
+ - 1
+ - 3
+ - 7
+ - 23
+ - 66
+ 4: 3.10.14
+ 5: 0.17.0
+ 6: 4.40.2
+ 8:
+ - 5
+ 9:
+ 1: transformers_trainer
+ 13: linux-x86_64
+ m:
+ - 1: train/global_step
+ 6:
+ - 3
+ - 1: train/loss
+ 5: 1
+ 6:
+ - 1
+ - 1: train/grad_norm
+ 5: 1
+ 6:
+ - 1
+ - 1: train/learning_rate
+ 5: 1
+ 6:
+ - 1
+ - 1: train/epoch
+ 5: 1
+ 6:
+ - 1
+vocab_size:
+ desc: null
+ value: 128256
+max_position_embeddings:
+ desc: null
+ value: 8192
+hidden_size:
+ desc: null
+ value: 4096
+intermediate_size:
+ desc: null
+ value: 14336
+num_hidden_layers:
+ desc: null
+ value: 32
+num_attention_heads:
+ desc: null
+ value: 32
+num_key_value_heads:
+ desc: null
+ value: 8
+hidden_act:
+ desc: null
+ value: silu
+initializer_range:
+ desc: null
+ value: 0.02
+rms_norm_eps:
+ desc: null
+ value: 1.0e-05
+pretraining_tp:
+ desc: null
+ value: 1
+use_cache:
+ desc: null
+ value: false
+rope_theta:
+ desc: null
+ value: 500000.0
+rope_scaling:
+ desc: null
+ value: null
+attention_bias:
+ desc: null
+ value: false
+attention_dropout:
+ desc: null
+ value: 0.0
+return_dict:
+ desc: null
+ value: true
+output_hidden_states:
+ desc: null
+ value: false
+output_attentions:
+ desc: null
+ value: false
+torchscript:
+ desc: null
+ value: false
+torch_dtype:
+ desc: null
+ value: bfloat16
+use_bfloat16:
+ desc: null
+ value: false
+tf_legacy_loss:
+ desc: null
+ value: false
+pruned_heads:
+ desc: null
+ value: {}
+tie_word_embeddings:
+ desc: null
+ value: false
+chunk_size_feed_forward:
+ desc: null
+ value: 0
+is_encoder_decoder:
+ desc: null
+ value: false
+is_decoder:
+ desc: null
+ value: false
+cross_attention_hidden_size:
+ desc: null
+ value: null
+add_cross_attention:
+ desc: null
+ value: false
+tie_encoder_decoder:
+ desc: null
+ value: false
+max_length:
+ desc: null
+ value: 20
+min_length:
+ desc: null
+ value: 0
+do_sample:
+ desc: null
+ value: false
+early_stopping:
+ desc: null
+ value: false
+num_beams:
+ desc: null
+ value: 1
+num_beam_groups:
+ desc: null
+ value: 1
+diversity_penalty:
+ desc: null
+ value: 0.0
+temperature:
+ desc: null
+ value: 1.0
+top_k:
+ desc: null
+ value: 50
+top_p:
+ desc: null
+ value: 1.0
+typical_p:
+ desc: null
+ value: 1.0
+repetition_penalty:
+ desc: null
+ value: 1.0
+length_penalty:
+ desc: null
+ value: 1.0
+no_repeat_ngram_size:
+ desc: null
+ value: 0
+encoder_no_repeat_ngram_size:
+ desc: null
+ value: 0
+bad_words_ids:
+ desc: null
+ value: null
+num_return_sequences:
+ desc: null
+ value: 1
+output_scores:
+ desc: null
+ value: false
+return_dict_in_generate:
+ desc: null
+ value: false
+forced_bos_token_id:
+ desc: null
+ value: null
+forced_eos_token_id:
+ desc: null
+ value: null
+remove_invalid_values:
+ desc: null
+ value: false
+exponential_decay_length_penalty:
+ desc: null
+ value: null
+suppress_tokens:
+ desc: null
+ value: null
+begin_suppress_tokens:
+ desc: null
+ value: null
+architectures:
+ desc: null
+ value:
+ - LlamaForCausalLM
+finetuning_task:
+ desc: null
+ value: null
+id2label:
+ desc: null
+ value:
+ '0': LABEL_0
+ '1': LABEL_1
+label2id:
+ desc: null
+ value:
+ LABEL_0: 0
+ LABEL_1: 1
+tokenizer_class:
+ desc: null
+ value: null
+prefix:
+ desc: null
+ value: null
+bos_token_id:
+ desc: null
+ value: 128000
+pad_token_id:
+ desc: null
+ value: null
+eos_token_id:
+ desc: null
+ value: 128009
+sep_token_id:
+ desc: null
+ value: null
+decoder_start_token_id:
+ desc: null
+ value: null
+task_specific_params:
+ desc: null
+ value: null
+problem_type:
+ desc: null
+ value: null
+_name_or_path:
+ desc: null
+ value: PygmalionAI/pyggel-ckpt-2947
+transformers_version:
+ desc: null
+ value: 4.40.2
+model_type:
+ desc: null
+ value: llama
+output_dir:
+ desc: null
+ value: ./pyggel
+overwrite_output_dir:
+ desc: null
+ value: false
+do_train:
+ desc: null
+ value: false
+do_eval:
+ desc: null
+ value: false
+do_predict:
+ desc: null
+ value: false
+evaluation_strategy:
+ desc: null
+ value: 'no'
+prediction_loss_only:
+ desc: null
+ value: false
+per_device_train_batch_size:
+ desc: null
+ value: 1
+per_device_eval_batch_size:
+ desc: null
+ value: 1
+per_gpu_train_batch_size:
+ desc: null
+ value: null
+per_gpu_eval_batch_size:
+ desc: null
+ value: null
+gradient_accumulation_steps:
+ desc: null
+ value: 1
+eval_accumulation_steps:
+ desc: null
+ value: 1
+eval_delay:
+ desc: null
+ value: 0
+learning_rate:
+ desc: null
+ value: 0.0001
+weight_decay:
+ desc: null
+ value: 0.01
+adam_beta1:
+ desc: null
+ value: 0.9
+adam_beta2:
+ desc: null
+ value: 0.999
+adam_epsilon:
+ desc: null
+ value: 1.0e-08
+max_grad_norm:
+ desc: null
+ value: 0.6
+num_train_epochs:
+ desc: null
+ value: 3
+max_steps:
+ desc: null
+ value: -1
+lr_scheduler_type:
+ desc: null
+ value: cosine
+lr_scheduler_kwargs:
+ desc: null
+ value: {}
+warmup_ratio:
+ desc: null
+ value: 0.0
+warmup_steps:
+ desc: null
+ value: 8628
+log_level:
+ desc: null
+ value: passive
+log_level_replica:
+ desc: null
+ value: warning
+log_on_each_node:
+ desc: null
+ value: true
+logging_dir:
+ desc: null
+ value: ./pyggel/runs/May16_03-27-11_d2827b6ba082
+logging_strategy:
+ desc: null
+ value: steps
+logging_first_step:
+ desc: null
+ value: false
+logging_steps:
+ desc: null
+ value: 1
+logging_nan_inf_filter:
+ desc: null
+ value: true
+save_strategy:
+ desc: null
+ value: steps
+save_steps:
+ desc: null
+ value: 0.022222222222222223
+save_total_limit:
+ desc: null
+ value: 4
+save_safetensors:
+ desc: null
+ value: true
+save_on_each_node:
+ desc: null
+ value: false
+save_only_model:
+ desc: null
+ value: false
+no_cuda:
+ desc: null
+ value: false
+use_cpu:
+ desc: null
+ value: false
+use_mps_device:
+ desc: null
+ value: false
+seed:
+ desc: null
+ value: 42
+data_seed:
+ desc: null
+ value: null
+jit_mode_eval:
+ desc: null
+ value: false
+use_ipex:
+ desc: null
+ value: false
+bf16:
+ desc: null
+ value: true
+fp16:
+ desc: null
+ value: false
+fp16_opt_level:
+ desc: null
+ value: O1
+half_precision_backend:
+ desc: null
+ value: auto
+bf16_full_eval:
+ desc: null
+ value: false
+fp16_full_eval:
+ desc: null
+ value: false
+tf32:
+ desc: null
+ value: false
+local_rank:
+ desc: null
+ value: 0
+ddp_backend:
+ desc: null
+ value: null
+tpu_num_cores:
+ desc: null
+ value: null
+tpu_metrics_debug:
+ desc: null
+ value: false
+debug:
+ desc: null
+ value: []
+dataloader_drop_last:
+ desc: null
+ value: false
+eval_steps:
+ desc: null
+ value: null
+dataloader_num_workers:
+ desc: null
+ value: 0
+dataloader_prefetch_factor:
+ desc: null
+ value: null
+past_index:
+ desc: null
+ value: -1
+run_name:
+ desc: null
+ value: ./pyggel
+disable_tqdm:
+ desc: null
+ value: false
+remove_unused_columns:
+ desc: null
+ value: true
+label_names:
+ desc: null
+ value: null
+load_best_model_at_end:
+ desc: null
+ value: false
+metric_for_best_model:
+ desc: null
+ value: null
+greater_is_better:
+ desc: null
+ value: null
+ignore_data_skip:
+ desc: null
+ value: false
+fsdp:
+ desc: null
+ value: []
+fsdp_min_num_params:
+ desc: null
+ value: 0
+fsdp_config:
+ desc: null
+ value:
+ min_num_params: 0
+ xla: false
+ xla_fsdp_v2: false
+ xla_fsdp_grad_ckpt: false
+fsdp_transformer_layer_cls_to_wrap:
+ desc: null
+ value: null
+accelerator_config:
+ desc: null
+ value:
+ split_batches: false
+ dispatch_batches: null
+ even_batches: true
+ use_seedable_sampler: true
+ gradient_accumulation_kwargs: null
+deepspeed:
+ desc: null
+ value: deepspeed_configs/zero1.json
+label_smoothing_factor:
+ desc: null
+ value: 0.0
+optim:
+ desc: null
+ value: adamw_bnb_8bit
+optim_args:
+ desc: null
+ value: amsgrad=True
+adafactor:
+ desc: null
+ value: false
+group_by_length:
+ desc: null
+ value: false
+length_column_name:
+ desc: null
+ value: length
+report_to:
+ desc: null
+ value:
+ - wandb
+ddp_find_unused_parameters:
+ desc: null
+ value: false
+ddp_bucket_cap_mb:
+ desc: null
+ value: null
+ddp_broadcast_buffers:
+ desc: null
+ value: null
+dataloader_pin_memory:
+ desc: null
+ value: true
+dataloader_persistent_workers:
+ desc: null
+ value: false
+skip_memory_metrics:
+ desc: null
+ value: true
+use_legacy_prediction_loop:
+ desc: null
+ value: false
+push_to_hub:
+ desc: null
+ value: true
+resume_from_checkpoint:
+ desc: null
+ value: null
+hub_model_id:
+ desc: null
+ value: Alignment-Lab-AI/pygmalion-3-1m-2.4
+hub_strategy:
+ desc: null
+ value: all_checkpoints
+hub_token:
+ desc: null
+ value:
+hub_private_repo:
+ desc: null
+ value: true
+hub_always_push:
+ desc: null
+ value: true
+gradient_checkpointing:
+ desc: null
+ value: unsloth
+gradient_checkpointing_kwargs:
+ desc: null
+ value:
+ use_reentrant: true
+include_inputs_for_metrics:
+ desc: null
+ value: false
+eval_do_concat_batches:
+ desc: null
+ value: true
+fp16_backend:
+ desc: null
+ value: auto
+push_to_hub_model_id:
+ desc: null
+ value: null
+push_to_hub_organization:
+ desc: null
+ value: null
+push_to_hub_token:
+ desc: null
+ value:
+mp_parameters:
+ desc: null
+ value: ''
+auto_find_batch_size:
+ desc: null
+ value: false
+full_determinism:
+ desc: null
+ value: false
+torchdynamo:
+ desc: null
+ value: null
+ray_scope:
+ desc: null
+ value: last
+ddp_timeout:
+ desc: null
+ value: 1800
+torch_compile:
+ desc: null
+ value: true
+torch_compile_backend:
+ desc: null
+ value: inductor
+torch_compile_mode:
+ desc: null
+ value: null
+dispatch_batches:
+ desc: null
+ value: null
+split_batches:
+ desc: null
+ value: null
+include_tokens_per_second:
+ desc: null
+ value: false
+include_num_input_tokens_seen:
+ desc: null
+ value: false
+neftune_noise_alpha:
+ desc: null
+ value: 15.0
+optim_target_modules:
+ desc: null
+ value: null
+lr_quadratic_warmup:
+ desc: null
+ value: false
+pretraining:
+ desc: null
+ value: false
+sample_packing:
+ desc: null
+ value: true
+multipack_real_batches:
+ desc: null
+ value: false
+eval_sample_packing:
+ desc: null
+ value: true
+sample_packing_efficiency:
+ desc: null
+ value: 0.92
+max_seq_length:
+ desc: null
+ value: 8192
+sample_packing_seq_len_multiplier:
+ desc: null
+ value: 1
+relora_steps:
+ desc: null
+ value: null
+relora_warmup_steps:
+ desc: null
+ value: null
+relora_anneal_steps:
+ desc: null
+ value: null
+relora_prune_ratio:
+ desc: null
+ value: 0.9
+bench_split:
+ desc: null
+ value: eval
+bench_dataset:
+ desc: null
+ value: pharaouk/dharma-1/dharma_1_mini.json
+do_bench_eval:
+ desc: null
+ value: false
+do_causal_lm_eval:
+ desc: null
+ value: false
+max_bench_samples:
+ desc: null
+ value: null
+bench_source_max_len:
+ desc: null
+ value: 2048
+cosine_min_lr_ratio:
+ desc: null
+ value: null
+cosine_constant_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_ratio:
+ desc: null
+ value: null
+loraplus_lr_embedding:
+ desc: null
+ value: 1.0e-06
+qlora:
+ desc: null
+ value: false
+orpo_alpha:
+ desc: null
+ value: null
+lisa_n_layers:
+ desc: null
+ value: null
+lisa_step_interval:
+ desc: null
+ value: null
+lisa_layers_attribute:
+ desc: null
+ value: null
+curriculum_sampling:
+ desc: null
+ value: null
diff --git a/wandb/run-20240516_032739-kmszoyl0/files/diff.patch b/wandb/run-20240516_032739-kmszoyl0/files/diff.patch
new file mode 100644
index 0000000000000000000000000000000000000000..152728c22c871d273aa4fb59226a4dde890d6b80
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/files/diff.patch
@@ -0,0 +1,93 @@
+diff --git a/src/axolotl/prompt_strategies/sharegpt.py b/src/axolotl/prompt_strategies/sharegpt.py
+index 5f0e7a8..b0e37b0 100644
+--- a/src/axolotl/prompt_strategies/sharegpt.py
++++ b/src/axolotl/prompt_strategies/sharegpt.py
+@@ -40,7 +40,7 @@ def register_chatml_template(system_message=None):
+
+
+ def register_llama3_template(system_message=None):
+- system_message = system_message or "You are a helpful assistant."
++ system_message = system_message or "You are great at roleplay"
+ register_conv_template(
+ Conversation(
+ name="llama3",
+diff --git a/src/axolotl/utils/config/models/input/v0_4_1/__init__.py b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py
+index f1c12b2..c0a3ff4 100644
+--- a/src/axolotl/utils/config/models/input/v0_4_1/__init__.py
++++ b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py
+@@ -145,7 +145,6 @@ class ChatTemplate(str, Enum):
+ cohere = "cohere" # pylint: disable=invalid-name
+ llama3 = "llama3" # pylint: disable=invalid-name
+
+-
+ class LoftQConfig(BaseModel):
+ """LoftQ configuration subset"""
+
+@@ -517,9 +516,6 @@ class AxolotlInputConfig(
+
+ sequence_len: int = Field(default=512)
+ min_sample_len: Optional[int] = None
+- max_prompt_len: int = Field(
+- default=512, metadata={"help": "maximum prompt length for RL training"}
+- )
+ sample_packing: Optional[bool] = None
+ eval_sample_packing: Optional[bool] = None
+ pad_to_sequence_len: Optional[bool] = None
+@@ -549,6 +545,11 @@ class AxolotlInputConfig(
+ flash_attn_fuse_mlp: Optional[bool] = None
+ flash_optimum: Optional[bool] = None
+
++ unsloth_cross_entropy_loss: Optional[bool] = None
++ unsloth_lora_mlp: Optional[bool] = None
++ unsloth_lora_qkv: Optional[bool] = None
++ unsloth_lora_o: Optional[bool] = None
++
+ deepspeed: Optional[Union[str, Dict[str, Any]]] = None
+ fsdp: Optional[List[str]] = None
+ fsdp_config: Optional[Dict[str, Any]] = None
+diff --git a/src/axolotl/utils/models.py b/src/axolotl/utils/models.py
+index e94a0f6..c4ca673 100644
+--- a/src/axolotl/utils/models.py
++++ b/src/axolotl/utils/models.py
+@@ -286,7 +286,7 @@ def load_tokenizer(cfg):
+ chat_template_string = chat_templates(cfg.chat_template)
+ if cfg.default_system_message and cfg.chat_template == "chatml":
+ chat_template_string = chat_template_string.replace(
+- "You are a helpful assistant.", cfg.default_system_message
++ "you are a highly intelligent sentient being", cfg.default_system_message
+ )
+
+ tokenizer.chat_template = chat_template_string
+@@ -390,6 +390,16 @@ def load_model(
+ "Shifted-sparse attention not currently implemented without flash attention."
+ )
+
++ if cfg.unsloth_cross_entropy_loss:
++ from axolotl.monkeypatch.unsloth_ import integrate_cross_entropy_loss_patch
++
++ integrate_cross_entropy_loss_patch()
++
++ if cfg.unsloth_lora_qkv or cfg.unsloth_lora_o:
++ from axolotl.monkeypatch.unsloth_ import patch_self_attn_lora
++
++ patch_self_attn_lora()
++
+ # Modify mistral derived models
+ if (
+ cfg.model_config_type == "mistral"
+@@ -828,6 +838,15 @@ def load_model(
+ if cfg.adapter is not None:
+ log_gpu_memory_usage(LOG, "after adapters", model.device)
+
++ if cfg.unsloth_lora_mlp:
++ from axolotl.monkeypatch.unsloth_ import integrate_lora_mlp_patch
++
++ integrate_lora_mlp_patch(model)
++ if cfg.unsloth_lora_qkv or cfg.unsloth_lora_o:
++ from axolotl.monkeypatch.unsloth_ import integrate_lora_patch
++
++ integrate_lora_patch(model, cfg)
++
+ # TODO resume_from_checkpoint handling
+ return model, lora_config
+
diff --git a/wandb/run-20240516_032739-kmszoyl0/files/output.log b/wandb/run-20240516_032739-kmszoyl0/files/output.log
new file mode 100644
index 0000000000000000000000000000000000000000..8770683d1078f02a58ea792afe6dd5708d8d2009
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/files/output.log
@@ -0,0 +1,29 @@
+[34m[1mwandb[39m[22m: [33mWARNING[39m Saving files without folders. If you want to preserve subdirectories pass base_path to wandb.save, i.e. wandb.save("/mnt/folder/file.h5", base_path="/mnt")
+ 0%| | 0/18756 [00:00, ?it/s]You're using a PreTrainedTokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
+/root/miniconda3/envs/py3.10/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
+ warnings.warn(
+[2024-05-16 03:27:41,617] [INFO] [axolotl.callbacks.on_train_begin:770] [PID:24653] [RANK:0] The Axolotl config has been saved to the WandB run under files.
+[2024-05-16 03:27:42,624] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:24653] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+[2024-05-16 03:27:43,601] [INFO] [axolotl.utils.samplers.multipack._len_est:184] [PID:24653] [RANK:0] packing_efficiency_estimate: 0.92 total_num_tokens per device: 47609834
+ 0%| | 1/18756 [00:12<63:21:30, 12.16s/it]
+
+ 0%| | 2/18756 [00:19<49:50:53, 9.57s/it]
+[2024-05-16 03:28:01,532] [INFO] [axolotl.callbacks.on_step_end:125] [PID:24653] [RANK:0] GPU memory usage while training: 19.590GB (+16.684GB cache, +1.237GB misc)
+
+ 0%| | 3/18756 [00:28<46:22:39, 8.90s/it]
+
+ 0%| | 4/18756 [00:35<44:17:47, 8.50s/it]
+
+ 0%| | 5/18756 [00:43<43:05:36, 8.27s/it]
+
+ 0%| | 6/18756 [00:51<42:28:05, 8.15s/it]
+
+ 0%| | 7/18756 [00:59<42:09:23, 8.09s/it]
+
+ 0%| | 8/18756 [01:07<41:55:39, 8.05s/it]
+
+ 0%| | 9/18756 [01:15<41:42:51, 8.01s/it]
+
+ 0%| | 10/18756 [01:23<41:36:29, 7.99s/it]
+
+ 0%| | 11/18756 [01:31<41:23:58, 7.95s/it]
diff --git a/wandb/run-20240516_032739-kmszoyl0/files/requirements.txt b/wandb/run-20240516_032739-kmszoyl0/files/requirements.txt
new file mode 100644
index 0000000000000000000000000000000000000000..98c6bc43747de05ee8bf70c159b9300b8af47644
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/files/requirements.txt
@@ -0,0 +1,193 @@
+GitPython==3.1.43
+Jinja2==3.1.4
+Markdown==3.6
+MarkupSafe==2.1.5
+PyYAML==6.0.1
+Pygments==2.18.0
+Werkzeug==3.0.3
+absl-py==2.1.0
+accelerate==0.30.1
+addict==2.4.0
+aiobotocore==2.7.0
+aiofiles==23.2.1
+aiohttp==3.9.5
+aioitertools==0.11.0
+aiosignal==1.3.1
+altair==5.3.0
+annotated-types==0.6.0
+anyio==4.3.0
+art==6.2
+async-timeout==4.0.3
+attrs==23.2.0
+awscli==1.32.99
+axolotl==0.4.0
+bitsandbytes==0.43.1
+botocore==1.31.64
+cachetools==5.3.3
+causal-conv1d==1.2.0.post2
+certifi==2024.2.2
+charset-normalizer==3.3.2
+click==8.1.7
+cmake==3.29.3
+colorama==0.4.6
+coloredlogs==15.0.1
+contourpy==1.2.1
+cycler==0.12.1
+datasets==2.19.1
+decorator==5.1.1
+deepspeed-kernels==0.0.1.dev1698255861
+deepspeed==0.14.2
+dill==0.3.7
+dnspython==2.6.1
+docker-pycreds==0.4.0
+docstring_parser==0.16
+docutils==0.16
+dropout_layer_norm==0.1
+einops==0.8.0
+email_validator==2.1.1
+evaluate==0.4.1
+exceptiongroup==1.2.1
+fastapi-cli==0.0.3
+fastapi==0.111.0
+fastcore==1.5.35
+ffmpy==0.3.2
+filelock==3.14.0
+fire==0.6.0
+flash_attn==2.5.8
+fonttools==4.51.0
+frozenlist==1.4.1
+fschat==0.2.36
+fsspec==2023.10.0
+galore-torch==1.0
+gcsfs==2023.10.0
+gitdb==4.0.11
+google-api-core==2.19.0
+google-auth-oauthlib==1.2.0
+google-auth==2.29.0
+google-cloud-core==2.4.1
+google-cloud-storage==2.16.0
+google-crc32c==1.5.0
+google-resumable-media==2.7.0
+googleapis-common-protos==1.63.0
+gradio==3.50.2
+gradio_client==0.6.1
+grpcio==1.63.0
+h11==0.14.0
+hf_transfer==0.1.6
+hjson==3.1.0
+httpcore==1.0.5
+httptools==0.6.1
+httpx==0.27.0
+huggingface-hub==0.23.0
+humanfriendly==10.0
+idna==3.7
+importlib_resources==6.4.0
+iniconfig==2.0.0
+jmespath==1.0.1
+joblib==1.4.2
+jsonschema-specifications==2023.12.1
+jsonschema==4.22.0
+kiwisolver==1.4.5
+llvmlite==0.42.0
+mamba-ssm==1.2.0.post1
+markdown-it-py==3.0.0
+markdown2==2.4.13
+matplotlib==3.8.4
+mdurl==0.1.2
+mpmath==1.3.0
+multidict==6.0.5
+multiprocess==0.70.15
+networkx==3.3
+nh3==0.2.17
+ninja==1.11.1.1
+numba==0.59.1
+numpy==1.26.4
+oauthlib==3.2.2
+optimum==1.16.2
+orjson==3.10.3
+packaging==23.2
+pandas==2.2.2
+peft==0.10.0
+pillow==10.3.0
+pip==24.0
+platformdirs==4.2.1
+pluggy==1.5.0
+prompt-toolkit==3.0.43
+proto-plus==1.23.0
+protobuf==4.25.3
+psutil==5.9.8
+py-cpuinfo==9.0.0
+pyarrow-hotfix==0.6
+pyarrow==16.0.0
+pyasn1==0.6.0
+pyasn1_modules==0.4.0
+pydantic==2.6.3
+pydantic_core==2.16.3
+pydub==0.25.1
+pynvml==11.5.0
+pyparsing==3.1.2
+pytest==8.2.0
+python-dateutil==2.9.0.post0
+python-dotenv==1.0.1
+python-multipart==0.0.9
+pytz==2024.1
+referencing==0.35.1
+regex==2024.5.10
+requests-oauthlib==2.0.0
+requests==2.31.0
+responses==0.18.0
+rich==13.7.1
+rpds-py==0.18.1
+rsa==4.7.2
+s3fs==2023.10.0
+s3transfer==0.10.1
+safetensors==0.4.3
+scikit-learn==1.2.2
+scipy==1.13.0
+semantic-version==2.10.0
+sentencepiece==0.2.0
+sentry-sdk==2.1.1
+setproctitle==1.3.3
+setuptools==69.5.1
+shellingham==1.5.4
+shortuuid==1.0.13
+shtab==1.7.1
+six==1.16.0
+smmap==5.0.1
+sniffio==1.3.1
+starlette==0.37.2
+svgwrite==1.4.3
+sympy==1.12
+tensorboard-data-server==0.7.2
+tensorboard==2.16.2
+termcolor==2.4.0
+threadpoolctl==3.5.0
+tiktoken==0.6.0
+tokenizers==0.19.1
+tomli==2.0.1
+toolz==0.12.1
+torch==2.1.2+cu118
+tqdm==4.66.4
+transformers==4.40.2
+triton==2.1.0
+trl==0.8.5
+typer==0.12.3
+typing_extensions==4.11.0
+tyro==0.8.4
+tzdata==2024.1
+ujson==5.9.0
+unsloth==2024.5
+urllib3==2.0.7
+uvicorn==0.29.0
+uvloop==0.19.0
+wandb==0.17.0
+watchfiles==0.21.0
+wavedrom==2.0.3.post3
+wcwidth==0.2.13
+websockets==11.0.3
+wheel==0.43.0
+wrapt==1.16.0
+xformers==0.0.23.post1+cu118
+xxhash==3.4.1
+yarl==1.9.4
+zstandard==0.22.0
\ No newline at end of file
diff --git a/wandb/run-20240516_032739-kmszoyl0/files/tmp/axolotl_config_h7qiqpxe.yml b/wandb/run-20240516_032739-kmszoyl0/files/tmp/axolotl_config_h7qiqpxe.yml
new file mode 100644
index 0000000000000000000000000000000000000000..0257223690de3c4af6b1fcbb8da42cb73368ffc5
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/files/tmp/axolotl_config_h7qiqpxe.yml
@@ -0,0 +1,66 @@
+base_model: PygmalionAI/pyggel-ckpt-2947
+load_in_8bit: false
+load_in_4bit: false
+strict: false
+
+datasets:
+ - path: PygmalionAI/unified-rp-dataset
+ type: sharegpt
+
+chat_template: llama3
+dataset_prepared_path: ./datasetstuff3
+hub_model_id: Alignment-Lab-AI/pygmalion-3-1m-2.4
+wandb_project: pyg-1-m-2.3
+hf_use_auth_token: true
+output_dir: ./pyg1m2.3
+
+wandb_watch: all
+hub_private_repo: true
+hub_strategy: all_checkpoints
+push_to_hub: true
+hf_use_auth_token: true
+output_dir: ./pyggel
+max_grad_norm: 0.6
+sequence_len: 8192
+sample_packing: true
+pad_to_sequence_len: true
+micro_batch_size: 1
+gradient_accumulation_steps: 1
+num_epochs: 3
+learning_rate: 0.0001
+optimizer: adamw_bnb_8bit
+optim_args:
+ amsgrad: true
+lr_scheduler: cosine
+train_on_inputs: true
+group_by_length: false
+bfloat16: false
+fp16:
+tf32: false
+neftune_noise_alpha: 15
+gradient_checkpointing: unsloth
+gradient_checkpointing_kwargs:
+ use_reentrant: true
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+unsloth_cross_entropy_loss: true
+#unsloth_lora_mlp: true
+#unsloth_lora_qkv: true
+#unsloth_lora_o: true
+flash_attn_cross_entropy: false
+flash_attn_rms_norm: true
+flash_attn_fuse_qkv: false
+flash_attn_fuse_mlp: true
+warmup_ratio: 0.5
+evals_per_step: 0.025
+eval_table_size:
+saves_per_epoch: 15
+debug:
+torch_compile: true
+rank:
+deepspeed: deepspeed_configs/zero1.json
+weight_decay: 0.01
+special_tokens:
+ eos_token: "<|eot_id|>"
+ pad_token: "<|end_of_text|>"
diff --git a/wandb/run-20240516_032739-kmszoyl0/files/wandb-metadata.json b/wandb/run-20240516_032739-kmszoyl0/files/wandb-metadata.json
new file mode 100644
index 0000000000000000000000000000000000000000..330fac9dff84cd6a64751c0f7fb277ea060ea27a
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/files/wandb-metadata.json
@@ -0,0 +1,565 @@
+{
+ "os": "Linux-5.15.0-106-generic-x86_64-with-glibc2.35",
+ "python": "3.10.14",
+ "heartbeatAt": "2024-05-16T03:27:40.625976",
+ "startedAt": "2024-05-16T03:27:39.977644",
+ "docker": null,
+ "cuda": null,
+ "args": [
+ "config.yml"
+ ],
+ "state": "running",
+ "program": "-m axolotl.cli.train",
+ "codePathLocal": null,
+ "git": {
+ "remote": "https://github.com/OpenAccess-AI-Collective/axolotl",
+ "commit": "039e2a03700730f333471d130dbf1f4d7922b357"
+ },
+ "email": null,
+ "root": "/workspace/disk1/axolotl",
+ "host": "d2827b6ba082",
+ "username": "root",
+ "executable": "/root/miniconda3/envs/py3.10/bin/python",
+ "cpu_count": 48,
+ "cpu_count_logical": 96,
+ "cpu_freq": {
+ "current": 1555.2083229166667,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ "cpu_freq_per_core": [
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.001,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2900.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3000.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 3400.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2300.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 2200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1200.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 1500.0,
+ "min": 800.0,
+ "max": 3500.0
+ },
+ {
+ "current": 800.0,
+ "min": 800.0,
+ "max": 3500.0
+ }
+ ],
+ "disk": {
+ "/": {
+ "total": 453.6786880493164,
+ "used": 151.98368072509766
+ }
+ },
+ "gpu": "NVIDIA A40",
+ "gpu_count": 10,
+ "gpu_devices": [
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ },
+ {
+ "name": "NVIDIA A40",
+ "memory_total": 48305799168
+ }
+ ],
+ "memory": {
+ "total": 503.51873779296875
+ }
+}
diff --git a/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json b/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
new file mode 100644
index 0000000000000000000000000000000000000000..67ca8c8e2a1d5edfa04de34275852b40d8d5fe39
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
@@ -0,0 +1 @@
+{"train/loss": 2.1299, "train/grad_norm": 22.632335662841797, "train/learning_rate": 1.2749188687992583e-07, "train/epoch": 0.0017594369801663468, "train/global_step": 11, "_timestamp": 1715830152.9823148, "_runtime": 92.9932188987732, "_step": 10}
\ No newline at end of file
diff --git a/wandb/run-20240516_032739-kmszoyl0/logs/debug-internal.log b/wandb/run-20240516_032739-kmszoyl0/logs/debug-internal.log
new file mode 100644
index 0000000000000000000000000000000000000000..fe979cae789bc1173bd85436f162519bfc46cc46
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/logs/debug-internal.log
@@ -0,0 +1,179 @@
+2024-05-16 03:27:39,985 INFO StreamThr :25247 [internal.py:wandb_internal():85] W&B internal server running at pid: 25247, started at: 2024-05-16 03:27:39.984560
+2024-05-16 03:27:39,987 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status
+2024-05-16 03:27:39,992 INFO WriterThread:25247 [datastore.py:open_for_write():87] open: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/run-kmszoyl0.wandb
+2024-05-16 03:27:39,993 DEBUG SenderThread:25247 [sender.py:send():378] send: header
+2024-05-16 03:27:40,003 DEBUG SenderThread:25247 [sender.py:send():378] send: run
+2024-05-16 03:27:40,392 INFO SenderThread:25247 [dir_watcher.py:__init__():211] watching files in: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files
+2024-05-16 03:27:40,393 INFO SenderThread:25247 [sender.py:_start_run_threads():1123] run started: kmszoyl0 with start time 1715830059.989096
+2024-05-16 03:27:40,425 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: check_version
+2024-05-16 03:27:40,426 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: check_version
+2024-05-16 03:27:40,511 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: run_start
+2024-05-16 03:27:40,545 DEBUG HandlerThread:25247 [system_info.py:__init__():26] System info init
+2024-05-16 03:27:40,545 DEBUG HandlerThread:25247 [system_info.py:__init__():41] System info init done
+2024-05-16 03:27:40,545 INFO HandlerThread:25247 [system_monitor.py:start():194] Starting system monitor
+2024-05-16 03:27:40,545 INFO SystemMonitor:25247 [system_monitor.py:_start():158] Starting system asset monitoring threads
+2024-05-16 03:27:40,546 INFO HandlerThread:25247 [system_monitor.py:probe():214] Collecting system info
+2024-05-16 03:27:40,546 INFO SystemMonitor:25247 [interfaces.py:start():188] Started cpu monitoring
+2024-05-16 03:27:40,547 INFO SystemMonitor:25247 [interfaces.py:start():188] Started disk monitoring
+2024-05-16 03:27:40,547 INFO SystemMonitor:25247 [interfaces.py:start():188] Started gpu monitoring
+2024-05-16 03:27:40,548 INFO SystemMonitor:25247 [interfaces.py:start():188] Started memory monitoring
+2024-05-16 03:27:40,549 INFO SystemMonitor:25247 [interfaces.py:start():188] Started network monitoring
+2024-05-16 03:27:40,625 DEBUG HandlerThread:25247 [system_info.py:probe():150] Probing system
+2024-05-16 03:27:40,628 DEBUG HandlerThread:25247 [system_info.py:_probe_git():135] Probing git
+2024-05-16 03:27:40,641 DEBUG HandlerThread:25247 [system_info.py:_probe_git():143] Probing git done
+2024-05-16 03:27:40,641 DEBUG HandlerThread:25247 [system_info.py:probe():198] Probing system done
+2024-05-16 03:27:40,641 DEBUG HandlerThread:25247 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-106-generic-x86_64-with-glibc2.35', 'python': '3.10.14', 'heartbeatAt': '2024-05-16T03:27:40.625976', 'startedAt': '2024-05-16T03:27:39.977644', 'docker': None, 'cuda': None, 'args': ('config.yml',), 'state': 'running', 'program': '-m axolotl.cli.train', 'codePathLocal': None, 'git': {'remote': 'https://github.com/OpenAccess-AI-Collective/axolotl', 'commit': '039e2a03700730f333471d130dbf1f4d7922b357'}, 'email': None, 'root': '/workspace/disk1/axolotl', 'host': 'd2827b6ba082', 'username': 'root', 'executable': '/root/miniconda3/envs/py3.10/bin/python', 'cpu_count': 48, 'cpu_count_logical': 96, 'cpu_freq': {'current': 1555.2083229166667, 'min': 800.0, 'max': 3500.0}, 'cpu_freq_per_core': [{'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.001, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2900.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 1800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 3000.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 3400.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2300.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 2200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1200.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}, {'current': 1500.0, 'min': 800.0, 'max': 3500.0}, {'current': 800.0, 'min': 800.0, 'max': 3500.0}], 'disk': {'/': {'total': 453.6786880493164, 'used': 151.98368072509766}}, 'gpu': 'NVIDIA A40', 'gpu_count': 10, 'gpu_devices': [{'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}, {'name': 'NVIDIA A40', 'memory_total': 48305799168}], 'memory': {'total': 503.51873779296875}}
+2024-05-16 03:27:40,641 INFO HandlerThread:25247 [system_monitor.py:probe():224] Finished collecting system info
+2024-05-16 03:27:40,642 INFO HandlerThread:25247 [system_monitor.py:probe():227] Publishing system info
+2024-05-16 03:27:40,642 DEBUG HandlerThread:25247 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
+2024-05-16 03:27:41,157 DEBUG HandlerThread:25247 [system_info.py:_save_conda():222] Saving conda packages done
+2024-05-16 03:27:41,157 DEBUG HandlerThread:25247 [system_info.py:_save_code():44] Saving code
+2024-05-16 03:27:41,157 WARNING HandlerThread:25247 [system_info.py:_save_code():46] unable to save code -- program entry not found
+2024-05-16 03:27:41,157 DEBUG HandlerThread:25247 [system_info.py:_save_patches():82] Saving git patches
+2024-05-16 03:27:41,183 DEBUG HandlerThread:25247 [system_info.py:_save_patches():124] Saving git patches done
+2024-05-16 03:27:41,184 INFO HandlerThread:25247 [system_monitor.py:probe():229] Finished publishing system info
+2024-05-16 03:27:41,189 DEBUG SenderThread:25247 [sender.py:send():378] send: files
+2024-05-16 03:27:41,189 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
+2024-05-16 03:27:41,190 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file diff.patch with policy now
+2024-05-16 03:27:41,395 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-metadata.json
+2024-05-16 03:27:41,396 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/conda-environment.yaml
+2024-05-16 03:27:41,396 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/diff.patch
+2024-05-16 03:27:41,409 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: python_packages
+2024-05-16 03:27:41,409 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: python_packages
+2024-05-16 03:27:41,410 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:27:41,410 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:27:41,614 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: log_artifact
+2024-05-16 03:27:41,745 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,745 DEBUG SenderThread:25247 [sender.py:send():378] send: config
+2024-05-16 03:27:41,747 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:41,747 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,747 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,747 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:41,748 WARNING SenderThread:25247 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
+2024-05-16 03:27:41,748 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,748 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:41,748 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: log_artifact
+2024-05-16 03:27:41,838 INFO wandb-upload_0:25247 [upload_job.py:push():130] Uploaded file /tmp/tmpz_h0aag_wandb/6blsqx0w-wandb-metadata.json
+2024-05-16 03:27:42,393 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/requirements.txt
+2024-05-16 03:27:42,394 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/tmp/axolotl_config_h7qiqpxe.yml
+2024-05-16 03:27:42,394 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/tmp
+2024-05-16 03:27:42,990 INFO wandb-upload_0:25247 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmptvdelfam
+2024-05-16 03:27:43,725 INFO SenderThread:25247 [sender.py:send_request_log_artifact():1455] logged artifact config-kmszoyl0 - {'id': 'QXJ0aWZhY3Q6ODM4MjMxMDk3', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NTQxMDQ4Mg==', 'latestArtifact': None}}
+2024-05-16 03:27:43,726 DEBUG SenderThread:25247 [sender.py:send():378] send: telemetry
+2024-05-16 03:27:43,727 DEBUG SenderThread:25247 [sender.py:send():378] send: files
+2024-05-16 03:27:43,727 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file tmp/axolotl_config_h7qiqpxe.yml with policy live
+2024-05-16 03:27:44,118 INFO wandb-upload_0:25247 [upload_job.py:push():130] Uploaded file /tmp/tmpz_h0aag_wandb/pz983g49-tmp/axolotl_config_h7qiqpxe.yml
+2024-05-16 03:27:44,395 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:27:45,622 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:27:46,396 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:27:50,623 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:27:51,779 INFO wandb-upload_1:25247 [upload_job.py:push():130] Uploaded file /tmp/tmpz_h0aag_wandb/s1zo24fv-diff.patch
+2024-05-16 03:27:53,781 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:27:53,786 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:53,787 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:53,787 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:53,788 DEBUG SenderThread:25247 [sender.py:send():378] send: metric
+2024-05-16 03:27:53,788 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:27:53,788 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:27:53,791 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:27:54,401 INFO Thread-12 :25247 [dir_watcher.py:_on_file_created():271] file/dir created: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:27:55,793 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:27:56,403 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:27:56,410 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:27:56,411 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:01,534 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:01,535 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:01,537 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:01,538 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:01,543 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:02,407 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:04,408 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:06,545 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:09,646 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:09,648 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:09,649 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:09,653 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:10,412 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:11,410 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:28:11,411 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:11,664 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:12,418 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/config.yaml
+2024-05-16 03:28:12,418 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:16,911 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:17,538 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:17,540 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:17,542 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:17,545 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:18,422 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:20,423 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:22,547 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:25,403 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:25,405 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:25,407 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:25,410 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:25,426 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:26,411 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:28:26,412 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:27,644 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:28,428 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:32,644 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:33,324 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:33,326 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:33,327 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:33,334 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:33,432 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:34,432 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:36,434 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:38,336 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:40,550 DEBUG SystemMonitor:25247 [system_monitor.py:_start():172] Starting system metrics aggregation loop
+2024-05-16 03:28:40,555 DEBUG SenderThread:25247 [sender.py:send():378] send: stats
+2024-05-16 03:28:41,296 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:41,299 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:41,300 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:41,301 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:41,411 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:28:41,412 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:41,437 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:42,438 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:43,629 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:48,630 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:49,254 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:49,257 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:49,259 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:49,262 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:49,443 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:50,444 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:54,264 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:28:57,171 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:28:57,172 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:28:57,178 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:28:57,360 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:28:57,360 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:28:57,361 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:28:57,448 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:28:58,449 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:28:59,363 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:04,363 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:05,121 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:29:05,124 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:29:05,125 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:29:05,129 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:29:05,453 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:29:06,454 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:29:10,131 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:10,559 DEBUG SenderThread:25247 [sender.py:send():378] send: stats
+2024-05-16 03:29:12,977 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: stop_status
+2024-05-16 03:29:12,978 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: stop_status
+2024-05-16 03:29:13,022 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: partial_history
+2024-05-16 03:29:13,201 DEBUG SenderThread:25247 [sender.py:send():378] send: history
+2024-05-16 03:29:13,202 DEBUG SenderThread:25247 [sender.py:send_request():405] send_request: summary_record
+2024-05-16 03:29:13,203 INFO SenderThread:25247 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end
+2024-05-16 03:29:13,459 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/wandb-summary.json
+2024-05-16 03:29:14,459 INFO Thread-12 :25247 [dir_watcher.py:_on_file_modified():288] file/dir modified: /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/files/output.log
+2024-05-16 03:29:15,204 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:20,205 DEBUG HandlerThread:25247 [handler.py:handle_request():158] handle_request: status_report
+2024-05-16 03:29:22,578 INFO memory :25247 [interfaces.py:monitor():138] Process proc.memory.rssMB has exited.
+2024-05-16 03:29:22,578 DEBUG SystemMonitor:25247 [system_monitor.py:_start():179] Finished system metrics aggregation loop
+2024-05-16 03:29:22,580 DEBUG SystemMonitor:25247 [system_monitor.py:_start():183] Publishing last batch of metrics
+2024-05-16 03:29:22,589 DEBUG SenderThread:25247 [sender.py:send():378] send: stats
+2024-05-16 03:29:23,536 INFO MainThread:25247 [internal.py:handle_exit():75] Internal process exited
diff --git a/wandb/run-20240516_032739-kmszoyl0/logs/debug.log b/wandb/run-20240516_032739-kmszoyl0/logs/debug.log
new file mode 100644
index 0000000000000000000000000000000000000000..67689dec5bd9a6acc4041f1c28375efdad1c362f
--- /dev/null
+++ b/wandb/run-20240516_032739-kmszoyl0/logs/debug.log
@@ -0,0 +1,31 @@
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Configure stats pid to 24653
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Loading settings from /workspace/disk1/axolotl/wandb/settings
+2024-05-16 03:27:39,980 WARNING MainThread:24653 [wandb_setup.py:_flush():76] Unknown environment variable: WANDB_WATCH
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Loading settings from environment variables: {'project': 'pyg-1-m-2.3'}
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
+2024-05-16 03:27:39,980 WARNING MainThread:24653 [wandb_setup.py:_flush():76] Could not find program at -m axolotl.cli.train
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m axolotl.cli.train'}
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_setup.py:_flush():76] Applying login settings: {}
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_init.py:_log_setup():520] Logging user logs to /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/logs/debug.log
+2024-05-16 03:27:39,980 INFO MainThread:24653 [wandb_init.py:_log_setup():521] Logging internal logs to /workspace/disk1/axolotl/wandb/run-20240516_032739-kmszoyl0/logs/debug-internal.log
+2024-05-16 03:27:39,981 INFO MainThread:24653 [wandb_init.py:init():560] calling init triggers
+2024-05-16 03:27:39,981 INFO MainThread:24653 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
+config: {}
+2024-05-16 03:27:39,981 INFO MainThread:24653 [wandb_init.py:init():610] starting backend
+2024-05-16 03:27:39,981 INFO MainThread:24653 [wandb_init.py:init():614] setting up manager
+2024-05-16 03:27:39,983 INFO MainThread:24653 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
+2024-05-16 03:27:39,988 INFO MainThread:24653 [wandb_init.py:init():622] backend started and connected
+2024-05-16 03:27:39,992 INFO MainThread:24653 [wandb_init.py:init():711] updated telemetry
+2024-05-16 03:27:40,002 INFO MainThread:24653 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
+2024-05-16 03:27:40,424 INFO MainThread:24653 [wandb_run.py:_on_init():2396] communicating current version
+2024-05-16 03:27:40,498 INFO MainThread:24653 [wandb_run.py:_on_init():2405] got version response
+2024-05-16 03:27:40,498 INFO MainThread:24653 [wandb_init.py:init():795] starting run threads in backend
+2024-05-16 03:27:41,410 INFO MainThread:24653 [wandb_run.py:_console_start():2374] atexit reg
+2024-05-16 03:27:41,410 INFO MainThread:24653 [wandb_run.py:_redirect():2229] redirect: wrap_raw
+2024-05-16 03:27:41,410 INFO MainThread:24653 [wandb_run.py:_redirect():2294] Wrapping output streams.
+2024-05-16 03:27:41,410 INFO MainThread:24653 [wandb_run.py:_redirect():2319] Redirects installed.
+2024-05-16 03:27:41,411 INFO MainThread:24653 [wandb_init.py:init():838] run started, returning control to user process
+2024-05-16 03:27:41,413 INFO MainThread:24653 [wandb_run.py:_config_callback():1376] config_cb None None {'vocab_size': 128256, 'max_position_embeddings': 8192, 'hidden_size': 4096, 'intermediate_size': 14336, 'num_hidden_layers': 32, 'num_attention_heads': 32, 'num_key_value_heads': 8, 'hidden_act': 'silu', 'initializer_range': 0.02, 'rms_norm_eps': 1e-05, 'pretraining_tp': 1, 'use_cache': False, 'rope_theta': 500000.0, 'rope_scaling': None, 'attention_bias': False, 'attention_dropout': 0.0, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'bfloat16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['LlamaForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 128000, 'pad_token_id': None, 'eos_token_id': 128009, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'PygmalionAI/pyggel-ckpt-2947', 'transformers_version': '4.40.2', 'model_type': 'llama', 'output_dir': './pyggel', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': 1, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 0.6, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'cosine', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 8628, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './pyggel/runs/May16_03-27-11_d2827b6ba082', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 1, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 0.022222222222222223, 'save_total_limit': 4, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': True, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': False, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': './pyggel', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'gradient_accumulation_kwargs': None}, 'deepspeed': 'deepspeed_configs/zero1.json', 'label_smoothing_factor': 0.0, 'optim': 'adamw_bnb_8bit', 'optim_args': 'amsgrad=True', 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': False, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': 'Alignment-Lab-AI/pygmalion-3-1m-2.4', 'hub_strategy': 'all_checkpoints', 'hub_token': '', 'hub_private_repo': True, 'hub_always_push': True, 'gradient_checkpointing': 'unsloth', 'gradient_checkpointing_kwargs': {'use_reentrant': True}, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': True, 'torch_compile_backend': 'inductor', 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': 15.0, 'optim_target_modules': None, 'lr_quadratic_warmup': False, 'pretraining': False, 'sample_packing': True, 'multipack_real_batches': False, 'eval_sample_packing': True, 'sample_packing_efficiency': 0.92, 'max_seq_length': 8192, 'sample_packing_seq_len_multiplier': 1, 'relora_steps': None, 'relora_warmup_steps': None, 'relora_anneal_steps': None, 'relora_prune_ratio': 0.9, 'bench_split': 'eval', 'bench_dataset': 'pharaouk/dharma-1/dharma_1_mini.json', 'do_bench_eval': False, 'do_causal_lm_eval': False, 'max_bench_samples': None, 'bench_source_max_len': 2048, 'cosine_min_lr_ratio': None, 'cosine_constant_lr_ratio': None, 'loraplus_lr_ratio': None, 'loraplus_lr_embedding': 1e-06, 'qlora': False, 'orpo_alpha': None, 'lisa_n_layers': None, 'lisa_step_interval': None, 'lisa_layers_attribute': None, 'curriculum_sampling': None}
+2024-05-16 03:27:41,415 INFO MainThread:24653 [wandb_watch.py:watch():51] Watching
diff --git a/wandb/run-20240516_032739-kmszoyl0/run-kmszoyl0.wandb b/wandb/run-20240516_032739-kmszoyl0/run-kmszoyl0.wandb
new file mode 100644
index 0000000000000000000000000000000000000000..6988df237053820e6fd3187ec3ddffb533d7cfc7
Binary files /dev/null and b/wandb/run-20240516_032739-kmszoyl0/run-kmszoyl0.wandb differ